Docker cheat sheet to handle the Docker universe

Before a while I got in touch with the wonderful world of the Docker (see https://www.docker.com) cosmos. I think it is one of the most important virtualisation technology at the moment. On this page I will just collect useful commands that helped me a lot in my daily work with Docker, Docker-Compose (https://docs.docker.com/compose/) and other stack…

Magento change attribute scope in setup script

This code snippet shows how to change attribute scope in a setup script: <?php /* @var $oInstaller Mage_Catalog_Model_Resource_Setup */ $oInstaller = $this; $oInstaller->startSetup(); $oInstaller->updateAttribute( ‘catalog_product’, ‘abc_attribute’, ‘is_global’, Mage_Catalog_Model_Resource_Eav_Attribute::SCOPE_STORE ); $oInstaller->updateAttribute( ‘catalog_product’, ‘xyz_attribute’, ‘is_global’, Mage_Catalog_Model_Resource_Eav_Attribute::SCOPE_STORE ); $oInstaller->endSetup(); // mark index as “reindex required” $aIndexerCodes = [ ‘catalog_product_attribute’, ‘catalog_product_flat’, ‘catalogsearch_fulltext’ ]; $oIndexer = Mage::getModel(‘index/process’); foreach ($aIndexerCodes as $sCode)…

AWS reinitialize Elastic Beanstalk environment

To initialize a Amazon Webservice Elastic Beanstalk environment you generally use the command: $> eb init But sometimes you need to reinitialize the environment, because you made a mistake or want to change the programming language or AWS region of the connected environment. In these cases you can reinitialize Elastic Beanstalk Environment by the -i parameter: $>…

Build a website crawler using Scrapy Framework

Before a while I discovered the website crawling framework Scrapy as mentioned in my earlier post Nice Python website crawler framework. Now I wrote a little website crawler using this genius framework. Here are the steps I’ve taken to get my blog https://www.ask-sheldon.com crawled: Installation of Scrapy Install Python package management system (pip): $> sudo apt-get install…