.. blogpost:: :title: Deep learning and other readings :keywords: machine learning, dask, OSM, Open Street Map :date: 2015-12-22 :categories: deep learning I came accross the following article `Evaluation of Deep Learning Toolkits `_ which studies a short list of libraries for deep learning: Caffe, CNTK, TensorFlow, Theano, Torch, and various angles: modeling capability, interfaces, model deployment, performance, architecture, ecosystem, cross-platform. It gives a nice overview and helps choosing the library which fits your needs. Once your deep models has been trained, how to use it? This question should be the first one to be answered. As machine learning and big data become more and more popular, people look for ways to simplify the implementation of complex chains of processings. Python is quite popular so here is one suggestion in that language for deep learning: `Blocks and Fuel: Frameworks for deep learning `_ *(Bart van Merriƫnboer, Dzmitry Bahdanau, Vincent Dumoulin, Dmitriy Serdyuk, David Warde-Farley, Jan Chorowski, Yoshua Bengio)*. It introduces `Fuel `_ which models pipelines of data processing. Finally, a nice tutorial on machine learning with Python: `PyData Seattle 2015 Scikit-learn Tutorial `_. The author's blog is nice too: `Pythonic Perambulations `_. See `Out-of-Core Dataframes in Python: Dask and OpenStreetMap `_. Some modules are hidden in his blog posts such as `gatspy `_ which plots timeseries in many ways or `supersmoother `_ to smooth timeseries or `line_profiler `_ in `Optimizing Python in the Real World: NumPy, Numba, and the NUFFT `_. Two other readings to conclude: `Why Python is Slow: Looking Under the Hood `_ and `Frequentism and Bayesianism: A Practical Introduction `_ still from the same source. **Addition (01/06/16)**: see also `Comparative Study of Caffe, Neon, Theano, and Torch for Deep Learning `_ Source: `XD Blogs `_.