Applied Deep Learning Resources
Applied Deep Learning Resources
A collection of research articles, blog posts, slides and code snippets about deep learning in applied settings. Including trained models and simple methods that can be used out of the box. Mainly focusing on Convolutional Neural Networks (CNN) but Recurrent Neural Networks (RNN), deep Q-Networks (DQN) and other interesting architectures will also be listed.
CNN
Latest overview of the CNNs can be found from the paper "Deep learning for visual understanding: A review" [link, PDF]
Another decent overview in Nature by LeCun, Bengio and Hinton: "Deep learning" [link, PDF]
ImageNet
ImageNet is the most important image classification and localization competition. Other data sets with results can be found from here: "Discover the current state of the art in objects classification." [link].

Prediction error of the ImageNet competition has been decreasing rapidly over the last 5 years: 
Main network architectures on ImageNet
AlexNet
Original paper: "ImageNet Classification with Deep Convolutional Neural Networks" [PDF]
Properties: 8 weight layers (5 convolutional and 2 fully connected), 60 million parameters, Rectified Linear Units (ReLUs), Local Response Normalization, Dropout

VGG
Original paper: "Very Deep Convolutional Networks for Large-Scale Image Recognition" [arxiv]
Properties: 19 weight layers, 144m parameters, 3x3 convolution filters, L2 regularised, Dropout, No Local Response Normalization
GoogLeNet
Original paper: "Going deeper with convolutions" [arxiv]
Lates upgrade to the model achieves even better scores with models and import to Torch: "Rethinking the Inception Architecture for Computer Vision" [arxiv], "Torch port of Inception V3" [github]
Properties: 22 layers, 7m parameters, Inception modules, 1x1 conv layers, ReLUs, Dropout, Mid-level outputs
Inception modules:

ResNet
Original paper: "Deep Residual Learning for Image Recognition" [arxiv]
Very nice slides: "Deep Residual Learning" [PDF]
Github: [github]
Properties: 152 layers, ReLUs, Batch Normalization (See "Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift" [arxiv]), less hacks (no dropout), more stable (different number of layers work as well) and lower complexity than VGG.
Main building block of the network:

Features are also very good and transferable with (faster) R-CNNs (see below):

Other architectures
Deep Learning for 3D shapes: "3D ShapeNets: A Deep Representation for Volumetric Shapes" [PDF]
Code and a model for faces: "Free and open source face recognition with deep neural networks." [github]
Fast neural networks which can perform arbitrary filters for images: "Deep Edge-Aware Filters" [PDF]
Lot's of different models in Caffe's "Model Zoo" [github]
Feature learning and object detection
"CNN Features off-the-shelf: an Astounding Baseline for Recognition" [arxiv]
First paper about R-CNN: "Rich feature hierarchies for accurate object detection and semantic segmentation" [PDF, slides]
"Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks" [arxiv, github, Slides]

- "An Empirical Evaluation of Deep Learning on Highway Driving" [arxiv]
 

- "Object Detectors Emerge in Deep Scene CNNs" [arxiv]
 

- Faster and better features: "Efficient Deep Feature Learning and Extraction via StochasticNets" [arxiv]
 
Other
- Code and models for automatic captions of images: "Deep Visual-Semantic Alignments for Generating Image Descriptions"[web poster, PDF, github]
 

- Google Deep Dream or neural networks on LSD: "Inceptionism: Going Deeper into Neural Networks" [link,deepdreamer.io/]
 
Deep dreaming from noise:

- Code and models for mixing arbitrary content and art style: "A Neural Algorithm of Artistic Style" [arxiv, deepart.io, a blog post, github.com/jcjohnson/neural-style]
 

- "Automatic Colorization" and it includes a pre-trained model [Link]
 

- "Learning visual similarity for product design with convolutional neural networks" [PDF]
 

Using images and image descriptions to improve search results: "Images Don’t Lie: Transferring Deep Visual Semantic Features to Large-Scale Multimodal Learning to Rank" [arxiv]
"How Google Translate squeezes deep learning onto a phone" [post]

- "What a Deep Neural Network thinks about your #selfie" [blog]
 
Top selfies according to the ConvNet:

"Recommending music on Spotify with deep learning" [github]
"DeepStereo: Learning to Predict New Views from the World's Imagery" [arxiv]

- Classifying street signs: "The power of Spatial Transformer Networks" [blog] with "Spatial Transformer Networks" [arxiv]
 

- "Pedestrian Detection with RCNN" [PDF]
 
DQN
Original paper: "Playing Atari with Deep Reinforcement Learning" [arxiv]
My popular science article about DQN: "Artificial General Intelligence that plays Atari video games: How did DeepMind do it?" [link]
DQN for RoboCup: "Deep Reinforcement Learning in Parameterized Action Space" [arxiv]
RNN
Original paper of the best RNN architecture: "Long short-term memory" [PDF]
Very good tutorial-like introduction to RNNs by Andrej Karpathy: "The Unreasonable Effectiveness of Recurrent Neural Networks" [link]
"Visualizing and Understanding Recurrent Networks" [arxiv]
"Composing Music With Recurrent Neural Networks" [blog]
Other promising or useful architectures
HTMs by Jeff Hawkins: "Continuous online sequence learning with an unsupervised neural network model" [arxiv]
Word2vec: "Efficient Estimation of Word Representations in Vector Space" [arxiv, Google code]
"Feedforward Sequential Memory Networks: A New Structure to Learn Long-term Dependency" [arxiv]
Framework benchmarks
- "Comparative Study of Caffe, Neon, Theano and Torch for deep learning" [arxiv]
 
Their summary: From our experiments, we observe that Theano and Torch are the most easily extensible frameworks. We observe that Torch is best suited for any deep architecture on CPU, followed by Theano. It also achieves the best performance on the GPU for large convolutional and fully connected networks, followed closely by Neon. Theano achieves the best performance on GPU for training and deployment of LSTM networks. Finally Caffe is the easiest for evaluating the performance of standard deep architectures.
Very good qualitative analysis: zer0n/deepframeworks: [github]
Just performance comparison: soumith/convnet-benchmarks: [github]
"Deep Learning Libraries by Language" [link]
Other resources
"An awesome list of (large-scale) public datasets on the Internet. (On-going collection)" [github]
"Model Zoo" [github]
Videos from "Deep Learning Summer School, Montreal 2015": http://videolectures.net/deeplearning2015_montreal/
Credits
Most of the snippets have come to my attention via internal mailing lists of Computational Neuroscience Lab at University of Tartu and London-based visual search company Dream It Get It. I am also reading a weekly newsletter by Data Elixir and checking research papers of the two main deep learning conferences: ICML and NIPS.
Applied Deep Learning Resources的更多相关文章
- (转) Deep Learning Resources
		
转自:http://www.jeremydjacksonphd.com/category/deep-learning/ Deep Learning Resources Posted on May 13 ...
 - why deep learning works
		
https://medium.com/towards-data-science/deep-learning-for-object-detection-a-comprehensive-review-73 ...
 - 深度学习阅读列表 Deep Learning Reading List
		
Reading List List of reading lists and survey papers: Books Deep Learning, Yoshua Bengio, Ian Goodfe ...
 - [C1W4] Neural Networks and Deep Learning - Deep Neural Networks
		
第四周:深层神经网络(Deep Neural Networks) 深层神经网络(Deep L-layer neural network) 目前为止我们学习了只有一个单独隐藏层的神经网络的正向传播和反向 ...
 - 论文笔记:A Review on Deep Learning Techniques Applied to Semantic Segmentation
		
A Review on Deep Learning Techniques Applied to Semantic Segmentation 2018-02-22 10:38:12 1. Intr ...
 - 机器学习(Machine Learning)&深度学习(Deep Learning)资料
		
<Brief History of Machine Learning> 介绍:这是一篇介绍机器学习历史的文章,介绍很全面,从感知机.神经网络.决策树.SVM.Adaboost到随机森林.D ...
 - 机器学习(Machine Learning)&深入学习(Deep Learning)资料
		
<Brief History of Machine Learning> 介绍:这是一篇介绍机器学习历史的文章,介绍很全面,从感知机.神经网络.决策树.SVM.Adaboost 到随机森林. ...
 - Machine and Deep Learning with Python
		
Machine and Deep Learning with Python Education Tutorials and courses Supervised learning superstiti ...
 - Understanding Convolution in Deep Learning
		
Understanding Convolution in Deep Learning Convolution is probably the most important concept in dee ...
 
随机推荐
- Mac运行exe的几种方法,欢迎补充!
			
1. 用wine直接运行exe.安装wine后有个放exe的文件夹,双击后会自动包装运行.看起来挺方便的,就怕暂用资源比较大: http://www.youtube.com/watch?v=eYISV ...
 - 《day15---多线程安全问题_JDK1.5的锁机制》
			
//15同步问题的分析案例以及解决思路 //两个客户到一个银行去存钱,每个客户一次存100,存3次. //问题,该程序是否有安全问题,如果有,写出分析过程,并定于解决方案. /* 发现运行结果: su ...
 - poj2184 背包
			
//Accepted 1492 KB 110 ms //背包 //把si看成weight,Fi看成value,这可以表示成当dp[j]=max(dp[j-weight[i]]+value[i]) // ...
 - Java中接口作为方法的返回
			
在<算法>中的散列表一节,在用拉链法实现散列表的API时要求实现以下一个方法: public Iterable<Key> keys() 我们知道Iterable是一个接口,那么 ...
 - Ubuntu 升级VisualBox后无法启动 Kernel driver not installed (rc=-1908)
			
VisualBox之所以在Linux上比传统的VMware快得多,关键一点就是它和Linux内核的结合比较紧密,这也是开源的优点. 不过Linux内核更新很频繁,每次更新内核后启动VirtualBox ...
 - BZOJ 2393 Cirno的完美算数教室
			
就是爆搜嘛. 先从大到小排个序能减去dfs树上很大的一部分.这个技巧要掌握. #include<iostream> #include<cstdio> #include<c ...
 - C#指针操作Marshal实例
			
static void Main(string[] args) { ,,,}; ,,,}; IntPtr pt = Marshal.AllocHGlobal(a.Length); //从source数 ...
 - hadoop权威指南 读书笔记
			
1 数据库和MapReduce对比 2 hadoop 包含的内容
 - 继承自CCObject的对象成员变量出错或者为空的问题
			
写了个类想让其作为某种数据集合,还可以自动销毁,所以就直接继承了最底层的CCObject,所以并不属于视图,也就不会被addChild到显示列表里,于是就造成了接下来遇到的一个情况:其所有的成员变量被 ...
 - 解决Eclipse Pydev中import时报错:Unresolved import
			
在安装 图像处理工具包 mahotas 后,在eclipse中尝试import mahotas时,出现Unresolved import错误,按快捷无法自动生成代码提示 但是,程序运行时可以通过,在命 ...