http://handong1587.github.io/deep_learning/2015/10/09/rnn-and-lstm.html  //RNN and LSTM

http://handong1587.github.io/deep_learning/2015/10/09/saliency-prediction.html //saliency Predection

http://handong1587.github.io/deep_learning/2015/10/09/scene-labeling.html //Scene Label

RNN and LSTM

Published: 09 Oct 2015  Category: deep_learning

Types of RNN

1) Plain Tanh Recurrent Nerual Networks

2) Gated Recurrent Neural Networks (GRU)

3) Long Short-Term Memory (LSTM)

Tutorials

A Beginner’s Guide to Recurrent Networks and LSTMs

http://deeplearning4j.org/lstm.html

A Deep Dive into Recurrent Neural Nets

http://nikhilbuduma.com/2015/01/11/a-deep-dive-into-recurrent-neural-networks/

Long Short-Term Memory: Tutorial on LSTM Recurrent Networks

http://people.idsia.ch/~juergen/lstm/index.htm

LSTM implementation explained

http://apaszke.github.io/lstm-explained.html

Recurrent Neural Networks Tutorial

Understanding LSTM Networks

Recurrent Neural Networks in DL4J

http://deeplearning4j.org/usingrnns.html

Train RNN

A Simple Way to Initialize Recurrent Networks of Rectified Linear Units

Sequence Level Training with Recurrent Neural Networks

Papers

Generating Sequences With Recurrent Neural Networks

DRAW: A Recurrent Neural Network For Image Generation

Unsupervised Learning of Video Representations using LSTMs(ICML2015)

LSTM: A Search Space Odyssey

Inferring Algorithmic Patterns with Stack-Augmented Recurrent Nets

A Critical Review of Recurrent Neural Networks for Sequence Learning

Scheduled Sampling for
Sequence Prediction with Recurrent Neural Networks(Winner of MSCOCO image
captioning challenge, 2015)

Visualizing and
Understanding Recurrent Networks(Andrej Karpathy, Justin Johnson, Fei-Fei Li)

Grid Long Short-Term
Memory

Depth-Gated LSTM

Deep Knowledge Tracing

Top-down Tree Long
Short-Term Memory Networks

Alternative structures
for character-level RNNs(INRIA & Facebook AI Research)

Pixel Recurrent Neural
Networks (Google DeepMind)

Long Short-Term
Memory-Networks for Machine Reading

Lipreading with Long
Short-Term Memory

Associative Long
Short-Term Memory

Representation of
linguistic form and function in recurrent neural networks

Architectural
Complexity Measures of Recurrent Neural Networks

Easy-First Dependency
Parsing with Hierarchical Tree LSTMs

Training Input-Output
Recurrent Neural Networks through Spectral Methods

Learn To Execute Programs

Learning to Execute

Neural
Programmer-Interpreters (Google DeepMind)

A
Programmer-Interpreter Neural Network Architecture for Prefrontal Cognitive
Control

Convolutional RNN: an
Enhanced Model for Extracting Features from Sequential Data

Attention Models

Recurrent Models of
Visual Attention
 (Google
DeepMind. NIPS2014)

Recurrent Model of
Visual Attention(Google DeepMind)

Show, Attend and Tell:
Neural Image Caption Generation with Visual Attention

A Neural Attention
Model for Abstractive Sentence Summarization(EMNLP 2015. Facebook AI Research)

Effective Approaches
to Attention-based Neural Machine Translation(EMNLP2015)

Generating Images from
Captions with Attention

Attention and Memory
in Deep Learning and NLP

Survey on the
attention based RNN model and its applications in computer vision

Train RNN

Training Recurrent
Neural Networks (PhD thesis)

Deep learning for
control using augmented Hessian-free optimization


Hierarchical Conflict
Propagation: Sequence Learning in a Recurrent Deep Neural Network

Recurrent Batch
Normalization

Optimizing Performance
of Recurrent Neural Networks on GPUs

Codes

NeuralTalk
(Deprecated): a Python+numpy project for learning Multimodal Recurrent Neural
Networks that describe images with sentences

NeuralTalk2: Efficient
Image Captioning code in Torch, runs on GPU

char-rnn in Blocks

Project:
pycaffe-recurrent

Using neural networks
for password cracking

Recurrent neural
networks for decoding CAPTCHAS

torch-rnn: Efficient,
reusable RNNs and LSTMs for torch

Deploying a model
trained with GPU in Torch into JavaScript, for everyone to use

LSTM implementation on
Caffe

Blog

Survey on
Attention-based Models Applied in NLP

http://yanran.li/peppypapers/2015/10/07/survey-attention-model-1.html

Survey on Advanced
Attention-based Models

http://yanran.li/peppypapers/2015/10/07/survey-attention-model-2.html

Online Representation
Learning in Recurrent Neural Language Models

http://www.marekrei.com/blog/online-representation-learning-in-recurrent-neural-language-models/

Fun with Recurrent
Neural Nets: One More Dive into CNTK and TensorFlow

http://esciencegroup.com/2016/03/04/fun-with-recurrent-neural-nets-one-more-dive-into-cntk-and-tensorflow/

Materials to
understand LSTM

https://medium.com/@shiyan/materials-to-understand-lstm-34387d6454c1#.4mt3bzoau

Understanding LSTM and
its diagrams (
★★★★★)

Persistent RNNs: 30
times faster RNN layers at small mini-batch sizes (Greg Diamos, Baidu Silicon
Valley AI Lab)

http://svail.github.io/persistent_rnns/

All of Recurrent
Neural Networks

https://medium.com/@jianqiangma/all-about-recurrent-neural-networks-9e5ae2936f6e#.q4s02elqg

Resources

Awesome Recurrent
Neural Networks - A curated list of resources dedicated to RNN

Jürgen Schmidhuber’s
page on Recurrent Neural Networks

http://people.idsia.ch/~juergen/rnn.html

Reading and
Questions

Are there any
Recurrent convolutional neural network network implementations out there ?

« Reinforcement LearningSaliency Prediction »

Saliency Prediction

 Published: 09 Oct 2015  Category: deep_learning

This task involves predicting the salient regions of an image given by human eye fixations.

Large-scale optimization of hierarchical features for saliency prediction in natural images

Predicting Eye Fixations using Convolutional Neural Networks

DeepFix: A Fully Convolutional Neural Network for predicting Human Eye Fixations

DeepSaliency: Multi-Task Deep Neural Network Model for Salient Object Detection

SuperCNN: A Superpixelwise Convolutional Neural Network for Salient Object Detection

Shallow and Deep Convolutional Networks for Saliency Prediction

Scene Labeling

 Published: 09 Oct 2015  Category: deep_learning

Papers

Learning hierarchical features for scene labeling

  • intro: “Their approach comprised of densely computing multi-scale CNN features for each pixel and aggregating them over image regions upon which they are classified. However, their methodstill required the post-processing step of generating over-segmented regions, like superpixels, for obtaining the final segmentation result. Additionally, the CNNs used for multi-scale feature learning were not very deep with only three convolution layers.”
  • paper: http://yann.lecun.com/exdb/publis/pdf/farabet-pami-13.pdf

Indoor Semantic Segmentation using depth information

Multi-modal unsupervised feature learning for rgb-d scene labeling

Using neon for Scene Recognition: Mini-Places2

Attend, Infer, Repeat: Fast Scene Understanding with Generative Models

Challenges

Large-scale Scene Understanding Challenge

RNN and LSTM saliency Predection Scene Label的更多相关文章

  1. RNN和LSTM

    一.RNN 全称为Recurrent Neural Network,意为循环神经网络,用于处理序列数据. 序列数据是指在不同时间点上收集到的数据,反映了某一事物.现象等随时间的变化状态或程度.即数据之 ...

  2. RNN、LSTM、Seq2Seq、Attention、Teacher forcing、Skip thought模型总结

    RNN RNN的发源: 单层的神经网络(只有一个细胞,f(wx+b),只有输入,没有输出和hidden state) 多个神经细胞(增加细胞个数和hidden state,hidden是f(wx+b) ...

  3. RNN 与 LSTM 的应用

    之前已经介绍过关于 Recurrent Neural Nnetwork 与 Long Short-Trem Memory 的网络结构与参数求解算法( 递归神经网络(Recurrent Neural N ...

  4. Naive RNN vs LSTM vs GRU

    0 Recurrent Neural Network 1 Naive RNN 2 LSTM peephole Naive RNN vs LSTM 记忆更新部分的操作,Naive RNN为乘法,LSTM ...

  5. TensorFlow之RNN:堆叠RNN、LSTM、GRU及双向LSTM

    RNN(Recurrent Neural Networks,循环神经网络)是一种具有短期记忆能力的神经网络模型,可以处理任意长度的序列,在自然语言处理中的应用非常广泛,比如机器翻译.文本生成.问答系统 ...

  6. 浅谈RNN、LSTM + Kreas实现及应用

    本文主要针对RNN与LSTM的结构及其原理进行详细的介绍,了解什么是RNN,RNN的1对N.N对1的结构,什么是LSTM,以及LSTM中的三门(input.ouput.forget),后续将利用深度学 ...

  7. 3. RNN神经网络-LSTM模型结构

    1. RNN神经网络模型原理 2. RNN神经网络模型的不同结构 3. RNN神经网络-LSTM模型结构 1. 前言 之前我们对RNN模型做了总结.由于RNN也有梯度消失的问题,因此很难处理长序列的数 ...

  8. RNN以及LSTM的介绍和公式梳理

    前言 好久没用正儿八经地写博客了,csdn居然也有了markdown的编辑器了,最近花了不少时间看RNN以及LSTM的论文,在组内『夜校』分享过了,再在这里总结一下发出来吧,按照我讲解的思路,理解RN ...

  9. 深度学习:浅谈RNN、LSTM+Kreas实现与应用

    主要针对RNN与LSTM的结构及其原理进行详细的介绍,了解什么是RNN,RNN的1对N.N对1的结构,什么是LSTM,以及LSTM中的三门(input.ouput.forget),后续将利用深度学习框 ...

随机推荐

  1. 【转载】芯片级拆解51、AVR、MSP430、凌阳61、PIC,5种单片机,多张显微照片

    先秀一张解剖照,放大裁剪,小米1S微距拍摄,800万像素摄像头很给力!今天等待被拆的是5个单片机芯片:(1)凌阳16位单片机SPCE061A ,这是我接触的第一个单片机,最高主频49MHz,32KB的 ...

  2. windows 代理服务器的搭建,提供Android 端访问公网.

    这段时间遇到一个情况,移动的网络收费.但是可以访问学校内部的网络,比如说学校官网图书馆之类了.所以我这里便想到一个方法,用学校内部一个可以访问互联网的主机充当代理服务器(我这里使用自己的电脑,非服务器 ...

  3. Android学习笔记(七)

    活动的启动模式 活动的启动模式一共有四种,分别是standard.singleTop.singleTask和singleInstance, 可以在AndroidManifest.xml中通过给< ...

  4. loadView、viewDidLoad、initWithCoder、initWithNibName、awakeFromNib的用法

    转载,原地址为:http://jianyu996.blog.163.com/blog/static/11211455520131226840879/ 请尊重原创: 1,无论XIB还是代码创建都会调用l ...

  5. PLSQL转义字符

    http://blog.csdn.net/cunxiyuan108/article/details/5800800

  6. Mathematica 计算矩阵的伴随矩阵

    AdjointMatrix[M_] := Module[{Ma, B, n, i, j},    Ma = Minors[M];    B = Ma;    n = Dimensions[M][[1] ...

  7. PHP发送请求头和接收打印请求头

    一.发送请求头 //发送地址 $url = 'http://127.0.0.1/2.php'; //请求头内容 $headers = array( 'Authorization: '.$basic, ...

  8. mysql登录基本语句

    默认密码:root mysql 显示所有的数据库,代码如下: mysql> show databases; mysql> show tables; MySQL显示命令二.显示命令 1.显示 ...

  9. 使用不同用户对Oracle数据库进行异机恢复,失败,错误:Backup file not found in NetBackup catalog

    最近做某数据库恢复演练,数据库版本是10.2.0.4,恢复控制文件一直报错,报错如下,经过反复折腾,原来恢复机上oracle用户不是oracle导致(我的是oraclev4),查看源库oracle用户 ...

  10. [IIS]IIS扫盲(八)

    iis - IIS之FTP服务器 一.建立你的FTP站点  第一个FTP站点(即“默认FTP站点”)的设置方法和更多FTP站点的建立方法请参照前文Web服务器中相关操作执行.需要注意的是,如果你要用一 ...