[深度学习论文笔记][Recurrent Neural Networks] Visualizing and Understanding Recurrent Networks
来源:互联网 发布:龙卷风网络收音机 mac 编辑:程序博客网 时间:2024/05/29 13:12
1 RNN
RNN has formWhere W varies between layers but is shared through time. ⃗ x is the input from the layer below.
It was observed that the back-propagation dynamics caused the gradients in an RNN to either vanish or explode.
2 LSTM
The exploding gradient concern can be alleviated with a heuristic of clipping the gradients, and LSTMs were designed to mitigate the vanishing gradient problem. In addition to a ⃗ LSTMs also maintain a memory vector ⃗ c . At each time step the hidden state vector h, LSTM can choose to read from, write to, or reset the cell using explicit gating mechanisms.The three vectors ⃗ i , ⃗ f , o ⃗ are thought of as binary gates that control whether each memory cell is updated, whether it is reset to zero, and whether its local state is revealed in
the hidden vector, respectively. The activations of these gates are based on the sigmoid function and hence allowed to range smoothly between zero and one to keep the model differentiable.
The vector g ⃗ ranges between -1 and 1 and is used to additively modify the memory contents. This additive interaction is a critical feature of the LSTM’s design, because during backpropagation a sum operation merely distributes gradients. This allows gradients on the memory cells ⃗ c to flow backwards through time uninterrupted for long time periods, or at least until the flow is disrupted with the multiplicative interaction of an active forget gate. See Fig. for an explanation.
3 GRU
This is a simple alternative to LSTM.The Gated Recurrent Unit (GRU) has the interpretation of computing a candidate hidden vector h ⃗ ̃ t and then smoothly interpolating towards it gated by ⃗ z .
- [深度学习论文笔记][Recurrent Neural Networks] Visualizing and Understanding Recurrent Networks
- [深度学习论文笔记][Visualizing] Visualizing and Understanding Convolutional Networks
- 论文笔记:Depth-Gated Recurrent Neural Networks
- 深度学习(2):Recurrent Neural Networks
- [深度学习论文笔记][Semantic Segmentation] Recurrent Convolutional Neural Networks for Scene Labeling
- [深度学习论文笔记][Visualizing] Understanding Neural Networks Through Deep Visualization
- CS231n学习笔记--10.Recurrent Neural Networks
- 论文笔记 Visualizing and understanding convolutional networks
- Visualizing and Understanding Convolutional Networks论文笔记
- [深度学习]Visualizing and Understanding Convolutional Networks阅读笔记
- speech recognition with deep recurrent neural networks-论文笔记
- 论文笔记:session-based recommendations with recurrent neural networks
- Gated Recurrent Neural Networks
- Recurrent Neural Networks Tutorial
- Recurrent Neural Networks - collections
- Recurrent Neural Networks regularization
- Recurrent Neural Networks
- Recurrent Neural Networks
- 【Android Studio安装配置教程】一、安装
- const限定形参的重载
- text控件监听键盘弹出和关闭示例代码
- HDU-2255(KM算法)
- Avatar Badge Button checkBox chip
- [深度学习论文笔记][Recurrent Neural Networks] Visualizing and Understanding Recurrent Networks
- Java篇-File类之常用操作
- H5的新特性
- Android 设计模式实战笔记 工厂方法模式
- 自定义IP里面的各个总线接口的定义
- CoreImage简单实用
- 关于"Mobile开发客户端的屏幕尺寸大小"的解答
- USB协通讯议--深入理解
- android studio 报错:Error:warning: Ignoring InnerClasses attribute