Embed presentation





















![Step 1: Forget Gate Layer• Decide what info to throw away• Look at h[t-1] and x[t] and output a number 0~1 to decide how much cell state to keep C[t-1]• E.g. When see a new subject, we want to forget the gender of the old subject](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-22-2048.jpg&f=jpg&w=240)
![Step 2: Input Gate Layer• Decide what info to add• A sigmoid: decide which value to update• A tanh layer: create a new candidate value C~[t]• E.g. add a new gender of the new subject](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-23-2048.jpg&f=jpg&w=240)
![Step 3: Combine step 1 & 2• Combine step 1 & 2• Multiply the old state by f[t]: to forget the things• Add i[t] * C~[t] : to add new candidate value (scaled)](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-24-2048.jpg&f=jpg&w=240)



![Variants on LSTM (2)• Coupled forgot and input gates: Not deciding separately f[t] * C[t-1] + (1-f[t]) * C~[t]](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-28-2048.jpg&f=jpg&w=240)






![Great references• [1] RNN: http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/?subscribe=success#blog_subscription-2• [2] LSTM: http://colah.github.io/posts/2015-08-Understanding-LSTMs/• [3] RNN Effectiveness: http://karpathy.github.io/2015/05/21/rnn-effectiveness/• [4] Backpropagation: http://cs231n.github.io/optimization-2/#backprop• [5] ML categories: http://enhancedatascience.com/2017/07/19/machine-learning-explained-supervised-learning-unsupervised-learning-and-reinforcement-learning/](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-35-2048.jpg&f=jpg&w=240)
Recurrent neural networks (RNNs) and long short-term memory (LSTM) networks can process sequential data like text and time series data. RNNs have memory and can perform the same task for every element in a sequence, but struggle with long-term dependencies. LSTMs address this issue using memory cells and gates that allow them to learn long-term dependencies. LSTMs have four interacting layers - a forget gate, input gate, cell state, and output gate that allow them to store and access information over long periods of time. RNNs and LSTMs are applied to tasks like language modeling, machine translation, speech recognition, and image caption generation.
Overview of RNN and LSTM as neural networks designed for processing sequential data.
Categorization of machine learning: Supervised, Unsupervised, and Reinforcement Learning with various examples.
Introduction to the K-Nearest Neighbor algorithm for classification.
Discussion on how neural networks can reveal hidden relationships within features.
Intro to RNNs, their sequential process capabilities, and applications in various fields.
Method of training RNNs using Backpropagation Through Time (BPTT).
Different types of RNNs including bidirectional and LSTM networks. Detailed explanation of LSTM architecture, including cell state and gated mechanisms.
Stepwise explanation of LSTM gate operations, including forget, input, and output decisions.
Different variations of LSTM including peephole connections, coupled gates, and GRUs.
Discussion on the effectiveness of RNN and LSTM models.
Examples of multiple applications of RNN models.
Concept of Turing-Completeness in relation to RNNs and their capability to simulate programs.
Application of RNNs on non-sequential data through sequential processing.
Resources showcasing interesting applications of RNN and LSTM architectures.
Compilation of references and resources for deeper understanding of RNNs, LSTMs, and ML.





















![Step 1: Forget Gate Layer• Decide what info to throw away• Look at h[t-1] and x[t] and output a number 0~1 to decide how much cell state to keep C[t-1]• E.g. When see a new subject, we want to forget the gender of the old subject](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-22-2048.jpg&f=jpg&w=240)
![Step 2: Input Gate Layer• Decide what info to add• A sigmoid: decide which value to update• A tanh layer: create a new candidate value C~[t]• E.g. add a new gender of the new subject](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-23-2048.jpg&f=jpg&w=240)
![Step 3: Combine step 1 & 2• Combine step 1 & 2• Multiply the old state by f[t]: to forget the things• Add i[t] * C~[t] : to add new candidate value (scaled)](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-24-2048.jpg&f=jpg&w=240)



![Variants on LSTM (2)• Coupled forgot and input gates: Not deciding separately f[t] * C[t-1] + (1-f[t]) * C~[t]](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-28-2048.jpg&f=jpg&w=240)






![Great references• [1] RNN: http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/?subscribe=success#blog_subscription-2• [2] LSTM: http://colah.github.io/posts/2015-08-Understanding-LSTMs/• [3] RNN Effectiveness: http://karpathy.github.io/2015/05/21/rnn-effectiveness/• [4] Backpropagation: http://cs231n.github.io/optimization-2/#backprop• [5] ML categories: http://enhancedatascience.com/2017/07/19/machine-learning-explained-supervised-learning-unsupervised-learning-and-reinforcement-learning/](/image.pl?url=https%3a%2f%2fimage.slidesharecdn.com%2fjpmrnnlstmsmalltalk-171208093736%2f75%2fRNN-LSTM-Neural-Network-for-Sequential-Data-35-2048.jpg&f=jpg&w=240)