Movatterモバイル変換


[0]ホーム

URL:


YH
Uploaded byYao-Chieh Hu
PPTX, PDF1,413 views

RNN & LSTM: Neural Network for Sequential Data

Recurrent neural networks (RNNs) and long short-term memory (LSTM) networks can process sequential data like text and time series data. RNNs have memory and can perform the same task for every element in a sequence, but struggle with long-term dependencies. LSTMs address this issue using memory cells and gates that allow them to learn long-term dependencies. LSTMs have four interacting layers - a forget gate, input gate, cell state, and output gate that allow them to store and access information over long periods of time. RNNs and LSTMs are applied to tasks like language modeling, machine translation, speech recognition, and image caption generation.

In this document
Powered by AI

Overview of RNN and LSTM as neural networks designed for processing sequential data.

Categorization of machine learning: Supervised, Unsupervised, and Reinforcement Learning with various examples.

Introduction to the K-Nearest Neighbor algorithm for classification.

Discussion on how neural networks can reveal hidden relationships within features.

Intro to RNNs, their sequential process capabilities, and applications in various fields.

Method of training RNNs using Backpropagation Through Time (BPTT).

Different types of RNNs including bidirectional and LSTM networks. Detailed explanation of LSTM architecture, including cell state and gated mechanisms.

Stepwise explanation of LSTM gate operations, including forget, input, and output decisions.

Different variations of LSTM including peephole connections, coupled gates, and GRUs.

Discussion on the effectiveness of RNN and LSTM models.

Examples of multiple applications of RNN models.

Concept of Turing-Completeness in relation to RNNs and their capability to simulate programs.

Application of RNNs on non-sequential data through sequential processing.

Resources showcasing interesting applications of RNN and LSTM architectures.

Compilation of references and resources for deeper understanding of RNNs, LSTMs, and ML.

Embed presentation

Downloaded 108 times
RNN & LSTMNeural Network for Sequential Data- Jeff Hu -
Machine Learning Categories• Supervised• Unsupervised• Reinforcement Learning
Supervised Machine Learning• Training Set: Inputs + Outputs• Learn a link between the inputs and the outputs• Linear and logistic regression• Support vector machine• K-nearest neighbors (k-NN)• Naive Bayes• Neural network• Gradient boosting• Classification trees and random forest
Unsupervised Machine Learning• Training Set: Inputs• Cluster the inputs• K-means• Hierarchical clustering• Mixture models• PCA• ICA• Auto-encoder
Reinforcement Learning• Training Set: N/A• Find the best way to earn the greatest reward• Utility learning• Q-learning
K-Nearest Neighbor
Neural Network > Machine Learning ?• Consider hidden relationships between features!
Recurrent Neural Network (RNN)
Benefits• Deal with sequential information• Perform the same task for every element of a sequence• Has memory• Can be unrolled like a chain
Application• Language Modeling and Generating Text• Machine Translation• Speech Recognition• Generating Image Descriptions
Training• Backpropagation Through Time (BPTT)
Variations• Bidirectional RNNs• Deep (Bidirectional) RNNs• LSTM networks
Long Short Term Memory Network(LSTM)
Memory Problem of RNN• Sometimes we need more context• RNN is unable to connect the information further in the past
Benefits of LSTM• Can learn long-term dependencies
Difference between RNN & LSTM• RNN: single layer (tanh)• LSTM: four interactive layers
Cell state• The conveyor belt
Gates (3 in total for LSTM)• A way that let information through• E.g. A sigmoid neural net layer & a pointwise multiplication operation
Optional Math – Sigmoid function
Step 1: Forget Gate Layer• Decide what info to throw away• Look at h[t-1] and x[t] and output a number 0~1 to decide how much cell state to keep C[t-1]• E.g. When see a new subject, we want to forget the gender of the old subject
Step 2: Input Gate Layer• Decide what info to add• A sigmoid: decide which value to update• A tanh layer: create a new candidate value C~[t]• E.g. add a new gender of the new subject
Step 3: Combine step 1 & 2• Combine step 1 & 2• Multiply the old state by f[t]: to forget the things• Add i[t] * C~[t] : to add new candidate value (scaled)
Step 4: Filter/output the Cell state• Decide what to output• sigmoid: decide which part to output• tanh: push the value to be between -1 ~ 1• Multiply them to only output the part we decided to• E.g. output a info related to a Verb• E.g. output whether the subject it singular or plural
Step 4: Filter/output the Cell state• Decide what to output
Variants on LSTM (1)• Peephole: let the gate layer look at the cell state (entire/ partial)
Variants on LSTM (2)• Coupled forgot and input gates: Not deciding separately f[t] * C[t-1] + (1-f[t]) * C~[t]
Variants on LSTM (3)• Gated Recurrent Unit (GRU): combine the forget and input layer into a single “update gate” merge the cell state and the hidden state simpler and popular
RNN / LSTM Effectiveness
Multiple types of RNN use cases
Turing-Complete• Running a fixed program with certain inputs and some internal variables (can simulatearbitrary programs)• Andrej Karpathy (Ph.D. @ Stanford):
Non-sequential data• Though the data is not in form of sequences, we can still use RNN by processit sequentially.
Some cool RNN/LSTM applications• http://karpathy.github.io/2015/05/21/rnn-effectiveness/
Great references• [1] RNN: http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/?subscribe=success#blog_subscription-2• [2] LSTM: http://colah.github.io/posts/2015-08-Understanding-LSTMs/• [3] RNN Effectiveness: http://karpathy.github.io/2015/05/21/rnn-effectiveness/• [4] Backpropagation: http://cs231n.github.io/optimization-2/#backprop• [5] ML categories: http://enhancedatascience.com/2017/07/19/machine-learning-explained-supervised-learning-unsupervised-learning-and-reinforcement-learning/

Recommended

PPTX
RNN-LSTM.pptx
PDF
Long Short Term Memory
 
PDF
LSTM
PPTX
PDF
LSTM Tutorial
PDF
Recurrent Neural Networks. Part 1: Theory
PDF
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
PDF
Backpropagation in RNN and LSTM
PDF
Recurrent neural networks rnn
PDF
Introduction to Recurrent Neural Network
PDF
Recurrent Neural Networks, LSTM and GRU
 
PDF
Rnn and lstm
PPTX
PPTX
Long Short Term Memory LSTM
PPTX
Introduction to CNN
PPTX
04 Multi-layer Feedforward Networks
PPTX
Deep Learning - RNN and CNN
PPTX
Recurrent Neural Network
PPT
Perceptron
PPTX
Activation functions
PDF
LSTM Basics
PDF
Deep Learning: Recurrent Neural Network (Chapter 10)
PPTX
Feedforward neural network
PPTX
Convolutional neural network
PPTX
Recurrent neural network
PPS
Neural Networks
PDF
RNN and its applications
PDF
Introduction to Recurrent Neural Network
 
PDF
Recurrent Neural Networks
PPTX
Introduction to deep learning

More Related Content

PPTX
RNN-LSTM.pptx
PDF
Long Short Term Memory
 
PDF
LSTM
PPTX
PDF
LSTM Tutorial
PDF
Recurrent Neural Networks. Part 1: Theory
PDF
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
PDF
Backpropagation in RNN and LSTM
RNN-LSTM.pptx
Long Short Term Memory
 
LSTM
LSTM Tutorial
Recurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
Backpropagation in RNN and LSTM

What's hot

PDF
Recurrent neural networks rnn
PDF
Introduction to Recurrent Neural Network
PDF
Recurrent Neural Networks, LSTM and GRU
 
PDF
Rnn and lstm
PPTX
PPTX
Long Short Term Memory LSTM
PPTX
Introduction to CNN
PPTX
04 Multi-layer Feedforward Networks
PPTX
Deep Learning - RNN and CNN
PPTX
Recurrent Neural Network
PPT
Perceptron
PPTX
Activation functions
PDF
LSTM Basics
PDF
Deep Learning: Recurrent Neural Network (Chapter 10)
PPTX
Feedforward neural network
PPTX
Convolutional neural network
PPTX
Recurrent neural network
PPS
Neural Networks
PDF
RNN and its applications
PDF
Introduction to Recurrent Neural Network
 
Recurrent neural networks rnn
Introduction to Recurrent Neural Network
Recurrent Neural Networks, LSTM and GRU
 
Rnn and lstm
Long Short Term Memory LSTM
Introduction to CNN
04 Multi-layer Feedforward Networks
Deep Learning - RNN and CNN
Recurrent Neural Network
Perceptron
Activation functions
LSTM Basics
Deep Learning: Recurrent Neural Network (Chapter 10)
Feedforward neural network
Convolutional neural network
Recurrent neural network
Neural Networks
RNN and its applications
Introduction to Recurrent Neural Network
 

Similar to RNN & LSTM: Neural Network for Sequential Data

PDF
Recurrent Neural Networks
PPTX
Introduction to deep learning
PDF
Sequence Modelling with Deep Learning
PDF
Deep Learning for Computer Vision: Recurrent Neural Networks (UPC 2016)
PPTX
RNN and LSTM model description and working advantages and disadvantages
PDF
Video Analysis with Recurrent Neural Networks (Master Computer Vision Barcelo...
PDF
Recurrent and Recursive Nets (part 2)
PPTX
recurrent_neural_networks_april_2020.pptx
PPTX
10.0 SequenceModeling-merged-compressed_edited.pptx
PPT
14889574 dl ml RNN Deeplearning MMMm.ppt
PDF
Recurrent Neural Networks RNN - Xavier Giro - UPC TelecomBCN Barcelona 2020
PPTX
DEEP LEARNING -Recurrent Neural Networks
PDF
rnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
PDF
Recurrent Neural Networks I (D2L2 Deep Learning for Speech and Language UPC 2...
PDF
Concepts of Temporal CNN, Recurrent Neural Network, Attention
PPT
Recurrent neural network power point presentation
PPT
Recurrent neural network power point presentation
PDF
Recurrent Neural Networks (D2L8 Insight@DCU Machine Learning Workshop 2017)
PDF
Recurrent Neural Networks A Deep Dive in 2025.pdf
PDF
Recurrent Neural Networks (D2L2 2017 UPC Deep Learning for Computer Vision)
Recurrent Neural Networks
Introduction to deep learning
Sequence Modelling with Deep Learning
Deep Learning for Computer Vision: Recurrent Neural Networks (UPC 2016)
RNN and LSTM model description and working advantages and disadvantages
Video Analysis with Recurrent Neural Networks (Master Computer Vision Barcelo...
Recurrent and Recursive Nets (part 2)
recurrent_neural_networks_april_2020.pptx
10.0 SequenceModeling-merged-compressed_edited.pptx
14889574 dl ml RNN Deeplearning MMMm.ppt
Recurrent Neural Networks RNN - Xavier Giro - UPC TelecomBCN Barcelona 2020
DEEP LEARNING -Recurrent Neural Networks
rnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnn
Recurrent Neural Networks I (D2L2 Deep Learning for Speech and Language UPC 2...
Concepts of Temporal CNN, Recurrent Neural Network, Attention
Recurrent neural network power point presentation
Recurrent neural network power point presentation
Recurrent Neural Networks (D2L8 Insight@DCU Machine Learning Workshop 2017)
Recurrent Neural Networks A Deep Dive in 2025.pdf
Recurrent Neural Networks (D2L2 2017 UPC Deep Learning for Computer Vision)

Recently uploaded

PPTX
Chapter-2 cyber security data sources.pptx
PDF
Interpretability and Explainability Module 4.pdf
PDF
Comparing Versions in MySQL Key Differences and Insights.pdf
PDF
MATERI HUAWEI AI ALL yang di upload langsung oleh dosen perguruan tinggi.pdf
PDF
IIT JEE 2026 Complete Study Planner.pdfg
PDF
From Microsoft SCOM to Dashboards | Grafana, SquaredUp, Power BI, Azure Workb...
PDF
Power BI Template Design Instructions.pdf
PDF
Approaches to Ransomware Protection with NetApp Monitoring
PDF
MariaDB Monitoring for Enhancing Performance, Availability, and Security
PDF
System Center 2025 Migration: Preparing for a Smooth Transition
PPTX
IR spectrophotometer is an analytical instrument
PPT
Transportation Management Plan (Effectiveness Study).ppt
PPT
atomic structure2.ppt qwgdwgdwidgqwuidg wdhwgdwgwydg qwgdwydgwydg87
PPTX
Correlation-Regression analysis -16.11.25.pptx
PPT
fgfdgdgdgdfgerst4t4tqwweqe3eawdsfsdsdsdd
PPT
(7) Qualitative Research Methodology (overview).ppt
PPTX
Understanding purpose, audience MSDA .pptx
PPTX
Chapter 4, Blocksmbjkhnl;m',\;lkjhcgjvbkln;m.pptx
PPTX
2B.Carbon-Neutral Technologies and Negative Emission Strategies for Net-Zero ...
PPTX
PPT PVA Training NDLM portal various data .pptx
Chapter-2 cyber security data sources.pptx
Interpretability and Explainability Module 4.pdf
Comparing Versions in MySQL Key Differences and Insights.pdf
MATERI HUAWEI AI ALL yang di upload langsung oleh dosen perguruan tinggi.pdf
IIT JEE 2026 Complete Study Planner.pdfg
From Microsoft SCOM to Dashboards | Grafana, SquaredUp, Power BI, Azure Workb...
Power BI Template Design Instructions.pdf
Approaches to Ransomware Protection with NetApp Monitoring
MariaDB Monitoring for Enhancing Performance, Availability, and Security
System Center 2025 Migration: Preparing for a Smooth Transition
IR spectrophotometer is an analytical instrument
Transportation Management Plan (Effectiveness Study).ppt
atomic structure2.ppt qwgdwgdwidgqwuidg wdhwgdwgwydg qwgdwydgwydg87
Correlation-Regression analysis -16.11.25.pptx
fgfdgdgdgdfgerst4t4tqwweqe3eawdsfsdsdsdd
(7) Qualitative Research Methodology (overview).ppt
Understanding purpose, audience MSDA .pptx
Chapter 4, Blocksmbjkhnl;m',\;lkjhcgjvbkln;m.pptx
2B.Carbon-Neutral Technologies and Negative Emission Strategies for Net-Zero ...
PPT PVA Training NDLM portal various data .pptx

RNN & LSTM: Neural Network for Sequential Data

  • 1.
    RNN & LSTMNeuralNetwork for Sequential Data- Jeff Hu -
  • 2.
    Machine Learning Categories•Supervised• Unsupervised• Reinforcement Learning
  • 3.
    Supervised Machine Learning•Training Set: Inputs + Outputs• Learn a link between the inputs and the outputs• Linear and logistic regression• Support vector machine• K-nearest neighbors (k-NN)• Naive Bayes• Neural network• Gradient boosting• Classification trees and random forest
  • 4.
    Unsupervised Machine Learning•Training Set: Inputs• Cluster the inputs• K-means• Hierarchical clustering• Mixture models• PCA• ICA• Auto-encoder
  • 5.
    Reinforcement Learning• TrainingSet: N/A• Find the best way to earn the greatest reward• Utility learning• Q-learning
  • 6.
  • 7.
    Neural Network >Machine Learning ?• Consider hidden relationships between features!
  • 8.
  • 10.
    Benefits• Deal withsequential information• Perform the same task for every element of a sequence• Has memory• Can be unrolled like a chain
  • 11.
    Application• Language Modelingand Generating Text• Machine Translation• Speech Recognition• Generating Image Descriptions
  • 12.
  • 13.
    Variations• Bidirectional RNNs•Deep (Bidirectional) RNNs• LSTM networks
  • 14.
    Long Short TermMemory Network(LSTM)
  • 16.
    Memory Problem ofRNN• Sometimes we need more context• RNN is unable to connect the information further in the past
  • 17.
    Benefits of LSTM•Can learn long-term dependencies
  • 18.
    Difference between RNN& LSTM• RNN: single layer (tanh)• LSTM: four interactive layers
  • 19.
    Cell state• Theconveyor belt
  • 20.
    Gates (3 intotal for LSTM)• A way that let information through• E.g. A sigmoid neural net layer & a pointwise multiplication operation
  • 21.
    Optional Math –Sigmoid function
  • 22.
    Step 1: ForgetGate Layer• Decide what info to throw away• Look at h[t-1] and x[t] and output a number 0~1 to decide how much cell state to keep C[t-1]• E.g. When see a new subject, we want to forget the gender of the old subject
  • 23.
    Step 2: InputGate Layer• Decide what info to add• A sigmoid: decide which value to update• A tanh layer: create a new candidate value C~[t]• E.g. add a new gender of the new subject
  • 24.
    Step 3: Combinestep 1 & 2• Combine step 1 & 2• Multiply the old state by f[t]: to forget the things• Add i[t] * C~[t] : to add new candidate value (scaled)
  • 25.
    Step 4: Filter/outputthe Cell state• Decide what to output• sigmoid: decide which part to output• tanh: push the value to be between -1 ~ 1• Multiply them to only output the part we decided to• E.g. output a info related to a Verb• E.g. output whether the subject it singular or plural
  • 26.
    Step 4: Filter/outputthe Cell state• Decide what to output
  • 27.
    Variants on LSTM(1)• Peephole: let the gate layer look at the cell state (entire/ partial)
  • 28.
    Variants on LSTM(2)• Coupled forgot and input gates: Not deciding separately f[t] * C[t-1] + (1-f[t]) * C~[t]
  • 29.
    Variants on LSTM(3)• Gated Recurrent Unit (GRU): combine the forget and input layer into a single “update gate” merge the cell state and the hidden state simpler and popular
  • 30.
    RNN / LSTMEffectiveness
  • 31.
    Multiple types ofRNN use cases
  • 32.
    Turing-Complete• Running afixed program with certain inputs and some internal variables (can simulatearbitrary programs)• Andrej Karpathy (Ph.D. @ Stanford):
  • 33.
    Non-sequential data• Thoughthe data is not in form of sequences, we can still use RNN by processit sequentially.
  • 34.
    Some cool RNN/LSTMapplications• http://karpathy.github.io/2015/05/21/rnn-effectiveness/
  • 35.
    Great references• [1]RNN: http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-1-introduction-to-rnns/?subscribe=success#blog_subscription-2• [2] LSTM: http://colah.github.io/posts/2015-08-Understanding-LSTMs/• [3] RNN Effectiveness: http://karpathy.github.io/2015/05/21/rnn-effectiveness/• [4] Backpropagation: http://cs231n.github.io/optimization-2/#backprop• [5] ML categories: http://enhancedatascience.com/2017/07/19/machine-learning-explained-supervised-learning-unsupervised-learning-and-reinforcement-learning/

[8]ページ先頭

©2009-2025 Movatter.jp