Lecture: Recurrent Neural Networks

Lecture: Recurrent Neural Networks

Administrative information


Title Recurrent Neural Networks
Duration 60 min
Module B
Lesson Type Lecture
Focus Technical - Deep Learning
Topic

Multiple (Recurrent Neural Networks (RNN), Backpropagation Through Time (BPTT), Long Short-Term Memory (LSTM))

 

Keywords


Recurrent Neural Networks (RNN), Backpropagation Trough Time (BPTT), Long Short-Term Memory (LSTM),

 

Learning Goals


  • Learning the fundamentals of Recurrent Neural Networks (RNN), Backpropagation Trough Time (BPTT) and Long Short-Term Memory (LSTM)

 

Expected Preparation


Learning Events to be Completed Before

Obligatory for Students

Optional for Students

None.

References and background for students:

None.

Recommended for Teachers

None.

Lesson Materials


 

The materials of this learning event are available under CC BY-NC-SA 4.0.

 

Instructions for Teachers


In the beginning a general overview of sequential data is recommended. Here, you can discuss the main challanges of modeling sequential data (including receptive field, multidimensionality and multiscale nature).

Later we introduce RNNs first, with the basic principles and show that training RNNs after unfolding is very simlar to training MLPs, but the input, recurrent and output weight matrices are shared.

We introduce BPTT (Backpropagation through time) and the truncated version.

Next, we discuss how the vanishing gradient problem makes RNNs impractical.

In order to solve the vanishing gradient, we introduce the LSTM architecture, which has an inner memory part (also reffered to as memory cell), which does not have any activatoin - so vanishing gradient does not occure here. Please make it really clear, that the gating mechanism is truly controlled by the data.

At the final part of the lecture, we show that LSTM (and RNN, indeed) layers can be stacked onto each other with one or two directions. (uni- and bidirectional networks)

Outline

  • Overview of sequential data
  • Recurrent neural networks basics
  • Backpropagation through time
  • Vanishing gradient
  • Long Short-Term Memory
  • Stacking RNN/LSTM layers
Time schedule
Duration (Min) Description
10 Sequential data introduction
15 Recurrent neural networks and Backpropagation through time
5 Vanishing gradients in RNNs
20 LSTMs
5 Stacking RNN/LSTM layers
5 Conclusions

More information

Click here for an overview of all lesson plans of the master human centred AI

Please visit the home page of the consortium HCAIM

Acknowledgements

The Human-Centered AI Masters programme was co-financed by the Connecting Europe Facility of the European Union Under Grant №CEF-TC-2020-1 Digital Skills 2020-EU-IA-0068.

The materials of this learning event are available under CC BY-NC-SA 4.0

 

The HCAIM consortium consists of three excellence centres, three SMEs and four Universities

HCAIM Consortium

Colofon

Het arrangement Lecture: Recurrent Neural Networks is gemaakt met Wikiwijs van Kennisnet. Wikiwijs is hét onderwijsplatform waar je leermiddelen zoekt, maakt en deelt.

Laatst gewijzigd
2024-05-15 11:16:30
Licentie

Dit lesmateriaal is gepubliceerd onder de Creative Commons Naamsvermelding-GelijkDelen 4.0 Internationale licentie. Dit houdt in dat je onder de voorwaarde van naamsvermelding en publicatie onder dezelfde licentie vrij bent om:

  • het werk te delen - te kopiëren, te verspreiden en door te geven via elk medium of bestandsformaat
  • het werk te bewerken - te remixen, te veranderen en afgeleide werken te maken
  • voor alle doeleinden, inclusief commerciële doeleinden.

Meer informatie over de CC Naamsvermelding-GelijkDelen 4.0 Internationale licentie.

Aanvullende informatie over dit lesmateriaal

Van dit lesmateriaal is de volgende aanvullende informatie beschikbaar:

Toelichting
.
Eindgebruiker
leerling/student
Moeilijkheidsgraad
gemiddeld
Studiebelasting
4 uur en 0 minuten

Gebruikte Wikiwijs Arrangementen

HCAIM Consortium. (z.d.).

Acknowledgement

https://maken.wikiwijs.nl/198386/Acknowledgement

HCAIM Consortium. (z.d.).

Lecture: Hyperparameter tuning

https://maken.wikiwijs.nl/200297/Lecture__Hyperparameter_tuning

close
Colofon
gemaakt met Wikiwijs van kennisnet-logo
open