Inhalt

[ 993MLPELSTU25 ] UE (*)LSTM & Transformers

Versionsauswahl
(*) Leider ist diese Information in Deutsch nicht verfügbar.
Workload Ausbildungslevel Studienfachbereich VerantwortlicheR Semesterstunden Anbietende Uni
1,5 ECTS M1 - Master 1. Jahr Artificial Intelligence Sepp Hochreiter 1,5 SSt Johannes Kepler Universität Linz
Detailinformationen
Quellcurriculum Masterstudium Artificial Intelligence 2025W
Lernergebnisse
Kompetenzen
(*)Students have hands-on experience in implementing Recurrent Neural Networks (RNNs), Long Short-Term Memory (LSTM) models, and Transformers using Python and the PyTorch framework. They have the practical skills to apply these models to real-world sequence prediction tasks, such as time-series forecasting, natural language processing, and other advanced sequential data applications.
Fertigkeiten Kenntnisse
(*)
  • Implementing RNNs for Sequential Data Prediction (k3)

Students can design, train, and optimize RNNs to predict sequential data, handling tasks like time-series forecasting and text generation.

  • Training LSTM Models to Handle Long-Term Dependencies (k4)

Students are able to implement and train LSTM networks in PyTorch to effectively solve sequence prediction tasks that require long-term memory retention, overcoming the vanishing gradient problem.

  • Working with Attention Mechanisms and Memory Models (k5)

Students can implement attention mechanisms and memory models such as Neural Turing Machines, enhancing model performance by allowing selective focus on relevant input data.

  • Building and Training Autoregressive and Transformer, Mamba, and xLSTM Models (k5)

Students are capable of implementing autoregressive sequence models like PixelRNN and advanced architectures like Transformers, applying these models to tasks such as language modeling and text generation

  • Applying Sequence Models to NLP Applications (k5)

Students can use RNNs, LSTMs, and Transformers for natural language processing tasks, implementing practical applications such as next character prediction, language modeling, and sentiment analysis.

(*)Students possess knowledge of sequence prediction fundamentals, covering RNNs, LSTMs, and the challenges associated with vanishing gradients. They also know of advanced techniques like attention models, memory models, Transformers, Mamba, and xLSTM, and how to apply them to real-world tasks such as time-series analysis and natural language processing.
Beurteilungskriterien (*)Assignments during the semester plus final exam
Lehrmethoden (*)Students are given assignments in 1-2 week intervals. Homework must be handed in. Results are to be presented and discussed in the course.
Abhaltungssprache Englisch
Lehrinhalte wechselnd? Nein
Frühere Varianten Decken ebenfalls die Anforderungen des Curriculums ab (von - bis)
993MLPELRNU19: UE LSTM and Recurrent Neural Nets (2019W-2025S)
Präsenzlehrveranstaltung
Teilungsziffer 35
Zuteilungsverfahren Direktzuteilung