Author: Bc. Daniel Pištek
Supervisor (Školitel): prof. Ing. Igor Farkaš, Dr.
University: Comenius University
Year: 2025
The introduction of Transformers in 2017 reshaped the landscape of deep learning. Originally proposed for sequence modelling, Transformers have since achieved widespread success across various domains. However, the scalability limitations of Transformers—particularly with respect to sequence length—have sparked renewed interest in novel recurrent models that are parallelizable during training, offer comparable performance, and scale more effectively.
1. Study and implement recent models of recurrent neural networks such as minimized versions of LSTM and GRU. 2. Compare the performance of these models of selected benchmark sequential tasks from the perspective of training time and complexity (number of trainable parameters, scalability). 3. Analyze the performance of these models using the methods of explainable AI.