A New Generation of Foundation Models Based on xLSTM

Activity: Talk or presentationInvited talkscience-to-science

Description

Foundation models, such as GPT and vision transformers, are almost all built on the famous Transformer architecture. Since about 2018 these transformers have replaced recurrent neural networks (RNNs) in natural language processing, computer vision, but also in other application areas. However, the computational costs of Transformers scale quadratically with sequence or context length which is one of the main reasons for the huge computational costs of AI systems across the world. xLSTM is a modern RNN, which is parallelizable similar to the Transformer, but its computational costs only scale linearly. Thus, xLSTM-based foundation models have the capacity to overtake Transformers as main components of AI systems. In this talk, we provide an overview of the development of AI systems from RNNs, to Transformers, LLMs and foundation models, and to the xLSTM architecture, its core component and its applications.
Period19 May 2025
Event titleCAIML Symposium 2025
Event typeOther
Conference number4
LocationWien, AustriaShow on map
Degree of RecognitionInternational

Fields of science

  • 101019 Stochastics
  • 102003 Image processing
  • 103029 Statistical physics
  • 101018 Statistics
  • 101017 Game theory
  • 102001 Artificial intelligence
  • 202017 Embedded systems
  • 101016 Optimisation
  • 101015 Operations research
  • 101014 Numerical mathematics
  • 101029 Mathematical statistics
  • 101028 Mathematical modelling
  • 101026 Time series analysis
  • 101024 Probability theory
  • 102032 Computational intelligence
  • 102004 Bioinformatics
  • 102013 Human-computer interaction
  • 101027 Dynamical systems
  • 305907 Medical statistics
  • 101004 Biomathematics
  • 305905 Medical informatics
  • 101031 Approximation theory
  • 102033 Data mining
  • 102 Computer Sciences
  • 305901 Computer-aided diagnosis and therapy
  • 102019 Machine learning
  • 106007 Biostatistics
  • 102018 Artificial neural networks
  • 106005 Bioinformatics
  • 202037 Signal processing
  • 202036 Sensor systems
  • 202035 Robotics

JKU Focus areas

  • Digital Transformation