Keynote: xLSTM: New Architectures for Large Language Models


Monday, November 18, 2024


9:05 am


Today’s LLMs such as ChatGPT show an impressive performance and have the potential to revolutionize our daily life. All these LLMs are based on the Transformer architecture with the Attention mechanism at its core. Due to the quadratic scaling with context length, Attention makes processing of long sequences very expensive. In this talk we present xLSTM, a novel architecture for LLMs that scales only linear in context length while still outperforming Transformers on language modeling.

Ready to attend?

Register now! Join your peers.

Register nowView Agenda
Newsletter Knowledge is everything! Sign up for our newsletter to receive:
  • 10% off your first ticket!
  • insights, interviews, tips, news, and much more about Machine Learning Week Europe
  • price break reminders