Beyond Transformers - New Sequence Processing Architectures - Day 6 Morning
Center for Language & Speech Processing(CLSP), JHU via YouTube
The Fastest Way to Become a Backend Developer Online
MIT Sloan: Lead AI Adoption Across Your Organization — Not Just Pilot It
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore cutting-edge sequence processing architectures that go beyond traditional transformer models in this comprehensive 3-hour tutorial from JSALT 2025. Learn from leading experts Jan Chorowski from Pathway, Marek Adamczyk from Warsaw University, and Adrian Łańcucki from NVIDIA as they present the latest developments in neural sequence modeling. Discover innovative architectural approaches that address the limitations of transformers and examine emerging paradigms for processing sequential data more efficiently. Gain insights into state-of-the-art methods that are shaping the future of natural language processing, speech recognition, and other sequence-based machine learning applications through detailed slide presentations and expert analysis from the Johns Hopkins University Center for Language & Speech Processing.
Syllabus
[slides] Day 6 morning - JSALT 2025 - Beyond transformers - new sequence processing architectures
Taught by
Center for Language & Speech Processing(CLSP), JHU