Attention is Not All You Need - Exploring Beyond LLMs in Machine Learning
MLCon | Machine Learning Conference via YouTube
Overview
Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Explore the broader machine learning landscape beyond large language models in this 34-minute keynote presentation that challenges the current LLM-centric approach to AI problem-solving. Discover why attention mechanisms and transformer-based models, while powerful, aren't universal solutions for every machine learning challenge. Learn about alternative technical advancements, innovative model architectures, and recent research developments that could be the missing piece for your next project's success. Gain insights into diverse approaches within the machine learning and deep learning ecosystem that extend beyond the current hype around large language models and multimodal systems. Broaden your understanding of when and how to apply different ML techniques effectively, moving beyond the "one-size-fits-all" mentality that often dominates current AI discourse.
Syllabus
Attention is not all you need - Keynote by Christoph Henkelmann
Taught by
MLCon | Machine Learning Conference