Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Coursera

Attention Mechanisms and Transformer Models Course

via Coursera

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
This deep learning course provides a comprehensive introduction to attention mechanisms and transformer models the foundation of modern GenAI systems. Begin by exploring the shift from traditional neural networks to attention-based architectures. Understand how additive, multiplicative, and self-attention improve model accuracy in NLP and vision tasks. Dive into the mechanics of self-attention and how it powers models like GPT and BERT. Progress to mastering multi-head attention and transformer components, and explore their role in advanced text and image generation. Gain real-world insights through demos featuring GPT, DALL·E, LLaMa, and BERT. To be successful in this course, you should have a basic understanding of neural networks, machine learning concepts, and Python programming. By the end of this course, you’ll be able to: - Explain how attention mechanisms enhance deep learning models - Implement and apply self-attention and multi-head attention - Understand transformer architecture and real-world use cases - Analyze leading GenAI models across NLP and image generation Ideal for AI developers, ML engineers, and data scientists.

Syllabus

  • Introduction to Attention Mechanism and Self-Attention
    • Explore the power of attention mechanisms in modern deep learning. Compare traditional neural architectures with attention-based models to see how additive, multiplicative, and self-attention boost accuracy in NLP and vision tasks. Grasp the core math and flow of self-attention, the engine behind Transformer giants like GPT and BERT and build a solid base for advanced AI development.
  • Multi-Head Attention, Transformers, and Their Applications
    • Master multi-head attention and transformer models in this advanced module. Learn how multi-head attention improves context understanding and powers leading transformer architectures. Explore transformer components, text and image generation workflows, and real-world use cases with models like GPT, BERT, LLaMa, and DALL·E. Ideal for building GenAI-powered applications.

Taught by

Priyanka Mehta

Reviews

Start your review of Attention Mechanisms and Transformer Models Course

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.