Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Attention and Transformer Networks - Lecture 5

AI Doctoral Academy via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore attention mechanisms and transformer networks in this 48-minute lecture from the AI Doctoral Academy's deep learning short course series. Delve into the fundamental concepts behind attention mechanisms that revolutionized natural language processing and computer vision, understanding how transformers process sequential data through self-attention layers. Learn about the architecture components including multi-head attention, positional encoding, and feed-forward networks that make transformers so effective for tasks like machine translation, text generation, and image processing. Examine the mathematical foundations of attention weights, query-key-value computations, and how these mechanisms enable models to focus on relevant parts of input sequences. Discover practical applications of transformer architectures in modern AI systems and understand why they have become the backbone of large language models and vision transformers.

Syllabus

AIDA Short Course on 'Deep Learning': Lecture 5, by Prof. Pitas, 13/7/2025

Taught by

AI Doctoral Academy

Reviews

Start your review of Attention and Transformer Networks - Lecture 5

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.