Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to build a Vision Transformer (ViT) from scratch in this hour-long tutorial that explores the revolutionary application of self-attention mechanisms in computer vision. Master key concepts including CLIP and SigLIP models, image preprocessing techniques, patch and position embeddings, multi-head attention, and MLP layers. Follow along with hands-on implementation using provided code in Google Colab to gain practical experience in constructing a complete Vision Transformer architecture. Visualize embeddings and understand how AI models process visual data through detailed explanations and step-by-step demonstrations. Progress from fundamental concepts to advanced implementation, concluding with a comprehensive recap that reinforces the entire development process.
Syllabus
0:00:00 Intro to Vision Transformer
0:03:48 CLIP Model
0:08:16 SigLIP vs CLIP
0:12:09 Image Preprocessing
0:15:32 Patch Embeddings
0:20:48 Position Embeddings
0:23:51 Embeddings Visualization
0:26:11 Embeddings Implementation
0:32:03 Multi-Head Attention
0:46:19 MLP Layers
0:49:18 Assembling the Full Vision Transformer
0:59:36 Recap
Taught by
freeCodeCamp.org