The Inner Workings of Large Language Models - Visualizing Self-Attention Mechanisms
Discover AI via YouTube
Introduction to Programming with Python
AI Engineer - Learn how to integrate AI into software applications
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the fundamental mechanics of Large Language Models (LLMs) in this 35-minute educational video that demystifies complex concepts through clear visualizations and beginner-friendly explanations. Dive into the self-attention mechanism powering models like ChatGPT and GPT-4, understand the key differentiators between various LLMs including their weights, pre-trained datasets, and architectural design structures, and learn about performance optimization through hardware, software, and architectural tuning. Master essential concepts about decoder-based Transformers, LangChain implementation, and Vector stores with their embeddings. Focus specifically on the decoder stack while examining real-world applications through examples like Claude from Anthropic and cutting-edge research from "AttentionViz: A Global View of Transformer Attention." Access interactive demonstrations and comprehensive documentation to reinforce understanding of these transformative AI technologies.
Syllabus
The inner workings of LLMs explained - VISUALIZE the self-attention mechanism
Taught by
Discover AI