Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Interpretability of LLMs - Sparse Autoencoders Part 1

UofU Data Science via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about sparse autoencoders as a fundamental technique for interpreting large language models in this 20-minute lecture from the University of Utah's CS 6966 course on LLM interpretability. Explore the theoretical foundations and practical applications of sparse autoencoders in understanding how neural networks process and represent information internally. Discover how these mathematical tools help researchers decode the "black box" nature of large language models by identifying interpretable features and patterns within their hidden representations. Access comprehensive course notes to supplement your understanding of this critical interpretability method used in modern AI research.

Syllabus

UUtah CS 6966 Interpretability of LLMs | Spring 2026 | Sparse autoencoders: Part 1

Taught by

UofU Data Science

Reviews

Start your review of Interpretability of LLMs - Sparse Autoencoders Part 1

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.