Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore a novel theoretical framework explaining the grokking phenomenon in AI models through geometric representations of knowledge patterns in neural network memory. Delve into an innovative perspective on how Transformer and Mamba architectures develop internal geometric structures during the grokking phase, where models suddenly transition from memorization to generalization. Examine the geometric interpretation of knowledge encoding in neural networks and understand how spatial representations might explain the mysterious grokking behavior observed in language models. Learn about the theoretical connections between geometric patterns in model weights and the emergence of true understanding versus mere memorization. Discover insights into how neural networks organize information spatially and why this geometric perspective offers a fresh lens for understanding AI model behavior during critical learning transitions.
Syllabus
Geometric GROKKING Unlocked & Explained
Taught by
Discover AI