Introduction to Programming with Python
Gain a Splash of New Skills - Coursera+ Annual Nearly 45% Off
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a 23-minute research presentation examining shared LoRA (Low-Rank Adaptation) subspaces as a solution for continual learning in artificial intelligence systems. Discover how researchers from Johns Hopkins University approach the challenge of enabling AI models to learn new tasks without forgetting previously acquired knowledge through innovative subspace sharing techniques. Learn about the mathematical foundations behind eigenvector analysis in AI contexts and understand how LoRA adaptations can be structured to maintain performance across multiple learning phases. Examine the practical implications of this approach for developing AI systems that can continuously adapt and expand their capabilities while preserving existing knowledge, addressing one of the fundamental challenges in machine learning known as catastrophic forgetting.
Syllabus
The Eigenvectors of AI: Shared LoRA Subspaces for Continual Learning
Taught by
Discover AI