Google AI Professional Certificate - Learn AI Skills That Get You Hired
Launch Your Cybersecurity Career in 6 Months
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the concept of essential sparsity in large pre-trained models through this insightful 1-hour 10-minute talk by Professor Atwas Wang from the University of Austin Texas. Delve into efficient methods for handling complex and expansive pre-trained transformer models in contemporary machine learning. Discover the threshold at which removing small-magnitude weights significantly impacts performance compared to lower levels of sparsity. Gain access to the project code on GitHub and learn about the research paper "The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter." Explore additional resources such as The Deep Dive newsletter for the latest AI research and industry trends, and Unify's blog for insights into the AI deployment stack. Connect with Unify through their website, GitHub, Discord, and Twitter to stay updated on AI advancements, transformers, language models, and sparsification techniques.
Syllabus
The Emergence of Essential Sparsity in Large Pre-trained Models
Taught by
Unify