Learn Excel & Financial Modeling the Way Finance Teams Actually Use Them
Live Online Classes in Design, Coding & AI — Small Classes, Free Retakes
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the controversial reality behind AI benchmarking systems in this 39-minute video that examines Apple's research on language model performance optimization. Discover how AI benchmark scores can be manipulated and learn which benchmarks provide reliable measurements for evaluating large language models. Investigate the relationship between computational requirements (FLOPS) and model accuracy, while understanding scaling laws that govern performance optimization in new LLMs. Examine the critical connection between model size and computational demands, and analyze findings from Apple's research paper "Language Models Improve When Pretraining Data Matches Target Tasks" by researchers from Apple, University of Washington, and Stanford. Gain insights into why traditional benchmarking approaches may be misleading and understand the implications for AI development and evaluation practices.
Syllabus
The Dark Truth behind AI Benchmarks (Apple)
Taught by
Discover AI