Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
In this lecture, Aviral Kumar from Carnegie Mellon University explores the optimization of test-time compute for large language models. Examine the formalization of test-time compute optimization as a meta reinforcement learning problem, providing a principled perspective through the lens of exploration and exploitation. Learn how this approach becomes increasingly relevant with larger test-time token budgets and why cumulative regret serves as an effective measurement tool. Discover a finetuning paradigm that optimizes intermediate tokens using dense rewards based on information gain, enabling novel solutions to difficult problems. Follow along as Kumar presents ablation analysis on state-of-the-art models to understand their behavior and potential improvements. The second part delves into theoretical results demonstrating why reinforcement learning with verification is critical for effective scaling of test-time compute, showing how RL outperforms expert cloning in terms of suboptimality decay rates when working with heterogeneous distributions. Based on Kumar's research blog and papers, this talk provides valuable insights into the future of language models and transformers, particularly regarding test-time compute optimization.
Syllabus
The Key Ingredients of Optimizing Test-Time Compute and What's Still Missing
Taught by
Simons Institute