Learn EDR Internals: Research & Development From The Masters
Most AI Pilots Fail to Scale. MIT Sloan Teaches You Why — and How to Fix It
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the intricacies of Reinforcement Learning from Human Feedback (RLHF) in this 53-minute talk by Wei Xiong from UIUC, presented at the Simons Institute. Delve into the mathematical foundations of RLHF, examining its formulation as a reverse-KL regularized contextual bandit problem and its statistical efficiency. Discover how continuous online exploration through human evaluator interactions enhances RLHF's effectiveness. Learn about a novel, provably efficient online iterative training framework that spawns innovative RLHF algorithms like iterative direct preference learning. Gain practical insights into creating state-of-the-art chatbots using open-source data, as demonstrated in the RLHFlow project. This talk, part of the "Emerging Generalization Settings" series, offers a deep dive into the cutting-edge techniques aligning large language models with human preferences.
Syllabus
Iterative preference learning methods for large language model post training
Taught by
Simons Institute