AI Product Expert Certification - Master Generative AI Skills
35% Off Finance Skills That Get You Hired - Code CFI35
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
This video explores a critical flaw in reasoning Large Language Models (LLMs) called "MiP-Overthinking," where AI models produce excessively long, redundant responses when faced with ill-posed questions that have missing premises. Learn how this phenomenon affects both reinforcement learning and supervised learning models, wasting the Chain-of-Thought reasoning capabilities in advanced systems like o1, o3, and R1. Discover the research findings that reveal how current training methods fail to encourage efficient thinking, and examine the detailed analyses of reasoning length, overthinking patterns, and critical thinking locations across different LLM types. Based on research by teams from the University of Maryland, Lehigh University, KIIT Bhubaneswar, KIMS Bhubaneswar, and Monash University.
Syllabus
When Smart AI Models Overthink Stupid Data (AI TRAP)
Taught by
Discover AI