Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a comprehensive evaluation of Anthropic's newly released Claude OPUS 4.6 models through rigorous causal reasoning testing in this 11-minute video analysis. Compare the performance of both the Thinking and Non-Thinking versions of Claude OPUS 4.6 as they tackle standardized causal reasoning challenges, despite Anthropic's claims that "Opus 4.6 extends the frontier of expert-level reasoning." Witness firsthand testing that reveals significant limitations in both model variants, including system crashes and validation failures during the assessment process. Follow the systematic testing methodology that examines initial results from the Non-Thinking model, observes the Thinking model's approach and subsequent crash, conducts validation runs, and identifies key performance problems. Gain insights into the current state of AI reasoning capabilities and understand the gap between marketing claims and actual performance through this detailed technical evaluation that forms part of a broader AI model testing series focused on causal reasoning benchmarks.
Syllabus
OPUS 4.6 TEST
OPUS 4.6 Non-Thinking First Result
OPUS 4.6 Thinking
OPUS 4.6 Thinking Crashed
OPUS 4.6 Validation Run
OPUS 4.6 Problems
Taught by
Discover AI