Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Memorization and Logical Reasoning Capabilities in Large Language Models

Simons Institute via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Watch a 48-minute research talk from Google Research scientist Chiyuan Zhang at the Simons Institute exploring how large language models (LLMs) handle logical reasoning tasks. Dive into findings about the complex relationship between memorization and genuine reasoning capabilities in LLMs through experiments with Knights and Knaves puzzles. Learn how these models can achieve near-perfect accuracy on training puzzles through memorization yet struggle with slight variations, while still showing improved generalization after fine-tuning. Examine detailed analyses including perturbation tests, cross difficulty-level transfer, model probing, and experiments with incorrect answers that reveal how LLMs balance memorization versus actual reasoning when solving logical problems. Understand the implications of per-sample memorization scoring for determining when models rely on memorized patterns versus engaging in true logical reasoning.

Syllabus

On Memorization of Large Language Models in Logical Reasoning

Taught by

Simons Institute

Reviews

Start your review of Memorization and Logical Reasoning Capabilities in Large Language Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.