Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore Stanford NeuroAI Lab's groundbreaking Probabilistic Structure Integration (PSI) framework that constructs self-improving visual world models from raw, non-linguistic data including 1.4 trillion tokens from internet video clips. Learn how this novel approach addresses key limitations in existing world models through coarse controllability and inflexible query interfaces by implementing a probabilistic graphical model approximated by neural predictor Ψ. Discover the three-step virtuous cycle comprising probabilistic prediction to build Ψ, structure extraction through zero-shot causal inference, and integration of extracted intermediates as new token types for continual training. Understand how the local random-access sequence model operates as an autoregressive transformer incorporating pointer tokens for arbitrary serialization orders of patch data, while the hierarchical local quantizer encodes patches into multi-resolution codes preserving spatial locality. Examine structure extraction techniques leveraging causal interventions including tracer counterfactuals for optical flow, motion hypotheticals for object segmentation, and viewpoint hypotheticals for depth estimation achieving state-of-the-art unsupervised results. Analyze the integration process that interleaves new structure tokens with RGB tokens in mixed sequences for autoregressive training refinement, and explore practical applications in physical video editing and robotic motion mapping with empirical scaling laws confirming efficient parameter utilization.