Why AI Is Chasing World Models Again
Read ArticleRead Original Articleadded Sep 2, 2025September 2, 2025

AI’s renewed focus on “world models” aims to give systems internal representations for prediction and planning, an idea with roots in 1940s psychology and early symbolic AI. Despite speculation that LLMs already contain such models, empirical work suggests they mostly encode fragmented heuristics that fail under slight changes. Labs are split between inducing models via multimodal training and inventing new architectures, but the potential gains in robustness and interpretability are driving intense interest.
Key Points
- World models — internal, simplified representations of reality — are seen as crucial for prediction, planning and robust decision-making in AI.
- Historically, world models inspired early symbolic AI but were rejected by robotics leader Rodney Brooks for being brittle, only to be revived by deep learning.
- Evidence suggests current LLMs rely on disconnected “bags of heuristics” rather than coherent, globally consistent models of the world.
- A navigation study showed LLMs can excel in familiar settings but fail under slight perturbations (e.g., 1% of streets blocked), underscoring the need for consistent internal representations.
- Major labs pursue different paths: DeepMind/OpenAI bet on multimodal data to induce world models, while Meta’s LeCun argues for new, non-generative architectures to explicitly build them.
Sentiment
Mixed but generally supportive of the need for robust world models; skeptical that current LLMs have them, and critical of the article’s omissions and the practical challenges (state, compute, optimization).
In Agreement
- Useful agents must plan with a model of the environment; whether learned or hand-coded, planning remains central.
- Current LLMs behave like bags of heuristics and struggle to maintain and update coherent state across steps (e.g., mazes, text adventures), matching the article’s critique.
- Explicit constraints and structured outputs (grammars/validators/DSLs) can reduce invalid actions and hallucinations, reflecting the appeal of more robust internal models.
- Learned world models (e.g., MuZero) demonstrate that planning-with-a-learned-model is viable and promising.
- Multimodal/embodied approaches and neuroscience-aligned architectures (e.g., place/grid cell analogs) could help world models emerge and generalize.
- World models should be adaptable and subordinate to incoming evidence (open-world assumption), addressing distribution shifts.
- The frame problem remains relevant: models need principled ways to update only what changes.
Opposed
- For many board and turn-based games, hand-coded rules plus search still outperform ML-only systems; the practical need for new world-model breakthroughs is overstated.
- Implementing a world model is often easy; the hard problem is optimizing/searching it fast enough—suggesting the article puts emphasis on the wrong bottleneck.
- LLMs with structured decoding and external constraints might be sufficient for many tasks, reducing the need for new architectures aimed at internal world models.
- Learned models like MuZero can be brittle, require massive compute, and still rely on legality constraints—so they are not clearly superior to explicit models.
- The article is criticized as incomplete or ‘useless,’ omitting key prior work (e.g., Ha & Schmidhuber’s World Models, Fei-Fei Li’s efforts) and oversimplifying the field.
- Some claim language has been ‘solved’ by scale and data alone, questioning the article’s emphasis on explicit or robust models for symbolic domains.