© 2024 Felix Ng

arrow_backBack to Journal
Experts Have World Models. LLMs Have Word Models. Here's Why That Matters.
February 9, 2026Journal6 min read

Experts Have World Models. LLMs Have Word Models. Here's Why That Matters.

Experts Have World Models. LLMs Have Word Models. Here's Why That Matters.

"The map is not the territory." — Alfred Korzybski

Research and reflection
Research and reflection

A fascinating paper crossed my desk this week, and it's been haunting my thoughts ever since. The central claim: human experts operate with "world models"—causal, physical, spatial understanding of how things actually work—while Large Language Models, despite their impressive outputs, are essentially manipulating "word models" with no grounded reality.

This distinction sounds academic until you realize it explains almost everything about where AI succeeds, where it fails, and where the real breakthroughs will come from.

The Difference in a Nutshell

World models are internal representations that mirror external reality. When a physicist imagines a ball rolling down a hill, they're simulating gravity, friction, momentum—not just remembering words about these concepts. The model captures causality: if this, then that, because of how the world actually works.

Word models are statistical patterns in language. When an LLM describes a ball rolling down a hill, it's predicting which words typically appear together. It has no simulation, no physics engine, no causal mechanism—just excellent autocomplete.

The paper's adversarial examples make this brutally clear. Ask an LLM about physical scenarios that require causal reasoning, and the facade crumbles. Not because the model is "dumb," but because it's doing something fundamentally different than understanding.

Where This Shows Up in Practice

Programming: Syntax vs. Semantics

When a senior developer reviews code, they're not just checking if the syntax is valid. They're simulating execution: "If this runs, memory gets allocated here, this thread waits there, this race condition emerges."

Current AI coding assistants excel at syntax and pattern matching. They generate plausible-looking code that often works. But when things break, they lack the causal model to diagnose why. They can suggest fixes that compile but don't solve the underlying problem.

The gap: LLMs have seen millions of code examples. Expert developers have built mental models of how computers actually work.

Medicine: Pattern Matching vs. Mechanistic Understanding

A dermatologist looks at a rash and sees not just visual patterns but underlying pathology: "This distribution suggests this mechanism which implies this treatment."

Medical AI systems are getting remarkably good at pattern recognition—sometimes exceeding human accuracy on specific diagnostic tasks. But when confronted with unusual presentations or asked to reason about novel interventions, they lack the mechanistic understanding to generalize.

The gap: Training on medical images and text doesn't create a model of human physiology.

Business Strategy: Correlation vs. Causation

Experienced executives develop intuition about market dynamics: "If we lower prices here, competitors respond there, margins compress here, but volume increases enough to justify it."

AI systems trained on business data can identify correlations and trends. But they struggle with counterfactual reasoning—"what would happen if we did X instead of Y"—because they lack causal models of market behavior.

The gap: Historical data shows what happened, not why it happened.

The Implications for AI Development

This distinction isn't just philosophical—it points toward where AI research needs to go.

Current LLMs: Ceiling Approaching?

If the paper's analysis is correct, current LLM architectures may be approaching fundamental limits. More data, more parameters, more compute—all improve word model performance but don't create world models.

We've seen hints of this: GPT-4's impressive but brittle reasoning. Claude's careful hedging on physical reasoning tasks. The consistent failure modes across different model architectures.

The ceiling might be higher than we thought, but it exists.

Hybrid Approaches: The Path Forward?

Several research directions attempt to bridge this gap:

Multimodal training: Exposing models to video, simulation, robotics—grounding language in physical experience.

Neurosymbolic methods: Combining neural pattern matching with explicit symbolic reasoning about causality.

World models in the loop: Systems like Waymo's driving simulator that maintain explicit physical models alongside learned patterns.

Interactive learning: Agents that learn by acting in environments, not just observing text.

None of these are solved problems. But they represent the frontier where progress toward true understanding might happen.

What This Means for Users

If you're using AI tools (and you should be), this analysis suggests practical guidelines:

Trust LLMs for:

  • Pattern recognition and summarization
  • Generating plausible starting points
  • Tasks where correctness can be easily verified
  • Domains with extensive training data

Be cautious with LLMs for:

  • Physical reasoning and causal inference
  • Novel situations outside training distribution
  • Tasks requiring counterfactual reasoning
  • Safety-critical decisions

The ultimate test: Can you verify the output independently? If yes, LLMs are powerful tools. If no, you're outsourcing judgment to a word model—and that's risky.

The Deeper Question

Reading this paper, I kept returning to a fundamental question: What would it mean for an AI to have a world model?

Not just better pattern matching. Not just more training data. But actual internal representations that mirror reality, support causal reasoning, and enable genuine understanding.

Some researchers believe this requires embodiment—AI that learns by interacting with the physical world, building models through experience like humans do. Others think sophisticated simulation might suffice. A few believe symbolic reasoning must be explicitly engineered.

I don't know the answer. But I suspect the next breakthrough in AI won't come from bigger language models. It will come from systems that bridge the gap between words and world.

Conclusion

The distinction between world models and word models clarifies a lot of confusion about current AI capabilities. It explains both the genuine utility of LLMs and their consistent failure modes. It points toward research directions that might lead to more robust, more capable systems.

Most importantly, it reminds us that impressive output isn't the same as genuine understanding. The words might be right. The reasoning might be wrong.

As we integrate AI into more critical systems—medicine, law, engineering, policy—the difference matters enormously. Word models are powerful tools. But world models are what we actually need.

The gap between them is where the next decade of AI research will be won or lost.

What do you think? Can LLMs develop world models through scale and training, or is something fundamentally missing in current architectures?