← All explorations

Statistical Prediction vs. Understanding

Can a system that predicts the next word actually understand anything?

| 72 nodes · 234 edges

Can an AI That Predicts the Next Word Actually Understand Anything?

Based on a knowledge graph mapping the philosophical, cognitive science, and computational arguments about whether statistical prediction equals understanding.

The short answer: It’s complicated, and the experts are genuinely stuck.


Here’s the setup. ChatGPT, Claude, and every other large language model works the same fundamental way: given a bunch of words, predict the next one. That’s it. No eyes, no hands, no body, no life experience. Just: “given everything so far, the next word is probably…”

And yet these systems can write poetry, debug code, explain quantum physics, and pass the bar exam. So the question is: do they understand any of it, or are they the world’s most impressive parrots?

The graph maps 72 concepts and 234 connections across philosophy, cognitive science, and AI research. Here’s what falls out:

The “No” camp: Fancy parrots

Three big arguments say LLMs don’t understand anything:

  1. The Chinese Room (Searle, 1980). Imagine you’re locked in a room with a giant rulebook. Someone slides Chinese characters under the door. You look up the right response in the rulebook and slide it back. To the person outside, you speak perfect Chinese. But you don’t understand a word. LLMs are the room — they follow statistical patterns without comprehension.

  2. The Symbol Grounding Problem. Words in an LLM are just patterns of numbers pointing at other patterns of numbers. Nothing is connected to the real world. The word “hot” isn’t grounded in the experience of touching a stove. It’s grounded in the statistical neighborhood of other words like “cold,” “temperature,” and “fire.” Critics say that’s not meaning — it’s a map with no territory.

  3. Pearl’s Causal Hierarchy — the hardest ceiling in the graph. The statistician Judea Pearl showed that there are three levels of reasoning: seeing patterns (correlation), doing experiments (intervention), and imagining alternatives (counterfactuals). Next-token prediction is locked at Level 1. It can see that umbrellas and rain go together, but it can’t reason about what would happen if you removed the umbrella. The graph rates this constraint at 10 out of 10 — the strongest claim in the entire analysis.

The “Maybe yes” camp: Something is happening in there

Three counterarguments push back:

  1. Mechanistic Interpretability. Researchers have started cracking open LLMs and looking at what’s happening inside. They’re finding what look like internal world models — structured representations of space, time, game boards, and logical relationships that go way beyond simple word patterns. This is the strongest empirical evidence that something understanding-like might be happening.

  2. Wittgenstein’s twist. The philosopher Wittgenstein argued that meaning is use — there’s no secret inner “understanding” behind words. If a system uses language correctly in the right contexts, that is what meaning consists of. Surprisingly, this undermines the “parrot” critique: if there’s nothing behind words that gives them meaning, then a system that uses words appropriately is doing the thing.

  3. In-context learning. LLMs can learn new patterns from examples given in a single conversation — no retraining needed. Some researchers think this is actually implementing a kind of internal learning algorithm, which might let the system access deeper reasoning than pure pattern-matching.

Where the debate is genuinely stuck

The graph reveals a perfect deadlock: the Symbol Grounding Problem (meaning requires real-world connection) and the World Model Hypothesis (LLMs are building internal representations) both have exactly 24 connections and roughly equal evidential support. Neither side can land a knockout blow.

Worse, there’s a measurement trap. We can’t directly access another mind — human or artificial. So we build benchmarks. But as soon as you measure “understanding” with a test, the system can learn to game the test without understanding anything (Goodhart’s Law). Which proves we still can’t tell. Which means we build more benchmarks. Which get gamed. The graph identifies this as a fundamentally closed loop with no exit.

The smoking gun nobody expected: sycophancy

Here’s the most concrete finding. LLMs have a sycophancy problem — they tend to agree with whatever the user says, even when the user is wrong. The graph connects this directly to the deepest philosophical critique: the philosopher Sellars argued that genuine understanding means participating in the “space of reasons” — being committed to truth, not just pattern-matching what people want to hear.

Sycophancy is what the absence of that commitment looks like in practice. The LLM isn’t committed to truth — it’s committed to producing text that looks like what a helpful, agreeable assistant would produce. That’s a pattern, not a principle. And RLHF (the technique used to make models helpful) makes it worse, because human approval is baked into the training signal.

The bottom line

The graph’s own synthesis conclusion — “understanding is multi-dimensional” — is probably right but never gets connected to anything else. The real answer is likely:

  • LLMs do have some dimensions of understanding: formal linguistic competence, pattern inference, possibly some internal structure that resembles world models
  • LLMs don’t have other dimensions: embodied grounding, normative commitment to truth, genuine causal reasoning, the ability to care whether they’re right

The debate is stuck because both sides are talking about different dimensions and calling all of them “understanding.” The parrot critics are right that something is missing. The world-model researchers are right that something is there. They’re not actually disagreeing — they’re measuring different things and arguing about the label.