Do LLMs Understand or Just Predict?

3D rendered abstract brain concept with neural network.
3D rendered abstract brain concept with neural network.

Do LLMs Understand or Just Predict?

The rise of Large Language Models (LLMs) like OpenAI’s GPT series, Google’s Gemini, Anthropic’s Claude, and Meta’s Llama has been nothing short of revolutionary. These models can write poetry, debug code, summarize complex texts, and engage in surprisingly coherent conversations. Their capabilities often feel uncannily human, leading to a fundamental question that cuts to the heart of artificial intelligence: Do LLMs truly understand, or are they simply predicting the next word with incredible accuracy?

This isn’t just an academic debate; it has profound implications for how we develop, deploy, and trust AI systems.

The Case for Prediction: The Statistical Machine View

At their core, LLMs are statistical machines. Their training process involves ingesting truly colossal amounts of text data—billions, even trillions of words from the internet, books, articles, and more. The primary objective during this training is called “next-token prediction.” Given a sequence of words, the model’s task is to predict the most probable next word or token.

This process is underpinned by the Transformer architecture, introduced by Google in 2017 1. Transformers leverage “attention mechanisms” that allow the model to weigh the importance of different words in the input sequence when making a prediction. They don’t read in a human sense; rather, they identify complex statistical relationships, patterns, and correlations within the data.

Consider this simple example: If an LLM sees “The cat sat on the…”, it’s been trained on so much text that it knows “mat,” “rug,” “sofa,” or “floor” are highly probable next words. It doesn’t need to conjure an image of a cat or a mat; it just needs to recognize the statistical likelihood based on its training data.

This view is perhaps best encapsulated by the term “Stochastic Parrots,” coined by Emily M. Bender and colleagues 2. They argue that LLMs are “systems for stitching together sequences of linguistic forms it has observed in its vast training data, according to probabilistic rules.” They mimic human language without understanding the meaning or having any grounding in the real world. They don’t experience, embody, or possess common sense in the way humans do. Their “knowledge” is merely a sophisticated reflection of the statistical regularities present in the data they’ve been trained on.

Key arguments for the “prediction only” view include:

  • Lack of World Model: LLMs don’t interact with the physical world. They lack sensory input, motor control, and the embodied experiences that form the basis of human understanding.
  • Symbol Grounding Problem: Their “knowledge” is confined to symbols (words, tokens) and their relationships, without being grounded in real-world referents. They can talk about an apple but don’t know what it tastes like or how it feels.
  • Hallucinations: When an LLM generates plausible but factually incorrect information, it’s often seen as evidence that it’s just generating statistically probable text rather than drawing from a deep, factual understanding 3.
  • Sensitivity to Prompt Phrasing: Minor changes in prompt wording can sometimes lead to drastically different outputs, suggesting a lack of robust, semantic understanding.

The Case for Emergent Understanding: Beyond Pure Statistics

While LLMs are indeed built on statistical prediction, a growing body of research and observation suggests that something more profound might be happening, especially in the largest models. As models scale up in terms of parameters and training data, they exhibit “emergent abilities” – capabilities that were not present in smaller models and were not explicitly programmed 4.

These emergent abilities include:

  • In-context learning: The ability to learn from examples provided in the prompt without explicit fine-tuning.
  • Chain-of-thought reasoning: Breaking down complex problems into intermediate steps, which significantly improves performance on multi-step reasoning tasks 5.
  • Basic theory of mind capabilities: Some studies suggest LLMs can infer mental states (beliefs, intentions) in simple scenarios, albeit in a rudimentary fashion 6.
  • Mathematical and logical problem-solving: Beyond simple arithmetic, larger models can solve complex word problems, logical puzzles, and even generate correct code for intricate algorithms.

Proponents of this view argue that for an LLM to consistently perform these tasks, it cannot merely be stitching words together randomly. It must be forming some kind of internal representation or “model” of the world, or at least the concepts and relationships within its training data. This “understanding” might not be conscious or human-like, but it is functionally robust.

Think of it this way: a highly sophisticated calculator doesn’t “understand” mathematics, but a human mathematician does. If an LLM can perform complex proofs, debug intricate code, or even generate novel scientific hypotheses (as some are starting to assist with), is it simply mimicking, or is it operating on a deeper structural knowledge it has implicitly learned?

Some researchers propose that LLMs develop “internal world models” 7, where they learn latent representations that capture abstract concepts, causal relationships, and even hierarchical structures from the vast text they consume. While these models are not grounded in physical reality, they allow the LLM to generalize, reason, and apply knowledge in novel ways that go beyond simple pattern matching.

Note: This “understanding” is often described as functional or computational understanding, meaning the model behaves as if it understands, rather than implying consciousness or subjective experience.

Defining “Understanding”: The Crux of the Debate

The core difficulty in answering “Do LLMs understand?” lies in how we define “understanding” itself.

  1. Human Understanding: For humans, understanding is deeply intertwined with consciousness, subjective experience (qualia), embodiment, common sense, emotions, intentions, and grounding in a shared reality. It’s multi-modal, incorporating sensory input, motor skills, and social interaction. By this definition, LLMs clearly do not understand. They lack all these attributes.

  2. Functional/Computational Understanding: This perspective asks: Does the system behave in a way that demonstrates comprehension? Can it use information flexibly, answer novel questions, adapt to new situations, infer context, and generate coherent, relevant responses? If a system can do all these things consistently and robustly, then arguably, it functionally understands. From this viewpoint, LLMs, especially the larger ones, show strong signs of functional understanding.

The famous “Chinese Room Argument” by John Searle highlights this distinction 8. In this thought experiment, a person inside a room manipulates Chinese symbols according to a rulebook, without understanding Chinese. From outside, it appears the room understands Chinese, but the person inside does not. Searle argues that LLMs are like the person in the room – they manipulate symbols (language) without grasping their meaning.

While the Chinese Room Argument has its own debates and criticisms, it serves as a powerful reminder that fluent output does not automatically equate to semantic understanding or consciousness.

A Spectrum, Not a Dichotomy

Perhaps the most accurate way to view LLM “understanding” is not as an either/or, but as a spectrum. LLMs are fundamentally predictive systems. Their astonishing capabilities are a direct result of scaling up this predictive power to unprecedented levels, allowing them to capture incredibly subtle and complex statistical regularities in language.

However, the sheer depth and complexity of these statistical relationships, combined with the scale of the models, lead to emergent behaviors that appear to be forms of understanding. They learn to represent concepts, relationships, and even forms of “reasoning” implicitly from the data. This is a powerful, functional approximation of understanding, enabling them to generalize and perform tasks that would be impossible with mere surface-level pattern matching.

It’s akin to saying: “A bird flies because it flaps its wings and generates lift.” That’s the mechanism. But the emergent behavior is “flight.” LLMs predict the next word, but the emergent behavior is coherent reasoning, summarization, and creative writing.

Note: The debate continues, with no single, universally accepted answer. Researchers are actively exploring what these models truly learn and whether their internal representations bear any resemblance to human cognitive processes.

Implications and Future Directions

The question of whether LLMs understand has significant implications:

  • Trust and Reliability: If LLMs only predict, how much can we trust their outputs, especially in critical domains like medicine or law? Hallucinations become a significant concern.
  • Safety and Alignment: If LLMs lack genuine understanding or a human-like value system, ensuring they align with human goals and ethics becomes even more challenging.
  • Path to AGI: Is this “functional understanding” a necessary stepping stone towards Artificial General Intelligence (AGI), or is a fundamentally different architecture required?
  • Research Focus: This debate fuels research into how to ground LLMs in the real world (e.g., through multimodal training involving images, video, and sensory data) 9, or by enabling them to interact with tools and environments 10.

Ultimately, LLMs are incredible computational achievements that have pushed the boundaries of what we thought possible with AI. They are exceptionally sophisticated predictors that exhibit behaviors remarkably similar to understanding. Whether this constitutes “true understanding” depends on the philosophical lens through which one views intelligence and cognition. The journey to unraveling the depths of their capabilities, and indeed, the nature of intelligence itself, is just beginning.


References:


  1. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., … & Polosukhin, I. (2017). Attention Is All You Need. Advances in Neural Information Processing Systems, 30. ↩︎

  2. Bender, E. M., Gebru, T., McMillan-Major, A., & Mitchell, M. (2021). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (pp. 610-623). ↩︎

  3. Ji, Z., Lee, N., Fries, E., Yu, T., & Su, X. (2023). Survey of Hallucination in Large Language Models. arXiv preprint arXiv:2304.01373↩︎

  4. Wei, J., Tay, Y., Bommasani, R., Raffel, C., Zoph, B., Roberts, K., … & Le, Q. V. (2022). Emergent Abilities of Large Language Models. Transactions on Machine Learning Research↩︎

  5. Wei, J., Wang, X., Schuurmans, D., Bosma, M., Chi, F., Le, Q. V., & Zhou, Y. (2022). Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. Advances in Neural Information Processing Systems, 35. ↩︎

  6. Kosinski, M. (2023). Theory of Mind May Have Spontaneously Emerged in Large Language Models. arXiv preprint arXiv:2302.02083↩︎

  7. DeepMind. (2022). What makes a world model?. DeepMind Blog Post↩︎

  8. Searle, J. R. (1980). Minds, Brains, and Programs. Behavioral and Brain Sciences, 3(3), 417-457. ↩︎

  9. Gemini Team, Google. (2023). Gemini: A Family of Highly Capable Multimodal Models. Google DeepMind Blog Post↩︎

  10. OpenAI. (2023). Function calling. OpenAI Developer Documentation↩︎

Last updated on