IQ Archive
Artificial Intelligence

Large Language Model (LLM)

What is a Large Language Model?

A Large Language Model (LLM) is a sophisticated artificial intelligence system built on deep learning techniques and massive datasets, typically involving substantial portions of the public internet. At their core, these models utilize a Transformer architecture, introduced by Google researchers in 2017, which allows the AI to weigh the importance of different words in a sentence (a mechanism called “attention”) to predict the most likely next token in a sequence.

Famous examples include OpenAI’s GPT-4, Anthropic’s Claude, and Google’s Gemini. While they began as text-predictors, their scale has allowed them to perform complex tasks like coding, creative writing, and logical deduction.

LLMs and Human Intelligence

From a psychometric perspective, LLMs present a fascinating case study because they effectively decouple Crystallized Intelligence from Fluid Intelligence and Consciousness.

  1. Crystallized Intelligence (Gc): LLMs possess a level of Gc that far exceeds any human. They have “read” more books, academic papers, and codebases than a human could in a thousand lifetimes. Their ability to retrieve and synthesize this information is superhuman.
  2. Fluid Intelligence (Gf): This is controversial. While LLMs can solve logic puzzles, they often do so by recognizing patterns in their training data rather than performing novel reasoning. However, modern models are showing increasing capability in “zero-shot” reasoning (solving problems they haven’t seen before), suggesting a form of synthetic fluid intelligence.

The “Stochastic Parrot” Debate

A major criticism, coined by linguist Emily M. Bender and her colleagues, is that LLMs are merely “Stochastic Parrots.” This theory argues that the models do not understand meaning; they simply stitch together linguistic forms based on probability without any reference to the real world.

For example, if you ask an LLM “What color is the sky?”, it answers “Blue” not because it has seen the sky or understands the concept of color, but because the words “sky” and “blue” appear together frequently in its dataset.

Emergent Properties and AGI

The counter-argument relies on Emergent Properties. In complex systems, “more is different.” When an LLM becomes large enough (billions of parameters), it begins to display abilities that were not explicitly programmed, such as the ability to translate languages or debug software.

This leads to the question of Artificial General Intelligence (AGI). If a machine can pass the Turing Test, score in the 90th percentile on the Bar Exam, and diagnose medical conditions better than a doctor, does it matter if it “understands” in the human sense? For the pragmatic definition of intelligence — “the ability to solve problems” — LLMs are currently the closest approximation to a non-biological mind.

How LLMs Actually Work: The Transformer Architecture

To evaluate LLM capabilities honestly, it helps to understand the mechanism underneath. The Transformer, introduced in the landmark 2017 paper “Attention Is All You Need” by Vaswani et al., works through a mechanism called self-attention: for each word (token) in a sequence, the model computes how much attention to pay to every other token when predicting what comes next.

During training on vast text corpora, the model adjusts billions of numerical parameters (weights) through gradient descent, learning to predict the next token given all preceding tokens. By the end of training, these weights implicitly encode an enormous amount of world knowledge, linguistic structure, and something that looks — from the outside — like reasoning.

Key architectural features that enable LLM capabilities:

  • Scale: Modern frontier models have hundreds of billions of parameters, trained on trillions of tokens of text. Scale alone has proven to unlock capabilities that smaller models entirely lack.
  • Context window: The amount of text the model can “hold in mind” at once — its working memory analogue. Early models had context windows of a few hundred tokens; modern models can process hundreds of thousands.
  • In-context learning: LLMs can adapt their behavior based on examples provided within the prompt, without any weight updates. This is a form of rapid, flexible learning that has no clear precedent in earlier AI systems.

Benchmarking LLM Intelligence: How Do They Score?

Researchers have administered a range of standardized cognitive assessments to LLMs, producing results that are both impressive and revealing:

  • SAT/GRE: Top models score in the 90th+ percentile on verbal and mathematical sections of these tests.
  • Bar Exam: GPT-4 scored approximately in the top 10% of human test-takers on the Uniform Bar Exam — a dramatic improvement over earlier models that scored near the bottom.
  • Medical licensing (USMLE): Frontier models pass the United States Medical Licensing Examination at or above the passing threshold.
  • Raven’s Progressive Matrices: Performance is more uneven — models can solve some matrix reasoning problems but fail on others in ways that suggest pattern-matching from training data rather than genuine novel reasoning.
  • Novel mathematical olympiad problems: Performance drops sharply on problems that require genuine mathematical insight rather than retrieval of known solutions.

This pattern — strong on crystallized knowledge tasks, weaker on genuine novelty — aligns with the psychometric framework and suggests LLMs currently excel at a specific subset of human cognitive abilities.

What LLMs Cannot Do: The Remaining Gaps

Despite impressive benchmark performance, LLMs have well-documented limitations that distinguish them from human general intelligence:

No persistent memory: LLMs have no memory across conversations. Each session starts from scratch. There is no accumulated personal history, no learning from experience, no sense of autobiography. This is a fundamental difference from human cognition, where autobiographical memory and experiential learning are central.

No grounded world model: LLMs learn from text about the world, not from direct sensorimotor engagement with it. This creates characteristic failure modes: they can describe how to catch a ball while having no understanding of trajectory, momentum, or gravity beyond statistical co-occurrences in training data.

Confabulation (“hallucination”): LLMs generate plausible-sounding text even when they lack the knowledge to answer accurately. Because the model is optimized to produce fluent, contextually appropriate text rather than to flag uncertainty, it will often generate confident falsehoods. This is the most practically dangerous limitation for real-world applications.

Fragile reasoning: On formal logic and mathematical reasoning tasks, LLMs can be derailed by superficial changes to problem wording that would not affect a human mathematician. This suggests their “reasoning” often relies on surface-level pattern matching rather than deep structural understanding.

The Psychometric Question: Is LLM Intelligence Real?

Perhaps the most intellectually interesting question LLMs raise for psychometrics is whether the concept of “intelligence” requires a specific type of substrate (biological neurons, embodied experience, consciousness) or whether it can be defined purely functionally — as any system that reliably solves problems that require intelligence to solve.

If we adopt the functional definition, frontier LLMs already qualify as highly intelligent on certain dimensions. If we require grounding in embodied experience, genuine novelty of reasoning, or consciousness, they fall well short.

This tension may ultimately force a revision of how we define and measure intelligence — extending frameworks that were built entirely around the human case to encompass genuinely novel cognitive architectures.

Conclusion

Large Language Models represent the most striking development in artificial intelligence since the field’s founding — and the most serious challenge to human cognitive exceptionalism in history. They do not replicate the full spectrum of human intelligence, but they exceed human performance on a widening range of crystallized cognitive tasks. Understanding what they can and cannot do, in rigorous psychometric terms, is essential for anyone navigating the increasingly AI-shaped world of the 21st century.

Related Terms

Artificial General Intelligence Turing Test Neural Networks Crystallized Intelligence Pattern Recognition
← Back to Glossary