The Two Wrong Models Everyone Starts With
Before we talk about what AI is, let's name what most people think it is. There are two dominant wrong models, and they each cause predictable problems.
Wrong Model 1: The Oracle
You ask it a question. It knows the answer. It's essentially a smarter Google — a vast database of truth that retrieves the correct answer when prompted correctly.
This model produces people who: trust AI outputs without verifying them, get confused when AI is confidently wrong, treat every response as a fact that just needs to be quoted.
Wrong Model 2: The Robot
It follows instructions mechanically, like a very sophisticated autocomplete. Very fast, very literal, but essentially just executing commands without anything like intelligence.
This model produces people who: dramatically underuse AI's capabilities, don't understand why careful prompting matters, give up quickly when results aren't what they expected.
Both models produce systematic mistakes. The Oracle model produces over-trust. The Robot model produces under-use. The accurate model produces neither.
What an LLM Actually Does
A Large Language Model is, fundamentally, a prediction engine.
It was trained on a massive dataset of text — books, websites, code, conversations, articles, documentation, essentially most of human-written language — and learned one thing with extraordinary precision:
Given this sequence of text, what word (technically, what "token") is most likely to come next?
That's not a metaphor. That's the literal mechanism. Every word in every AI response is selected through a probabilistic process of predicting the most contextually appropriate continuation of what you wrote.
What emerged from doing this at massive scale on virtually all of human-written language is something genuinely remarkable: a system that can generate coherent, contextually appropriate text on virtually any topic, reason through problems step by step, write code, translate languages, analyze documents, and explain concepts at multiple levels of complexity.
But the mechanism is prediction, not understanding. This distinction has real consequences.
The Alien Translator
Here's the mental model that actually works: imagine an entity that has read essentially everything humans have ever written, but has never experienced anything.
It has never seen the color red, felt cold, made a decision with real stakes, or cared about an outcome. It has never been surprised. It has never been confused. It has no preferences, no fear, no curiosity.
What it has is an extraordinarily rich statistical model of how humans write about all of these things. It knows that sentences about grief tend to have certain patterns. It knows that good arguments have certain structures. It knows what a persuasive essay sounds like, what a confident medical diagnosis sounds like, what a wrong answer stated confidently looks like.
It doesn't know these things the way a person knows things. It knows them the way a language knows them — as patterns in how things are expressed.
This model predicts your actual experience with AI far better than either wrong model:
- Why it can write beautifully about grief without feeling it: it's pattern-matching on how grief is written about, not experiencing grief itself
- Why it hallucinates: when it doesn't have a strong pattern match for a specific fact, it generates something that looks like a correct answer — the shape of truth without the substance of it
- Why prompting matters: you're not talking to someone who reads your mind; you're feeding context to a prediction engine that will generate the most plausible continuation of what you wrote
- Why it can be wrong with total confidence: confidence is a tone pattern, not a verification signal — the model learned that certain types of answers sound confident, and it reproduces that tone whether or not the content is accurate
What AI Is Good At — and Where It Fails
Once you hold the right model, the capabilities and limitations of AI become predictable rather than mysterious.
Reliably good:
- Generating well-structured, contextually appropriate text on any topic
- Drafting, rewriting, summarizing, and reformatting content
- Explaining concepts at different levels of complexity
- Writing code that follows common patterns
- Synthesizing multiple sources into a coherent summary
- Generating options, variations, and alternatives
- Following complex multi-part instructions
Unreliable — verify before using:
- Specific factual claims (especially recent events, statistics, citations)
- Precise mathematical calculation (use a calculator or code interpreter)
- Information about events after the model's training cutoff
- Anything where being precisely right matters and you can't check it
- Generating genuinely original ideas (it generates novelty that looks original)
- Knowing what it doesn't know — it rarely signals uncertainty accurately