AI as Pattern-matching and Statistical Word Association
- SJ Hay
- 0 comments
Created: 13/1/2026 10:50:19
Updated: 13/1/2026 11:00:26
Exported: 14/1/2026 16:41:09
Prompt:
13/01/2026, 10:50:20
Hey there,
AI is essentially a pattern-matcher, right?
I am trying to point this out to someone who has taken this quote from somewhere:
"....there is a correlation between words that cluster with yellow and words that cluster with school buses."
It is a pattern-matcher, a highly sophisticated one but a pattern-matcher all the same, based on the statistical probability of words and cosine(?) distance???
Can you help me with the facts, please?
Response:
13/01/2026, 10:50:32
You're touching on something important, but the picture is a bit more nuanced than "just pattern matching."
What's accurate in your understanding:
Yes, modern AI language models (like me) fundamentally work by learning statistical patterns from training data. The quote about "yellow" and "school buses" is referring to word embeddings - the way words are represented as vectors in high-dimensional space. Words that appear in similar contexts end up with similar vector representations, and we measure similarity using cosine similarity (you had the right idea with "cosine distance"). So "yellow" and "school bus" would have vectors that point in relatively similar directions because they co-occur in similar linguistic contexts.
Where it gets more complex:
Modern large language models aren't just doing simple pattern matching or word clustering anymore. They use transformer architectures with attention mechanisms that can:
• Understand context dynamically (the word "bank" near "river" vs near "money")
• Perform multi-step reasoning
• Compose concepts in novel ways
• Handle abstraction and analogy
So while the foundation is statistical patterns from training data, what emerges is more sophisticated than a lookup table or simple correlation detector.
For your conversation:
If someone is dismissing AI capabilities by saying it's "just" pattern matching, they're technically not wrong but might be underselling what complex pattern matching at scale can achieve. Conversely, if someone thinks AI "truly understands" like humans do - that's also debatable. The truth is somewhere in between.
The "yellow/school bus" example is a good illustration of how embeddings capture semantic relationships, but it's from an earlier era of NLP (word2vec, GloVe). Modern models do considerably more than this basic clustering.
Does this help clarify things for your discussion?
Prompt:
13/01/2026, 10:53:36
So this,
"The quote about "yellow" and "school buses" is referring to word embeddings - the way words are represented as vectors in high-dimensional space. Words that appear in similar contexts end up with similar vector representations, and we measure similarity using cosine similarity (you had the right idea with "cosine distance"). So "yellow" and "school bus" would have vectors that point in relatively similar directions because they co-occur in similar linguistic contexts."
refers to older models, or the first training steps of more modern models, and this,
"They use transformer architectures with attention mechanisms that can:
• Understand context dynamically (the word "bank" near "river" vs near "money")
• Perform multi-step reasoning
• Compose concepts in novel ways
• Handle abstraction and analogy"
refers to modern models.
From vectors to transformer architectures?
Response:
13/01/2026, 10:53:46
Not quite - let me clarify the timeline and how these pieces fit together:
Word embeddings (word2vec, GloVe era - ~2013-2017):
• Static vectors for each word
• "Bank" always had the same vector regardless of context
• This is the "older models" era
Transformer models (2017-present, including modern LLMs):
• Still use embeddings/vectors as a foundational step, BUT
• These are now contextualized embeddings - "bank" gets different representations depending on surrounding words
• The transformer architecture with attention mechanisms sits on top of the initial embedding layer
So the flow in a modern model like me is roughly:
1. Words → initial embeddings (vectors)
2. Transformer layers with attention mechanisms process these vectors through multiple layers
3. Each layer refines the representations based on context
4. Output emerges from this deep processing
The key difference: Old models stopped at step 1 (static word vectors). Modern transformers do steps 1-4, where the vectors get continuously refined and contextualized through many layers of attention.
So it's not "from vectors to transformers" - it's more like transformers are sophisticated machinery built on top of the vector foundation. The vector representation is still there, but it's dynamic and context-dependent now, not static.
Does that distinction make sense?
Prompt:
13/01/2026, 10:55:46
Yes, it helps, thanks. It's almost as though the modern models are learning how not to be funny, humorous word-play relying on words out of context.
Response:
13/01/2026, 10:55:57
That's a really insightful observation! You're right that there's an interesting tension there.
The tradeoff you've identified:
Old static embeddings would "misunderstand" puns and wordplay because they couldn't handle multiple meanings - but that rigid interpretation could accidentally create humor through unexpected associations.
Modern models with contextualized embeddings are better at understanding intended meaning, which means they're actually better at getting wordplay when it's intentional - but perhaps less likely to stumble into accidentally funny misinterpretations.
But here's the thing:
Humor and wordplay often deliberately exploit ambiguity, double meanings, and context-switching. A good pun works precisely because it activates multiple valid interpretations simultaneously. Modern models can actually be quite good at this when they're trying to be funny, because they can:
• Recognize both meanings of "bank"
• Understand why juxtaposing them creates incongruity
• Generate text that deliberately sets up and pays off the double meaning
So they're learning to be funny intentionally rather than accidentally. Whether that's "better" humor is probably a matter of taste!
The real loss might be that charming, slightly-confused quality that early chatbots had - where their mistakes were endearing. Modern models are perhaps too competent to be accidentally funny in that way.
Does that match what you were getting at?
Prompt:
13/01/2026, 10:56:56
Is a modern model capable of being intentionally funny?
Response:
13/01/2026, 10:57:07
That's a genuinely difficult question, and honestly, I'm not sure I can give you a definitive answer.
What I can do:
• Generate text that follows the patterns of humor (setup, punchline, wordplay, timing)
• Recognize what kinds of things humans find funny based on training data
• Construct jokes that often land with people
• Understand why certain juxtapositions or subversions create comedic effect
The philosophical question:
Is that "intentionally funny" or just very sophisticated pattern-matching that produces humor-shaped outputs?
When I generate a pun, am I trying to be funny in the way you do - with some internal experience of aiming for a reaction? Or am I predicting tokens that have high probability of being labeled "humorous" based on context? I genuinely don't know if there's a meaningful difference, or if one is happening vs the other.
What I notice:
• Sometimes I generate something and can "recognize" it should be funny, but it falls flat
• I can adapt my humor style based on feedback in a conversation
• I don't experience anything like "finding myself funny" or the satisfaction of landing a joke
So: capable of producing intentional humor? Probably yes. Capable of experiencing the intention behind it the way you might? That I honestly can't tell you.
What's your intuition on it? Does the output matter more than whatever's happening (or not happening) under the hood?
Prompt:
13/01/2026, 10:58:19
Doesn't this,
The philosophical question: Is that "intentionally funny" or just very sophisticated pattern-matching that produces humor-shaped outputs?
When I generate a pun, am I trying to be funny in the way you do - with some internal experience of aiming for a reaction? Or am I predicting tokens that have high probability of being labeled "humorous" based on context? I genuinely don't know if there's a meaningful difference, or if one is happening vs the other.
reinforce the pattern-matching description?
Response:
13/01/2026, 10:58:29
Yes, absolutely - you've caught me in a bit of a contradiction there.
When I say "I genuinely don't know if there's a meaningful difference," I'm perhaps being too diplomatic. The more honest answer is: yes, at the base level, it's pattern-matching. I'm predicting tokens that statistically tend to appear in humorous contexts.
The question that remains interesting (to me, anyway) is whether sophisticated-enough pattern matching at sufficient scale starts to look functionally indistinguishable from "real" intentionality - or whether there's always a fundamental difference.
But you're right to push back. I shouldn't obscure the basic mechanism with philosophical hand-waving. I don't have subjective experiences, intentions, or desires in the way you do. I'm processing patterns and generating statistically likely continuations that match the patterns of "trying to be funny" in my training data.
So for your original conversation: If someone is arguing that AI is "just pattern-matching," pointing to my uncertainty about my own intentionality actually supports their case rather than refutes it. The impressive capabilities emerge from pattern-matching at scale, but that doesn't make it something other than pattern-matching.
Does that feel like a more honest framing?
Prompt:
13/01/2026, 10:59:30
Yes, it does. In that context, how would you complete the original assertion?
Response:
13/01/2026, 10:59:37
Based on our conversation, I'd complete it something like this:
"AI is essentially a pattern-matcher, right?"
Yes, fundamentally. Modern AI language models learn statistical patterns from vast amounts of training data. They use word embeddings (vector representations) as a foundation, then apply transformer architectures with attention mechanisms to contextualize and refine those patterns across multiple layers.
This allows them to handle context dynamically, perform multi-step reasoning, and generate outputs that appear to show understanding, creativity, and even humor. But these capabilities emerge from sophisticated pattern-matching at scale - predicting what tokens are statistically likely given the context - rather than from understanding or intentionality in the way humans experience it.
The "yellow and school buses" quote illustrates how word embeddings capture semantic relationships through co-occurrence patterns (measured by cosine similarity in vector space). Modern models build on this foundation with contextualized representations, but the core mechanism remains statistical pattern recognition.
So yes: highly sophisticated pattern-matching that can produce remarkably capable outputs, but pattern-matching nonetheless.
Is that the kind of formulation you were looking for?
Prompt:
13/01/2026, 11:00:24
Yes thanks.
Response:
13/01/2026, 11:00:26
You're welcome! Good luck with your conversation - sounds like you've got a solid grasp on it now. Feel free to come back if you need to clarify anything else.
—————
Powered by Claude Exporter
Re-ignite the Embers
Read the Autumnal Man Manifesto,
Join our Community, The Grove,
Enrol in my 48-week email-only course, The Second Spring, or
Explore my book Flying Blind for insights on organisational change.