The Fundamental Problem: Computers Need Numbers
Nina notices something odd. She asks her model to summarize two articles — one about a "river bank" and one about a "bank loan." The summaries are weirdly similar. The model seems to treat "bank" the same way in both.
Why? Because computers don't understand words. They understand numbers. The word "bank" means something completely different next to "river" than it does next to "loan" — but to a machine, it's just a string of characters until someone figures out how to turn meaning into numbers.
For decades, this was the central challenge of language AI. Every technique — from the earliest word counters to modern Transformers — is a progressively better answer to this one question.
This isn't historical trivia. It's the story of why LLMs behave the way they do when you prompt them.