The Mechanism That Makes LLMs Actually Understand Language: Self-Attention Explained
Static embeddings can't tell 'bank' the financial institution from 'bank' the riverbank. Self-attention is how language models fix that โ by rewriting each token's meaning based on what surrounds it.
Johannes Hayer
johanneshayer