Back to Blog

The Wall Between You and the Model: Tokens, Encoders, and Embeddings

GPT-4 can't count letters. Here's why — and how embeddings give token IDs the meaning the tokenizer strips away.

Johannes Hayer

Johannes Hayer

johanneshayer

    The Wall Between You and the Model: Tokens, Encoders, and Embeddings