3

Sampling from Your Language Model One Byte at a Time
Tokenization is used almost universally by modern language models, enabling efficient text representation using multi-byte or …
Broken Tokens? Your Language Model can Secretly Handle Non-Canonical Tokenizations
Modern tokenizers employ deterministic algorithms to map text into a single “canonical” token sequence, yet the same string …
Tuning Language Models by Proxy
We develop an algorithm for “tuning” language models at decoding-time!