
AINLP
Why GPT Can't Count: Visualizing the BPE Tokenization Trap
Interactive visualization of Byte Pair Encoding (BPE) tokenization. Discover why GPT thinks 9.11 > 9.9 and watch the tokenizer chop your text in real-time.
7 min read
Loading...
Make machines understand human language
4 posts

Interactive visualization of Byte Pair Encoding (BPE) tokenization. Discover why GPT thinks 9.11 > 9.9 and watch the tokenizer chop your text in real-time.

Interactive visualization of RNN vs Transformer architecture. See why RNNs forget and how Transformers solve the vanishing gradient problem.

See how large language models decide which words matter in a sentence. Interactive Query, Key, Value simulator shows self-attention live.

Interactive visualization of Word2Vec and word embeddings. See why good and great were strangers in one-hot encoding but neighbors in vector space.