LLM Intuition Tool v2.1

Toy model (n-grams + heuristics), but more “LLM-like”: readable subword-ish tokens (stems + -ing), softmax-style sampling controls, and a toy attention strip. Predictions update as you type.

Analyze Text
Next-word Predictor

1) Paste training text

Walkthrough highlights current token (blue), then +1 (green), +2 (yellow), +3 (red). It increments 1–4 token phrases and stores next-token maps for contexts of length 1, 2, 3, and 4. The loss plot is a toy online NLL under an interpolated n-gram model.
Animated walkthrough Idle
Step /
Current:
+1:
+2:
+3:
Live counts
Vocab:
Unique bigrams:
Unique trigrams:
Unique 4-grams:
Just added
1:
2:
3:
4:
Toy loss (NLL)
avg:
Lower = higher probability on the true next token. This is not backprop—just “counts getting better”.
final view only
for heatmap + toy attention
optional
final heatmap
walkthrough
human-friendly subwords

2) What the model learned

No model yet
Tokens
Vocabulary
Unique bigrams
Unique trigrams
Unique 4-grams
Top tokens (click to highlight)count
Token–token correlation heatmap
low
high
Click a top token to highlight its row/column in the heatmap.
Top 2-token phrases
Top 3-token phrases
Top 4-token phrases
Note
Predictor shows a single probability-ordered list, but still includes a “by context length” breakdown below for intuition.