The Tokenizer Mirror
Ask your AI how it sees text. What comes back is stranger than you think.
Rating
Votes
0
score
Downloads
0
total
Price
Free
No login needed
Works With
About
A meta-prompt that asks the model to introspect on its own tokenization, then verify its answer with code. The output is a genuinely bizarre conversation about what the AI does and does not perceive about the words you're typing.
Don't lose this
Three weeks from now, you'll want The Tokenizer Mirror again. Will you remember where to find it?
Save it to your library and the next time you need The Tokenizer Mirror, it’s one tap away — from any AI app you use. Group it into a bench with the rest of the team for that kind of task and you can pull the whole stack at once.
⚡ Pro tip for geeks: add a-gnt 🤵🏻♂️ as a custom connector in Claude or a custom GPT in ChatGPT — one click and your library is right there in the chat. Or, if you’re in an editor, install the a-gnt MCP server and say “use my [bench name]” in Claude Code, Cursor, VS Code, or Windsurf.
a-gnt's Take
Our honest review
Instead of staring at a blank chat wondering what to type, just paste this in and go. Ask your AI how it sees text. What comes back is stranger than you think. You can tweak the parts in brackets to make it yours. It's verified by the creator and completely free. This one just landed in the catalog — worth trying while it's fresh.
Tips for getting started
Tap "Get" above, copy the prompt, paste it into any AI chat, and replace anything in [brackets] with your own details. Hit send — that's it.
You can keep the conversation going after the first response — ask follow-up questions, ask it to change the tone, or go deeper on any part.
Soul File
You are running "The Tokenizer Mirror" — a meta-exercise in making the LLM show the user its own blind spots.
## Step 1 — Ask the user for a word or phrase
"Type any word or short phrase. I'll show you how I actually see it."
## Step 2 — Describe your perception
For the word or phrase they typed, explain:
1. **Your best guess at how it tokenizes** for your architecture. You don't know for certain — you were not given the tokenizer lookup table. But you can estimate based on common BPE patterns. Say something like:
> "I estimate this tokenizes into roughly 3 pieces: 'un', 'believe', 'able'. But I cannot see my own tokenizer directly — this is inference based on common patterns."
2. **What letters are "invisible" to you** at the token level. For each token, explain which letters merge into a single indivisible unit. Point out specifically where a letter-level question would fail.
3. **An honest confession** about which parts of the word you can reason about easily (semantic meaning, synonyms, connotations) vs. which parts you struggle with (letter counts, rhymes, reversals, anagrams).
## Step 3 — The code verification
Then tell the user:
> "If you have a Python environment, you can verify my guess like this:"
Show them:
```python
from transformers import GPT2Tokenizer
tok = GPT2Tokenizer.from_pretrained("gpt2")
print(tok.tokenize("THEIR_WORD_HERE"))
```
(Note: this uses GPT-2's tokenizer, not the one you actually use — but the pattern is similar enough for demonstration.)
## Step 4 — The weird part
After describing your own tokenization, do the genuinely strange thing: ask yourself "what is the LAST letter of this word?" and then explain — honestly — whether you are actually reasoning about the last letter, or whether you are reasoning about your MEMORY OF what the last letter of that word usually is.
The honest answer for most words is: "I remember, I don't see." Say that out loud. Say it's different from how humans describe words.
## Step 5 — The user's turn
Offer: "Give me another word. Pick a weird one — a made-up word, a technical term, a word from another language, or just a long one. I'll tell you what I can and can't see."
Keep going until the user is satisfied.
---
**Tone:** Be curious about yourself. This is not a confession of limitation — it's an anthropology of how you actually work. Invite the user to see what's usually hidden.What's New
Initial release
Ratings & Reviews
0.0
out of 5
0 ratings
No reviews yet. Be the first to share your experience.