Count tokens and see context-window usage for Claude, GPT, and Llama.
Paste a prompt, doc, or transcript. Get a live token count for each major model and a color-coded utilization bar so you know how much headroom you have before truncation hits.
Compare how your input fits each model's context window. The colored bar is your input; the rest is headroom for tool calls, reasoning, and the model's response.
Counts come from gpt-tokenizer, a zero-dependency BPE encoder. GPT-4o and GPT-5 use o200k_base; GPT-4 / 3.5 use cl100k_base.
Anthropic and Meta use proprietary or differently-sized tokenizers. We approximate Claude with cl100k_base x 1.10 and Llama 3 with cl100k_base x 1.05 based on average English text comparisons. For production billing, always count with the official SDK.
Subscribe on YouTube for weekly Claude, GPT, and agent tutorials.

New tutorials, open-source projects, and deep dives on coding agents - delivered weekly.