The ultimate AI token counter for your IDE.
Supports 35+ models including GPT-5, Claude 4.5, Gemini 3, DeepSeek V3, and Llama 3.
Optimized for developers building with LLMs.
LLM Tokenizer gives you instant visibility into your token usage directly within your IDE. Whether you're optimizing prompts, estimating API costs, or ensuring your context window limits aren't exceeded, LLM Tokenizer removes the guesswork.
- Check Context Limits: Know instantly if your file fits within the context window of your favorite AI model.
- Estimate Costs: Get a clear sense of input token volume before sending requests to expensive APIs.
- Optimize RAG Pipelines: Analyze folder-level token counts to better chunk your knowledge base.
Stop copying and pasting into web calculators. Get precise counts right where you code.
Features
- Real-time Token Count: View the token count of the active file in the Status Bar
- 37 AI Models: OpenAI, Anthropic, Google, xAI, DeepSeek, Meta, and more
- Folder Analysis: Right-click a folder to count tokens recursively
- Grouped Model Selection: Models organized by provider
- Persistent Preferences: Selected model is remembered across sessions
Supported Models
| Provider |
Models |
| OpenAI |
GPT-5.2, GPT-OSS 120B, GPT-4o, GPT-4o Mini, o1, o3-mini |
| Anthropic |
Claude Sonnet/Opus/Haiku 4.5, Claude 3.5 Sonnet, Claude 3 Opus/Haiku |
| Google |
Gemini 3 Flash/Pro, Gemini 2.5 Flash/Pro/Lite, Gemini 2.0/1.5 |
| xAI |
Grok 4.1 Fast, Grok 4 Fast, Grok Code Fast 1 |
| DeepSeek |
DeepSeek V3.2, DeepSeek V3 |
| Meta |
Llama 3.2, CodeLlama |
| Zhipu |
GLM 4.7, GLM 4.6, GLM 4.5 |
| Others |
Mistral Large, Qwen 2.5 Coder, Kimi K2.5, MiMo-V2-Flash, MiniMax M2.1 |
Usage
- Open a file: Token count appears in Status Bar (bottom right)
- Click Status Bar item to change model
- Right-click file/folder → Count Tokens
Accuracy Notes
| Provider |
Method |
Accuracy |
| OpenAI |
tiktoken (exact) |
~100% |
| Claude |
cl100k_base + 1.05x |
~95% |
| Gemini |
4 chars/token |
~90% |
| DeepSeek |
3.33 chars/token |
~90% |
| Others |
cl100k_base proxy |
~85-95% |
Requirements
VS Code 1.85.0+
Changelog
See CHANGELOG.md for detailed release notes.
Author: Matteo Teodori
| |