For Beginners
What is this tool?
Section titled “What is this tool?”Polyglot MCP is a translation server that runs on your own computer. It connects to AI coding assistants like Claude Code or Claude Desktop and lets them translate text between 57 languages automatically. When you ask Claude to translate something, Polyglot handles it behind the scenes using a Google translation model called TranslateGemma running on your GPU through a program called Ollama.
The key difference from cloud translation services (Google Translate, DeepL, etc.) is that nothing leaves your machine. Your text stays local, there are no API keys to manage, no usage limits, and no subscription fees.
Who is this for?
Section titled “Who is this for?”- Developers who need to translate README files, documentation, or UI strings as part of their workflow
- Teams who want translation integrated into their AI coding assistant without sending text to third-party services
- Anyone using Claude Code or Claude Desktop who wants local, private translation
You do not need to be a machine learning expert. If you can install a program and run a terminal command, you can use Polyglot MCP.
Prerequisites
Section titled “Prerequisites”Before you start, you need:
- Node.js 18 or newer — Download from nodejs.org. Run
node --versionto check. - Ollama — Download from ollama.com. This is the program that runs the translation model on your GPU.
- A GPU with at least 4 GB of VRAM — The smallest model (4B) needs about 3.3 GB. The recommended model (12B) needs about 8.1 GB. An NVIDIA GPU is recommended, but Ollama also supports AMD and Apple Silicon.
- An MCP client — Claude Code or Claude Desktop. This is the AI assistant that talks to Polyglot MCP.
- Basic terminal skills — You need to be able to open a terminal and run commands.
Your first 5 minutes
Section titled “Your first 5 minutes”Step 1: Start Ollama. Open a terminal and run:
ollama serveIf Ollama is already running (e.g., as a system service), you can skip this.
Step 2: Add Polyglot to your MCP client. For Claude Code, add this to your .mcp.json file:
{ "mcpServers": { "polyglot": { "command": "npx", "args": ["-y", "@mcptoolshop/polyglot-mcp"] } }}Step 3: Ask Claude to translate something. In Claude Code or Claude Desktop, type a message like:
Translate “Hello, how are you?” to Japanese
Claude will automatically use the translate tool. The first time, it downloads the TranslateGemma 12B model (about 8.1 GB), which takes a few minutes. After that, translations complete in under a second.
Step 4: Try markdown translation. Give Claude a markdown document and ask it to translate the whole thing:
Translate this README to Spanish
Polyglot preserves all code blocks, links, badges, and table formatting — only the prose gets translated.
Step 5: Check your setup. If something is not working, ask Claude:
Check the polyglot translation status
This runs the check_status tool, which tells you whether Ollama is running and which models are installed.
Common mistakes
Section titled “Common mistakes”Ollama is not running. Polyglot tries to auto-start Ollama, but this can fail if Ollama is not installed or not on your PATH. Fix: run ollama serve in a separate terminal before using Polyglot.
Not enough VRAM. If the 12B model causes out-of-memory errors, switch to the smaller 4B model. Set the environment variable before starting Polyglot:
POLYGLOT_MODEL=translategemma:4b npx @mcptoolshop/polyglot-mcpSlow first translation. The first translation after starting Ollama takes about 15 seconds because the model needs to load into GPU memory. All subsequent translations are fast (~600ms). This is normal behavior, not a bug.
Wrong language code. Polyglot accepts both language codes (en, ja, zh) and full names (English, Japanese, Chinese (Simplified)). If you get an “unsupported language” error, ask Claude to list_languages to see all 57 supported options.
Cache file appears in your project. Polyglot creates a .polyglot-cache.json file next to translated files to speed up re-translations. Add it to your .gitignore if you do not want it committed.
Next steps
Section titled “Next steps”- Getting Started — Detailed install instructions and model selection
- Tools — Full parameter reference for all 5 MCP tools
- Features — Learn about caching, fuzzy matching, glossaries, and validation
- Reference — Environment variables, architecture, supported languages
Glossary
Section titled “Glossary”| Term | Definition |
|---|---|
| MCP | Model Context Protocol — a standard for connecting AI assistants to external tools |
| Ollama | A program that runs large language models locally on your computer |
| TranslateGemma | A Google translation model designed for high-quality multilingual translation |
| GPU | Graphics Processing Unit — the hardware that runs the translation model |
| VRAM | Video RAM — memory on your GPU; determines which model sizes you can run |
| Semaphore | A concurrency limiter that prevents too many simultaneous GPU requests |
| Segment cache | A local file that stores previous translations to avoid re-translating unchanged text |
| Translation memory | Fuzzy matching that reuses cached translations when source text is similar but not identical |
| NDJSON | Newline-Delimited JSON — the streaming format Ollama uses to send tokens as they are generated |
| stdio | Standard input/output — the communication channel between Polyglot MCP and the AI assistant |