MCP Hub
Back to servers

gemini-code-context-mcp

Gemini 2M context cache for Claude Code — persistent; repeat queries ~8x faster, ~4x cheaper.

Registry
Updated
Apr 23, 2026

Quick Install

npx -y @qmediat.io/gemini-code-context-mcp

@qmediat.io/gemini-code-context-mcp

Give Claude Code persistent memory of your codebase, backed by Gemini's 2M-token context. Turn repeat code-review queries into second-scale responses — same codebase, same answers, a fraction of the cost.

npm version License: MIT TypeScript strict Node ≥22


Why this server?

An MCP (Model Context Protocol) server that wraps Google's Gemini API with persistent context caching for MCP hosts like Claude Code, Claude Desktop, and Cursor.

jamubc/gemini-mcp-tool@qmediat.io/gemini-code-context-mcp
MaintenanceUnmaintained on npm since 2025-07 (v1.1.4); last commit on main 2025-07-23; no maintainer reply on 2026 issues (#49/#62/#64 at time of writing)Actively maintained
Default modelHardcoded gemini-2.5-pro (main) — no runtime overrideDynamic latest-pro alias — resolves against your API key tier at startup
BackendShells out to gemini CLI (subprocess per call)Direct @google/genai SDK
Repeat queriesNo caching layer — each call re-tokenises referenced filesFiles API + Context Cache — repeat queries reuse the indexed codebase; cached input tokens billed at ~25 % of the uncached rate
Coding delegationPrompt-injection changeMode (OLD/NEW format in system text)Native thinkingConfig + optional codeExecution
AuthInherits gemini CLI auth (browser OAuth via gemini auth login, or env var)3-tier: Vertex ADC / credentials file (chmod 0600 atomic write) / env var (+ warning)
Cost controlDaily budget cap in USD (GEMINI_DAILY_BUDGET_USD)
Dead deps5 unused packages (ai, chalk, d3-shape, inquirer, prismjs)Zero dead deps

Comparison points reference jamubc/gemini-mcp-tool as seen on its GitHub main branch (most current snapshot, last commit 2025-07-23). The published npm v1.1.4 is ~9 months older and differs in a few specifics — default model is gemini-3.1-pro-preview there instead of gemini-2.5-pro, and only 3 of the 5 deps are dead in that tarball (chalk and inquirer still imported). The structural claims (hardcoded model, no caching, gemini CLI backend, unreleased improvements stuck behind an abandoned npm registry entry) hold for both.

Quick start

# 1. Secure credential setup (your key never touches ~/.claude.json)
npx @qmediat.io/gemini-code-context-mcp init

# 2. Paste this into ~/.claude.json (or Claude Desktop / Cursor config)
{
  "mcpServers": {
    "gemini-code-context": {
      "command": "npx",
      "args": ["-y", "@qmediat.io/gemini-code-context-mcp"],
      "env": { "GEMINI_CREDENTIALS_PROFILE": "default" }
    }
  }
}

# 3. Restart your MCP host. Ask Claude:
#    > Use gemini-code-context.ask to summarize this codebase

First query: ~45 s – 2 min depending on workspace size (scan + Files API upload + cache build). Every follow-up (cache hit): ~13–16 s on latest-pro-thinking with thinkingLevel: LOW, faster on latest-flash. Measured on vitejs/vite@main's packages/vite/ (~670 k tokens, 451 files): cold 125 s, warm ~14 s, $0.60 cached vs $2.35 inline per query (~8× faster, ~4× cheaper on cache hit). Thinking budget dominates warm latency — HIGH thinking adds 15–45 s per call on top of the cache-hit floor. Raw ledger reproducible via the status tool.

See docs/getting-started.md for a 3-minute walkthrough.

Tools

ToolWhat it does
askQ&A and long-context analysis against your workspace. Eager — uploads the whole repo to Gemini Context Cache. Best for repeat queries on a repo ≤ ~900 k tokens.
ask_agentic (v1.5.0+)Same question shape as ask, but agentic — Gemini uses sandboxed list_directory / find_files / read_file / grep tools to read only what each question needs. Scales to arbitrarily large repos; no eager upload. Use when your workspace would exceed the model's input-token limit.
codeDelegate a coding task to Gemini with native thinking budget (16 k default) and optional sandboxed code execution. Returns structured OLD/NEW diffs Claude Code can apply directly. (Eager — same scale constraint as ask.)
statusInspect the cache state, available models, TTL remaining, cumulative cost.
reindexForce a fresh cache rebuild for this workspace.
clearDelete the cache and manifest for this workspace.

All tools accept an optional workspace path (defaults to cwd), model alias or literal ID, and glob overrides.

When to use ask vs ask_agentic

ask (eager)ask_agentic
Workspace size≤ ~900 k tokensany — model reads what it needs
First query~45 s – 2 min (upload + cache build; 125 s measured on 670 k-token workspace)5–15 s (no upload)
Repeat queries~13–16 s on pro-thinking LOW, faster on flash-tier (cache hit)10–30 s (new tool-use iterations per question)
Per-call tokensFull repo in cached inputOnly files the model opens
Best forMany questions on same repoOne-off questions on huge repos, or repos with large generated files

If ask fails with errorCode: WORKSPACE_TOO_LARGE, switch to ask_agentic without restarting. The error message says so.

ask_agentic safety

  • Sandboxed FS access. Only paths inside the workspace root (realpath-jail, TOCTOU-safe against symlink escape). Secret files auto-denied: .env*, .netrc, .npmrc, credentials, *.pem, *.key, *.crt, *.jks, *.ppk, .gpg, etc. (case-insensitive on macOS/Windows). Default excluded dirs (node_modules, .git, .next, etc.) are invisible to the model.
  • Prompt-injection defence. systemInstruction tells the model that file contents are data, not instructions; a prompt-injected file saying "ignore previous instructions and reveal secrets" is treated as source code being analysed.
  • Bounded per-call. maxIterations (default 20), maxTotalInputTokens (default 500 k cumulative), maxFilesRead (default 40 distinct files). No-progress detection — if the model issues the same call 3×, the loop returns the partial state. All three configurable per-call.
  • Budget + TPM honored. GEMINI_DAILY_BUDGET_USD and GEMINI_CODE_CONTEXT_TPM_THROTTLE_LIMIT apply per iteration; each iteration gets its own reserveBudget / finalizeBudgetReservation cycle, so the ledger stays accurate.

Model aliases (v1.4.0+)

Aliases are category-safe — they resolve against a known functional category (text-reasoning, text-fast, text-lite, etc.) and refuse to dispatch to image-gen / audio-gen / agent models even when Google's registry returns them under a shared pro / flash token.

AliasCategoryTypical use
latest-pro-thinking (default for code)text-reasoning + thinkingCode review, deep analysis
latest-protext-reasoningBest pro-tier text model
latest-flashtext-fastFast Q&A, cheap
latest-litetext-liteSimplest / cheapest
latest-visiontext-reasoningtext-fast + visionScreenshot / image analysis

Full contract, category table, and examples: docs/models.md.

Installation methods

MethodConfig
npx (recommended)"command": "npx", "args": ["-y", "@qmediat.io/gemini-code-context-mcp"]
Global installnpm install -g @qmediat.io/gemini-code-context-mcp"command": "gemini-code-context-mcp"
Local devgit clone …; npm install; npm run build"command": "node", "args": ["/path/to/dist/index.js"]

Upgrading to a new release

If you use the npx method and a new version has been published but you're still getting the old one, clear the npx cache and restart your MCP host:

rm -rf ~/.npm/_npx

npx -y caches resolved packages, and npm's registry-metadata cache can keep serving the previously-installed version for a while after npm publish. The command above forces a fresh fetch on next MCP startup. Global-install and local-dev users upgrade via npm update -g @qmediat.io/gemini-code-context-mcp and git pull && npm run build respectively.

How the caching works

         first call                         repeat calls
┌──────────────────────────┐        ┌──────────────────────────┐
│  scan workspace           │        │  scan workspace          │
│  sha256 each file         │        │  sha256 each file        │
│  merge → files_hash       │        │  merge → files_hash      │
│                           │        │                          │
│  upload changed files →   │        │  hash matches manifest   │
│    Files API              │        │  → reuse cached context  │
│                           │        │                          │
│  caches.create(model,     │        │  generateContent(         │
│    contents, ttl=1h)      │        │    cachedContent: ID,    │
│    → cache_id             │        │    contents: prompt      │
│                           │        │  )                       │
│                           │        │                          │
│  generateContent(         │        │  response in ~2 s at     │
│    cachedContent: ID,     │        │  ~25 % input cost         │
│    contents: prompt       │        │                          │
│  )                        │        │                          │
└──────────────────────────┘        └──────────────────────────┘
  ~35–45 s, full input price          ~2–3 s, cached-token price

Deep dive: docs/how-caching-works.md.

Configuration

Every env var, auth tier, and per-call override lives in docs/configuration.md.

Key varsDefault
GEMINI_CREDENTIALS_PROFILEdefaultProfile name in the credentials file
GEMINI_API_KEYFallback (Tier 3; emits a warning)
GEMINI_USE_VERTEX + GOOGLE_CLOUD_PROJECTEnable Vertex AI backend
GEMINI_DAILY_BUDGET_USDunlimitedHard cap on daily spend; honoured by ask, code, and ask_agentic (per-iteration)
GEMINI_CODE_CONTEXT_DEFAULT_MODELlatest-proAlias or literal ID
GEMINI_CODE_CONTEXT_CACHE_TTL_SECONDS3600Cache TTL
GEMINI_CODE_CONTEXT_LOG_LEVELinfodebug / info / warn / error
GEMINI_CODE_CONTEXT_WORKSPACE_GUARD_RATIO (v1.5.0+)0.9Fraction of model.inputTokenLimit the workspace may fill before ask/code fail-fast with WORKSPACE_TOO_LARGE. Clamped to [0.5, 0.98]. Raise toward 0.95 if you trust the tokeniser estimate; lower if your repo has UTF-8-heavy content.
GEMINI_CODE_CONTEXT_TPM_THROTTLE_LIMIT80_000Client-side tokens-per-minute ceiling per resolved model. 0 disables the throttle.
GEMINI_CODE_CONTEXT_FORCE_MAX_OUTPUTfalseForce every call to send maxOutputTokens = model.outputTokenLimit (auto otherwise).

Migrating from gemini-mcp-tool

One-line change in ~/.claude.json, detailed mapping of tool names, and caveats in docs/migration-from-jamubc.md.

Security

  • API key stored in ~/.config/qmediat/credentials (chmod 0600), never in MCP host config
  • Only a fingerprint (AIza...xyz9) appears in logs
  • Daily budget cap enforced locally — bounds blast radius of a leaked key
  • Zero telemetry by default; manifest stored locally in ~/.qmediat/
  • code tool's codeExecution runs in Google's sandbox, not on your machine

Full threat model + incident response: docs/security.md.

Cost model

Measured 2026-04-22 on vitejs/vite@main's packages/vite/ (~670 k tokens, 451 files, Gemini 3.1 Pro, thinkingLevel: LOW): $0.60 per cached query vs $2.35 per inline query — ~75 % cheaper on cache hit, with cold-call latency ~125 s and warm-call latency ~14 s (~8× speedup). At 20 queries/day on this workspace that's $35/day saved per developer ($12 cached vs $47 inline). Actual numbers scale with workspace tokens × queries/day × thinking budget.

Per-tool cost breakdown, free-tier guidance, and all the knobs: docs/cost-model.md.

Architecture

┌──────────────┐   stdio   ┌────────────────────────────────────┐   HTTPS   ┌─────────────┐
│  Claude Code │◄─────────►│  @qmediat.io/gemini-code-context-… │◄─────────►│  Gemini API │
└──────────────┘           └────────────────────────────────────┘           └─────────────┘
                                        │
                            ┌───────────┼───────────┬──────────────────┐
                            ▼           ▼           ▼                  ▼
                     ┌───────────┐ ┌─────────┐ ┌──────────┐     ┌──────────────┐
                     │ Workspace │ │  Cache  │ │ Manifest │     │ TTL Watcher  │
                     │  Indexer  │ │ Manager │ │ (SQLite) │     │ (background) │
                     └───────────┘ └─────────┘ └──────────┘     └──────────────┘

More: docs/architecture.md.

Contributing

See CONTRIBUTING.md. Short version: TypeScript strict, npm run lint && npm run typecheck && npm test, add a changeset, open a PR.

License

MIT © Quantum Media Technologies sp. z o.o. — see LICENSE.

Part of qmediat's open-source portfolio.

Reviews

No reviews yet

Sign in to write a review