MCP Hub
Back to servers

LLM Wiki Kit

Enables creation of persistent, compounding knowledge bases using Karpathy's LLM Wiki pattern with LLM-maintained markdown wikis. Supports automated ingestion, cross-referencing, synthesis, and linting of sources as an alternative to traditional RAG systems.

glama
Stars
17
Forks
8
Updated
Apr 8, 2026
Validated
Apr 9, 2026

📚 llm-wiki-kit

Stop re-explaining your research to your AI agent every session.

License: MIT Python 3.10+


llm-wiki-kit gives your AI agent a persistent, structured memory that compounds over time. Drop PDFs, URLs, YouTube videos — your agent builds a wiki, connects the dots, and remembers everything across sessions.

Based on Karpathy's LLM Wiki pattern. Works with Claude, Codex, Cursor, Windsurf, and any MCP-compatible agent.


The Problem

Every time you start a new chat:

You: "Remember that paper on speculative decoding I shared last week?"
Agent: "I don't have access to previous conversations..."
You: *sighs, re-uploads PDF, re-explains context*

You're constantly re-teaching your agent things it should already know.

The Solution

With llm-wiki-kit, your agent maintains its own knowledge base:

You: "What did we learn about speculative decoding?"
Agent: *searches wiki* "Based on the 3 papers you've shared, the Eagle 
       architecture shows the best efficiency tradeoffs because..."

The wiki persists. Cross-references build up. Your agent gets smarter with every source you add.


⚡ Quickstart (2 minutes)

1. Install

pip install "llm-wiki-kit[all] @ git+https://github.com/iamsashank09/llm-wiki-kit.git"

2. Initialize a wiki

mkdir my-research && cd my-research
llm-wiki-kit init --agent claude

3. Connect your agent

Add to Claude Desktop config (claude_desktop_config.json):

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}
Other agents (Codex, Cursor, Windsurf)

OpenAI Codex

codex mcp add llm-wiki-kit -- llm-wiki-kit serve --root /path/to/my-research

Cursor

Add to .cursor/mcp.json:

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}

Windsurf

Add to ~/.codeium/windsurf/mcp_config.json:

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}

4. Use it

You: "Ingest this paper: raw/attention-is-all-you-need.pdf"
Agent: *creates wiki pages, cross-references concepts, updates index*

You: "Now ingest https://youtube.com/watch?v=kCc8FmEb1nY"
Agent: *extracts transcript, links to existing transformer concepts*

You: "How does the attention mechanism in the paper relate to Karpathy's explanation?"
Agent: *searches wiki, synthesizes answer from both sources*

Your agent now has persistent memory that survives across sessions.


🔥 What Makes This Different

FeatureWhy It Matters
Multi-format ingestPDFs, URLs, YouTube, markdown — just drop it in
Auto cross-referencingAgent builds [[wiki links]] between related concepts
Persistent across sessionsStart fresh chats without losing context
Full-text searchAgent finds relevant pages instantly (SQLite FTS5)
Health checkswiki_lint catches broken links, orphan pages, contradictions
Zero lock-inIt's just markdown files in a folder — view in Obsidian, VS Code, anywhere
Works with any MCP agentClaude, Codex, Cursor, Windsurf, and more

📥 Supported Sources

Your agent can ingest anything:

Drop this...Get this...
raw/paper.pdfExtracted text, page markers, metadata
https://arxiv.org/abs/...Clean article content, auto-saved to raw/
https://youtube.com/watch?v=...Full transcript with timestamps
raw/notes.mdDirect markdown ingestion

Install what you need:

pip install "llm-wiki-kit[pdf]"      # PDF support
pip install "llm-wiki-kit[web]"      # URL extraction  
pip install "llm-wiki-kit[youtube]"  # YouTube transcripts
pip install "llm-wiki-kit[all]"      # Everything

🧠 How It Works

┌─────────────────────────────────────────────────────────┐
│  YOU                                                    │
│  "Ingest this paper. How does it relate to X?"         │
└───────────────────────┬─────────────────────────────────┘
                        │
┌───────────────────────▼─────────────────────────────────┐
│  WIKI (agent-maintained)                                │
│                                                         │
│  ┌──────────────┐  ┌──────────────┐  ┌──────────────┐  │
│  │ concepts/    │  │ sources/     │  │ synthesis/   │  │
│  │ attention.md │◄─┤ paper-1.md   │──► cache.md     │  │
│  │ [[linked]]   │  │ [[linked]]   │  │ [[linked]]   │  │
│  └──────────────┘  └──────────────┘  └──────────────┘  │
│                                                         │
│  + index.md (table of contents)                        │
│  + log.md (what happened when)                         │
└───────────────────────┬─────────────────────────────────┘
                        │
┌───────────────────────▼─────────────────────────────────┐
│  RAW SOURCES (immutable)                                │
│  paper.pdf, article.html, transcript.md                 │
└─────────────────────────────────────────────────────────┘

The agent reads raw sources, writes wiki pages, and maintains the connections. You never touch the wiki directly — the agent does all the work.


🛠 Available Tools

Your agent gets these MCP tools:

ToolWhat it does
wiki_ingestProcess any source (file, URL, YouTube)
wiki_write_pageCreate or update a wiki page
wiki_read_pageRead a specific page
wiki_searchFull-text search across all pages
wiki_lintFind broken links, orphans, empty pages
wiki_statusOverview: page count, sources, recent activity
wiki_logAppend to the operation log

💡 Use Cases

Research: Feed papers into your wiki over weeks. Ask synthesis questions that span all your reading.

Technical onboarding: Ingest a codebase's docs. Your agent answers architecture questions from accumulated context.

Competitive intel: Add market reports, earnings calls, news. Agent maintains a living landscape that updates as you add more.

Learning: Watch YouTube tutorials, read blog posts. Agent builds a personalized wiki of everything you've studied.

Book notes: Ingest chapters as you read. Agent tracks characters, themes, plot threads, and connections.


🔍 Pro Tips

  • Use Obsidian to visualize your wiki's graph — it's just a folder of markdown files
  • Git init your wiki directory — get version history for free
  • Let the agent link aggressively — the value compounds in the connections
  • Run lint periodically — catches contradictions and gaps in your knowledge base
  • Start small — even 5-10 sources produce a surprisingly useful wiki

📦 Development

git clone https://github.com/iamsashank09/llm-wiki-kit
cd llm-wiki-kit
uv venv && source .venv/bin/activate
uv pip install -e ".[all]"

🙏 Credits

Based on the LLM Wiki idea by Andrej Karpathy.

📄 License

MIT — do whatever you want with it.

Reviews

No reviews yet

Sign in to write a review