Skip to content

iamsashank09/llm-wiki-kit

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

14 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸ“š llm-wiki-kit

Stop re-explaining your research to your AI agent every session.

License: MIT Python 3.10+


llm-wiki-kit gives your AI agent a persistent, structured memory that compounds over time. Drop PDFs, URLs, YouTube videos β€” your agent builds a wiki, connects the dots, and remembers everything across sessions.

Based on Karpathy's LLM Wiki pattern. Works with Claude, Codex, Cursor, Windsurf, and any MCP-compatible agent.

llm-wiki-kit.mp4.mp4

The Problem

Every time you start a new chat:

You: "Remember that paper on speculative decoding I shared last week?"
Agent: "I don't have access to previous conversations..."
You: *sighs, re-uploads PDF, re-explains context*

You're constantly re-teaching your agent things it should already know.

The Solution

With llm-wiki-kit, your agent maintains its own knowledge base:

You: "What did we learn about speculative decoding?"
Agent: *searches wiki* "Based on the 3 papers you've shared, the Eagle 
       architecture shows the best efficiency tradeoffs because..."

The wiki persists. Cross-references build up. Your agent gets smarter with every source you add.


⚑ Quickstart (2 minutes)

1. Install

pip install "llm-wiki-kit[all] @ git+https://github.com/iamsashank09/llm-wiki-kit.git"

2. Initialize a wiki

mkdir my-research && cd my-research
llm-wiki-kit init --agent claude

3. Connect your agent

Add to Claude Desktop config (claude_desktop_config.json):

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}
Other agents (Codex, Cursor, Windsurf)

OpenAI Codex

codex mcp add llm-wiki-kit -- llm-wiki-kit serve --root /path/to/my-research

Cursor

Add to .cursor/mcp.json:

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}

Windsurf

Add to ~/.codeium/windsurf/mcp_config.json:

{
  "mcpServers": {
    "llm-wiki-kit": {
      "command": "llm-wiki-kit",
      "args": ["serve", "--root", "/path/to/my-research"]
    }
  }
}

4. Use it

You: "Ingest this paper: raw/attention-is-all-you-need.pdf"
Agent: *creates wiki pages, cross-references concepts, updates index*

You: "Now ingest https://youtube.com/watch?v=kCc8FmEb1nY"
Agent: *extracts transcript, links to existing transformer concepts*

You: "How does the attention mechanism in the paper relate to Karpathy's explanation?"
Agent: *searches wiki, synthesizes answer from both sources*

Your agent now has persistent memory that survives across sessions.


πŸ”₯ What Makes This Different

Feature Why It Matters
Multi-format ingest PDFs, URLs, YouTube, markdown β€” just drop it in
Auto cross-referencing Agent builds [[wiki links]] between related concepts
Persistent across sessions Start fresh chats without losing context
Full-text search Agent finds relevant pages instantly (SQLite FTS5)
Health checks wiki_lint catches broken links, orphan pages, contradictions
Graph visualization wiki_graph generates an interactive HTML map of your knowledge (see below)
Zero lock-in It's just markdown files in a folder β€” view in Obsidian, VS Code, anywhere
Works with any MCP agent Claude, Codex, Cursor, Windsurf, and more

πŸ“₯ Supported Sources

Your agent can ingest anything:

Drop this... Get this...
raw/paper.pdf Extracted text, page markers, metadata
https://arxiv.org/abs/... Clean article content, auto-saved to raw/
https://youtube.com/watch?v=... Full transcript with timestamps
raw/notes.md Direct markdown ingestion

Install what you need:

pip install "llm-wiki-kit[pdf]"      # PDF support
pip install "llm-wiki-kit[web]"      # URL extraction  
pip install "llm-wiki-kit[youtube]"  # YouTube transcripts
pip install "llm-wiki-kit[all]"      # Everything

🧠 How It Works

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚  YOU                                                    β”‚
β”‚  "Ingest this paper. How does it relate to X?"         β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                        β”‚
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β–Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚  WIKI (agent-maintained)                                β”‚
β”‚                                                         β”‚
β”‚  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”  β”‚
β”‚  β”‚ concepts/    β”‚  β”‚ sources/     β”‚  β”‚ synthesis/   β”‚  β”‚
β”‚  β”‚ attention.md │◄── paper-1.md   │──► cache.md     β”‚  β”‚
β”‚  β”‚ [[linked]]   β”‚  β”‚ [[linked]]   β”‚  β”‚ [[linked]]   β”‚  β”‚
β”‚  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜  β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜  β”‚
β”‚                                                         β”‚
β”‚  + index.md (table of contents)                        β”‚
β”‚  + log.md (what happened when)                         β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                        β”‚
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β–Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚  RAW SOURCES (immutable)                                β”‚
β”‚  paper.pdf, article.html, transcript.md                 β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

The agent reads raw sources, writes wiki pages, and maintains the connections. You never touch the wiki directly β€” the agent does all the work.


πŸ“Š Knowledge Graph

wiki_graph generates an interactive HTML visualization of your wiki's structure:

llm-wiki-kit-graph

Nodes are color-coded by type (sources, concepts, synthesis). Click and drag to explore connections.


πŸ›  Available Tools

Your agent gets these MCP tools:

Tool What it does
wiki_ingest Process any source (file, URL, YouTube)
wiki_write_page Create or update a wiki page
wiki_read_page Read a specific page
wiki_search Full-text search across all pages
wiki_lint Find broken links, orphans, empty pages
wiki_status Overview: page count, sources, recent activity
wiki_log Append to the operation log
wiki_graph Generate interactive HTML graph visualization

πŸ’‘ Use Cases

Research: Feed papers into your wiki over weeks. Ask synthesis questions that span all your reading.

Technical onboarding: Ingest a codebase's docs. Your agent answers architecture questions from accumulated context.

Competitive intel: Add market reports, earnings calls, news. Agent maintains a living landscape that updates as you add more.

Learning: Watch YouTube tutorials, read blog posts. Agent builds a personalized wiki of everything you've studied.

Book notes: Ingest chapters as you read. Agent tracks characters, themes, plot threads, and connections.


πŸ” Pro Tips

  • Use Obsidian to visualize your wiki's graph β€” it's just a folder of markdown files
  • Git init your wiki directory β€” get version history for free
  • Let the agent link aggressively β€” the value compounds in the connections
  • Run lint periodically β€” catches contradictions and gaps in your knowledge base
  • Start small β€” even 5-10 sources produce a surprisingly useful wiki

πŸ“¦ Development

git clone https://github.com/iamsashank09/llm-wiki-kit
cd llm-wiki-kit
uv venv && source .venv/bin/activate
uv pip install -e ".[all]"

πŸ™ Credits

Based on the LLM Wiki idea by Andrej Karpathy.

πŸ“„ License

MIT β€” do whatever you want with it.

About

Stop re-explaining your research to your AI agent. Persistent, LLM-maintained wikis that compound over time. Drop PDFs, URLs, YouTube - your agent remembers forever. Based on Karpathy's LLM Wiki pattern.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages