Your AI agent wastes 70% of tokens reading irrelevant files. Sverklo indexes your codebase locally, so agents find the right code instantly — fewer tokens, faster results, better answers.
The problem
Features
Hybrid retrieval with real ONNX embeddings, graph-based ranking, and session memory. Everything runs locally.
BM25 text matching combined with vector semantic search, PageRank graph ranking, and reciprocal rank fusion for precise results.
all-MiniLM-L6-v2 runs locally via ONNX runtime. No API keys, no network calls, no data leaving your machine.
TypeScript, JavaScript, Python, Go, Rust, Java, C, C++, Ruby, and PHP. Tree-sitter parsing for accurate symbol extraction.
Remember decisions, patterns, and context across sessions. Git-state linking, quality scoring, and automatic staleness detection.
Auto-indexes your codebase, respects .gitignore, incremental updates with file watcher. Just run it and search.
Responses are trimmed to fit within LLM context windows. Your agent gets the most relevant code without wasting tokens.
How it works
30 files indexed in 500ms. Search results in under 50ms. All local.
File watcher detects changes, respects .gitignore
Tree-sitter extracts symbols, types, references
ONNX model creates vector embeddings locally
BM25 + vector index + PageRank graph built
Hybrid retrieval with RRF fusion, ranked results
MCP Tools
Code search and session memory designed for how AI agents work.
Open Source
All 10 tools, all features, MIT licensed. Runs locally on your machine — your code never leaves your computer.
Quick Start
Install Sverklo
Add to your AI agent
Start coding. The embedding model (~90MB) downloads automatically on first run. After that, instant.
Optional: Open the dashboard to see what's indexed