Your AI agent wastes 70% of tokens reading irrelevant files. Sverklo indexes your codebase locally, so agents find the right code instantly — fewer tokens, faster results, better answers.
The problem
Features
Hybrid retrieval with real ONNX embeddings, graph-based ranking, and session memory. Everything runs locally.
BM25 text matching combined with vector semantic search, PageRank graph ranking, and reciprocal rank fusion for precise results.
all-MiniLM-L6-v2 runs locally via ONNX runtime. No API keys, no network calls, no data leaving your machine.
TypeScript, JavaScript, Python, Go, Rust, Java, C, C++, Ruby, and PHP. Tree-sitter parsing for accurate symbol extraction.
Remember decisions, patterns, and context across sessions. Git-state linking, quality scoring, and automatic staleness detection.
Auto-indexes your codebase, respects .gitignore, incremental updates with file watcher. Just run it and search.
Responses are trimmed to fit within LLM context windows. Your agent gets the most relevant code without wasting tokens.
How it works
30 files indexed in 500ms. Search results in under 50ms. All local.
File watcher detects changes, respects .gitignore
Tree-sitter extracts symbols, types, references
ONNX model creates vector embeddings locally
BM25 + vector index + PageRank graph built
Hybrid retrieval with RRF fusion, ranked results
MCP Tools
Code search and session memory designed for how AI agents work.
Pricing
Full-featured code search and memory, MIT licensed. Pro and Team tiers are coming for advanced needs.
MIT licensed, forever
Coming soon
Coming soon
Quick Start
Download the embedding model (one-time, ~90MB)
Start the MCP server in your project directory
Add to Claude Code (or any MCP client)
Start coding. Your agent now has deep codebase search and persistent memory.