AGPL-3.0 · open source · self-hosted

Stop explaining your codebase to your AI. Every. Single. Session.

ByteBell indexes your codebase into a knowledge graph so your AI tools actually understand your code across models, copilots, sessions, cache and memory.

// works with Claude Code Cursor Cline Codex Windsurf Zed Aider Continue Roo Code Goose OpenHands Tabby
~/your-project
# index once. remember forever.
$ bytebell boot
$ bytebell index ./
$ claude mcp add bytebell http://127.0.0.1:8080/mcp
# your AI now remembers your codebase.
# switch to cursor? context stays.
# new session tomorrow? still there.
# swap to deepseek? same quality.
proven in production

Built by infra engineers. Validated by real teams.

Not a weekend project. ByteBell is production infrastructure built by a team that shipped systems handling millions of users.

500K+
Files indexed across production codebases
live customer deployments
100+
Repositories in a single org deployment
enterprise pilots
<4%
Hallucination rate with receipt verification
vs 15-30% without indexed context
the problem

Every AI coding tool has amnesia.

New session? Zero memory. Switch tools? Everything gone. Context full? Compacted. No history. No business context. No continuity.

session reset

Your AI forgets everything between sessions

Spent an hour debugging yesterday and the AI understood the entire flow? Today it starts from scratch. 20 minutes re-explaining what you were doing.

context death

"Context left until auto-compact: 0%"

Window fills up, compaction kicks in, details are gone. Quality collapses mid-session. After 3-4 compactions, you're getting answers from summaries of summaries.

tool lock-in

Switch tools, lose everything

Claude Code for backend, Cursor for frontend, Codex for a quick fix. Each one starts blind. A better model drops? Start over. Your context is trapped in whichever tool you used last.

invisible history

No one knows why this code exists

What business requirement drove this commit? What design decisions shaped this service? Which conversation produced this AI-generated function? Nobody tracks it. Nobody can answer.

the fix

Persistent structured memory. Any tool. Any model.

Index once. Query forever. The knowledge graph survives everything — sessions, tools, models, team changes.

01

Sessions carry forward

Every file has an LLM-generated purpose, summary, and business context in Neo4j. Tomorrow's session starts where today's left off. No re-reading.

persistent
02

One graph, every tool

Claude Code → Cursor → Codex → DeepSeek. The knowledge graph doesn't care which tool or model you use. Same context through MCP, always.

tool-agnostic
03

Open source models match frontier

DeepSeek-V4 and Kimi-2.6 hit frontier accuracy when retrieval is handled. Better input = better output regardless of the model.

model-agnostic
04

Business context and design history on every file

Every file carries a businessContext explaining why it exists. Design decisions evolve with the code. AI-generated code stays traceable to the conversation that produced it.

traceability
05

Monorepo or multi-repo — same experience

Semantic entities are org-scoped, not repo-scoped. One repo or fifty, the graph connects everything through shared keywords, concepts, and contracts.

scales
economics

The math your engineering budget cares about.

Every AI session without ByteBell re-reads your codebase at full token price. With ByteBell, your AI queries structured metadata. The difference is 80%+.

without bytebell
Per session cost$6 – $10
Context usage>80%
Compactions per session3 – 4×
Monorepo accuracybaseline
Cross-repo accuracyfails
Monthly / developer (5 sessions/day) ~$1,200
with bytebell
Per session cost$0.04 – $0.20
Context usage~30%
Compactions per session0
Monorepo accuracy+40%
Cross-repo accuracyaccurate
Monthly / developer (unlimited sessions/day) ~$13
$1.2
per million tokens ingested (one-time file analysis)
$0.20
per million MCP input/output tokens
SHA-256
diff per file — re-index only what changed

Measured on production deployments across 500K+ files. Costs based on OpenRouter pricing.

architecture

No vectors. Graph + fulltext only.

We tried embeddings. Two functions named process() in different services embed to similar vectors but share zero context. Vectors flatten call graphs, inheritance, and import trees into a single point. We went a different way.

vectors

Flatten everything into a point

Cosine similarity measures token overlap, not structural relationships. Dependency graphs, inheritance chains, import trees — all lost. Retrieval precision on real codebases is consistently poor.

knowledge graph

Preserve how code actually connects

9-channel parallel search across purpose, summary, business context, functions, classes, imports, keywords, paths, and semantic fields in Neo4j. Structural relationships preserved. Fulltext, not cosine.

purpose classes functions imports keywords paths semantic integration business
quickstart

Running in five commands.

Bun + Docker. Local daemon. Binds to 127.0.0.1. No cloud, no telemetry.

step 01

Set your LLM provider

Point at OpenRouter with any model — Claude, GPT, DeepSeek, Kimi, or a local model. bytebell set openrouter-api-key sk-or-...

step 02

Boot the stack

bytebell boot — spins up Neo4j, MongoDB, and Redis in Docker. One command.

step 03

Index your codebase

bytebell index ./ — LLM analysis per file: purpose, summary, business context, classes, functions, imports. SHA-256 diff means re-indexing only processes changed files.

step 04

Connect any MCP tool

claude mcp add bytebell http://127.0.0.1:8080/mcp — your AI queries the graph instead of re-reading files. 10-15 tool calls per question.

benchmarks

Measured on SWE-bench Pro. Not vibes.

Tested across 500K+ files, 100+ repositories, multiple models.

+5%
Accuracy gain over SOTA
60%
Cost reduction
10–15
Tool calls per question
<4%
Hallucination rate

→ full comparison: Sourcegraph · Augment · GraphRAG · GitNexus

what developers actually care about

Built for how you actually work in 2026.

95% of developers use AI tools weekly. They juggle 2-4 tools. They switch models when something better drops. They need accuracy, not "almost right." ByteBell was built around these realities.

First-pass accuracy
#1 developer frustration
66% of devs waste time fixing "almost right" AI code. ByteBell's graph gives your AI precise, structured context — not a noisy 200-file dump. Hallucination drops from 15-30% to under 4%.
🔀
Switch tools freely
devs juggle 2-4 tools daily
Claude Code leads at 46% satisfaction, but devs still use Cursor, Codex, Copilot for different tasks. ByteBell's MCP graph works with all of them. Switch tools without losing a single insight.
🧪
Try new models instantly
the ai landscape shifts weekly
DeepSeek-V4 drops? Kimi-2.6 benchmarks better? Try it immediately. Your context isn't locked to any model. ByteBell serves the same graph to whichever model you point at.
🔒
Code stays on your machine
75% won't share proprietary code
Binds to 127.0.0.1. No cloud account. No telemetry. No code leaving your infra. Air-gapped deployment for teams that can't compromise. Your graph, your servers, your rules.
95%
devs use AI weekly
66%
fix "almost right" code
10–15
AI tools per developer
46%
distrust AI accuracy
75%
won't share code externally
sources: Stack Overflow 2025 Survey · Pragmatic Engineer 2026 Survey · JetBrains AI Pulse Jan 2026
pricing

Start free. Scale when ready.

No tricks. The open source version is real and complete. Paid plans add capacity and managed hosting.

Free
Individual developers
$0
forever · self-hosted
  • Index up to 1M tokens
  • All MCP tools included
  • Any model, any tool
  • Runs on your machine
  • No credit card
Clone the repo
Pro
Teams and power users
$13/user/mo
billed monthly
  • Index up to 5M tokens
  • 25M MCP input/output tokens
  • Managed hosting available
  • Conversation history
  • Priority indexing
  • Dashboard available
Custom
On-prem · air-gapped · unlimited
$250+/mo
starts at 75M token analysis · 100M MCP I/O
  • 75M token analysis (one-time ingest)
  • 100M MCP input/output tokens
  • Custom context
  • Business context persistence
  • Cross-user histories
  • Feedback loop
  • Admin and user dashboard

Your AI should remember your code.

Five commands. Persistent context. Open source. Any tool. Any model.

★ Star on GitHub