v1.0 · public beta

Benchmarks
The numbers that matter in production.
Under the hood
Why not just a RAG or vector DB?
A vector DB answers "what words are most similar?" A memory engine answers "what's true right now?" Those are different queries, and they need different data structures.
Dimension
Naive RAG
"what words are most similar?"
XMem
"what's true right now?"
✕
✓
✕
✓
✕
✓
✕
✓
Integration
✕
✓
How it works
One write. One recall.
A belief engine in the middle.
Click a node on the timeline to see how memory builds up with XTrace
use cases
Our memory is the difference between an agent and a chatbot.
Support agents
Know the user's plan, their last three tickets, and the policy changes since they signed up. Stop asking "what's your order number?"
Coding agents
Remember which patterns this repo favors, what conventions were deprecated last month, and which files the user owns.
Voice agents
<80ms p50 is the latency budget you need. Facts-only recall means short prompts — fewer tokens, lower cost, faster TTFB.
Research agents
Track what's been read, what's been ruled out, and which claims are still load- bearing. Supersede chains = reproducible research.
Companions
Users want an AI that actually knows them. Not one that forgot their dog's name between sessions.
Workflow / ops agents
Long-running agents that span days and tools. Lineage matters here more than anywhere — XMem gives you audit + replay.
For CTOs
The memory layer you'd build if you had a year.
Every team building agents writes some version of this themselves. De-duping facts. Hand-rolled entity resolution. Deciding whose contradiction to trust. It's a year of distraction from your product.
XMem is that layer — built, battle-tested, benchmarked, compliant. You ship the agent. We ship the brain.
Homomorphically encrypted by default
Our own encrypted vector database keeps embeddings searchable while encrypted — we never see your data in the clear.
0
Zero training on your data. Ever.
Your facts are yours. Never used to train models. Export-on-exit as JSON.
Python SDK — open source.
Apache 2.0 licensed. Self-host the SDK; point it at our managed engine or your own deploy. TypeScript and Go coming soon.
Drop-in. No lock-in.
Three calls to integrate. One call to export everything. We don't hold your data hostage.
Pricing
Start free. Scale when it works.
Every plan includes every feature. You're only paying for volume.
Priced on
adds
(events written) and
recalls
(context requests). Everything else — facts, episodes, artifacts, supersede chains — is just how we think about what's inside.
Same engine, two surfaces
Your agent has memory. Your whole team should too.
Most XMem customers ship their agent, then roll memory out to the rest of the team. That's Memory Hub: the same belief engine, surfaced as a product humans can use.
you are here
XMem
Memory for your agent
A drop-in SDK. Your code, your agent, your stack. One engine per application.
·
Python SDK (TS + Go coming soon)
·
Fact-level belief revision
·
Scoped per agent · per user
·
$0 → $199 → custom
