Skip to Content
FAQ

FAQ

General

What is Engram?

Engram is a memory service for AI agents. It stores complete, uncompressed conversation transcripts and makes them searchable via semantic search. It’s accessible through the Model Context Protocol (MCP).

How is Engram different from Mem0, Zep, or Supermemory?

Those products extract and compress conversations into distilled “memories.” Engram stores the full, verbatim transcript. When you search Engram, you get back the actual conversation — not a summary.

This matters because summaries lose context. A memory like “user prefers TypeScript” doesn’t capture why they prefer it, what they’ve tried before, or the nuances of their preference. The full conversation does.

What is MCP?

The Model Context Protocol  is an open standard for connecting AI models to external tools and data sources. Engram implements MCP, so it works with any MCP-compatible client — Claude Desktop, Claude Code, Cursor, Windsurf, and custom agents.

Data and Privacy

Where is my data stored?

On Cloudflare’s network. Conversations and messages are stored in D1 (SQLite at the edge). Search embeddings are stored in Vectorize. Both are managed by Cloudflare with automatic replication.

Is my data encrypted?

Data is encrypted at rest and in transit by Cloudflare. API keys are hashed with SHA-256 before storage — the raw key is never stored.

Can one organization see another’s data?

No. Every query is scoped by organization_id. The ID is determined by the API key used, and all database queries and vector searches filter by it. There is no API to query across organizations.

Can I export my data?

You can retrieve all conversations via list_conversations and get_conversation with pagination. Full export tooling is planned for a future release.

Can I delete all my data?

Delete conversations individually with delete_conversation. This removes the conversation, all messages, all chunks, and all vector embeddings. Bulk deletion is planned for a future release.

Technical

What embedding model does Engram use?

@cf/baai/bge-base-en-v1.5 — a 768-dimensional embedding model that runs on Cloudflare Workers AI at no cost. It’s optimized for semantic similarity tasks.

How does semantic search work?

When messages are appended, they’re grouped into overlapping chunks (windows of 5 messages). Each chunk is embedded into a 768-dimensional vector. When you search, your query is embedded with the same model and compared against all stored vectors using cosine similarity. The most similar chunks are returned with their original messages.

What are the size limits?

ResourceLimit
D1 storage5GB free, 10GB paid
Vectorize vectors5M
Messages per append200
Search results50 max
Message contentNo hard limit (practical limit ~100KB)

Typically under 100ms for the full pipeline: embedding the query (~20ms), vector search (~10ms), fetching chunks and messages from D1 (~20ms). Actual latency depends on the user’s distance from the nearest Cloudflare data center.

Does Engram support streaming?

The MCP transport uses Streamable HTTP, which supports server-sent events. However, Engram’s tool responses are complete JSON objects — there’s no token-by-token streaming of results.

Can I use Engram without MCP?

Not yet. The current MVP only exposes MCP tools. A REST API is planned for Phase 2.

Pricing

Is there a free tier?

Pricing is not yet announced.

Last updated on