The memory layer
for AI agents.

Most AI memory tools store flat facts and search by similarity. Engram extracts structured knowledge — typed, ranked, and organized — so your agents actually understand what they know.

Open source · 3 API calls · Works with Claude, ChatGPT, and any LLM

Works with

LangGraph CrewAI Claude OpenAI

Memory is more than a key-value store.

You've probably tried the basics: stuff the chat history in the prompt, bolt on a vector search, or store flat facts in a database. It works — until it doesn't.

Context stuffing

10,000 tokens of chat history. Your agent has the information — buried in noise. Costs spike. Latency doubles. And it still misses the one fact that matters.

Basic RAG

Your retriever returns "I live in Boston" AND "I moved to Vancouver." Your agent picks one at random. The user corrects it for the third time.

Flat memory

You store 500 facts. Which ones matter right now? They all look the same. No priority, no type, no decay. Your agent dumps everything into the prompt and hopes for the best.

These approaches store information. Engram builds knowledge.

The Solution

What makes Engram different.

Not just storage and retrieval. Engram understands, organizes, and prioritizes what your agent knows.

Every memory has a type.

Engram doesn't store flat strings. Every piece of knowledge is classified — FACT, DECISION, PREFERENCE, GOAL, PROCEDURE, PRINCIPLE — so your agent knows the difference between what a user likes and what they decided.

Not all memories are equal.

Every bullet gets a salience score based on importance and recency. Critical decisions rank higher than casual mentions. Stale knowledge decays over time. Your agent always knows what matters most right now.

Fits your context window.

Tell Engram your token budget. It assembles the most important, most relevant knowledge that fits — not just the most similar. No overflow, no truncation, no wasted tokens.

Handles updates, not just appends.

"I live in Boston" followed by "I moved to Vancouver" doesn't return both. Engram detects the contradiction, supersedes the old fact, and keeps a version history.

Patterns emerge automatically.

After enough conversations, Engram detects recurring patterns and forms schemas: "user prefers concise technical answers" or "team always reviews PRs before merging." Structure emerges from usage.

One memory, every model.

Connect Engram to Claude via MCP, ChatGPT via Actions, or any agent via REST API. Tell Claude something in the morning, ask ChatGPT about it at night. Same knowledge, everywhere.

How it works

Three API calls. Structured memory.

No vector database to manage. No retrieval pipeline to tune. Your agent commits what it learns, recalls what it needs, and knowledge self-organizes over time.

01

Commit

Your agent saves what it learned.

Raw conversation goes in. Typed, structured knowledge comes out. Facts, decisions, preferences, goals — each extracted, classified, and scored for importance.

terminal
curl -X POST api.engram.so/contexts/abc/commit \
  -H "Authorization: Bearer eng_..." \
  -d '{
    "input": "User prefers dark mode,
     lives in Vancouver,
     runs an AI consultancy"
  }'

→ Extracted: 3 bullets (FACT, PREFERENCE, DECISION)

→ Superseded: 1 bullet (location: Boston → Vancouver)

→ Salience: 0.85, 0.72, 0.91

02

Materialize

Your agent recalls what it needs.

Not a similarity search. Engram assembles the most important, most relevant knowledge within your token budget. What comes back is ready to inject into any prompt.

terminal
curl -X POST api.engram.so/contexts/abc/materialize \
  -H "Authorization: Bearer eng_..." \
  -d '{
    "query": "user preferences"
  }'

→ Returned: 12 bullets, 2,400 tokens (budget: 4,000)

→ Ranked by: relevance × salience

→ Includes: 2 schemas, 0 contradictions

03

Consolidate

Knowledge self-organizes.

Over time, Engram deduplicates, merges related facts, forms schemas from recurring patterns, and decays stale knowledge. Your context stays clean no matter how many conversations flow through.

terminal
# Automatic background process
# Dedup, merge, schema formation,
# salience decay, contradiction resolution

→ 500 conversations → 47 active bullets, 3 schemas

→ 12 duplicates merged, 2 contradictions resolved

→ 8 stale bullets archived (salience < 0.05)

Not just another RAG

Beyond search. Structured understanding.

RAG finds similar chunks. Flat memory stores facts. Engram organizes knowledge.

 
RAG
Flat Memory
Engram
Storage
Chunks raw text
Key-value facts
Typed bullets (FACT, DECISION, etc.)
Retrieval
Similarity search
Similarity search
Salience + relevance within token budget
Duplicates
Pile up forever
Overwrite or duplicate
Merge, deduplicate, link related
Contradictions
Returns both old and new
Last-write-wins
Detects and resolves with version history
Over time
Grows linearly
Grows linearly
Stabilizes via consolidation + decay
Structure
None — raw chunks
Flat list
Typed bullets, schemas, concept graph
Cross-platform
Model-specific embeddings
Depends
Model-agnostic by design

Integrations

Start with Claude. Continue with ChatGPT. Nothing is lost.

Engram is model-agnostic by design. Your users build up knowledge in one AI, and it's available everywhere. Same memory, every platform.

Morning · Claude

"We decided on a freemium model with usage-based Pro."

↓ commit

Engram

DECISION: freemium + usage-based Pro

salience: 0.91

Evening · ChatGPT

"What pricing model did we discuss?"

↓ materialize

claude-mcp
{
  "mcpServers": {
    "engram": {
      "command": "npx",
      "args": ["-y", "engram-mcp"],
      "env": {
        "ENGRAM_API_URL": "https://api.engram.so",
        "ENGRAM_API_KEY": "eng_..."
      }
    }
  }
}

Pricing

Free to start. Open source forever.

Self-host the full engine for free. Use the cloud for managed hosting and dashboards.

Free

$0 /month

Everything you need to add memory to your AI agent.

  • 3 contexts
  • 1,000 bullets per context
  • BYOK (bring your own LLM keys)
  • Community support
  • Dashboard access
  • Claude MCP + ChatGPT Actions
Get Started

Enterprise

Custom

Everything in Free, plus power features and dedicated infrastructure.

  • Unlimited contexts
  • 50,000 bullets per context
  • Bullet editing & management
  • Full graph explorer
  • Dedicated tenant + data isolation
  • Custom SLAs + dedicated support
  • SSO / SAML + Audit logs
Talk to Sales

All plans include the open-source SDK. Self-host for free, forever.

Open Source

Fully open source. Not open-core.

The complete Engram engine — extraction, consolidation, materialization, concept graph — is MIT licensed. Self-host it, fork it, extend it. The cloud adds managed hosting and a dashboard, but the engine is the same.

$ pip install engram
MIT License Python

Add memory to your agent in 5 minutes.

Free tier. No credit card. Open source.