Lumetra Launches Engram, an MCP-Native Memory Layer Scoring 91.6% on LongMemEval

By: PRLog
ⓘ This article is third-party content and does not represent the views of this site. We make no guarantees regarding its accuracy or completeness.
Memory that shows its work. Every recall traces back to a stored memory and a knowledge-graph edge. Bring your own model. Plug in via MCP, REST, or official SDKs.

SEATTLE - May 15, 2026 - PRLog -- Lumetra today announced the general availability of Engram, a memory layer for AI agents. After a year of invitation-only beta access, Engram is opening its doors to all developers.

TL;DR
  • 91.6% on LongMemEval-S (458/500) out of the box. Methodology and results published openly.
  • Every recall is auditable: semantic retrieval plus an automatically maintained knowledge graph, so you can see which memory and which edge produced an answer.
  • BYOM by default: frontier, open-source, or self-hosted models. No inference lock-in.
  • Plug in three ways: MCP server, REST API, or official TypeScript, Python, and Go SDKs.

What Engram Does

An agent that talked with a user last week recalls their preferences today, and surfaces the exact stored memory and graph edge that produced the answer. Engram ingests conversation, extracts atomic facts and relationships, and stores them where they can be retrieved semantically and explained structurally.

Retrieval fuses three signals: keyword (BM25), semantic vector search, and traversal of the knowledge graph. Recall doesn't fail when a question is rephrased or when the answer depends on an implicit connection between memories.

For developers, that means memory stops being a black box. The recall path is inspectable end-to-end: every retrieved fact is grounded in a stored memory; every connection is grounded in a graph edge. If a recall is wrong, you can see why.

Memory That Shows Its Work

Memory products that bundle inference and hide the retrieval path require trusting the vendor about what their system did and why. Engram's design choice is the opposite: the system shows its work. This is the difference between a memory product and memory infrastructure.

Bring Your Own Model

Engram is BYOM by default. Developers connect their preferred LLM (frontier, open-source, or self-hosted), and Engram handles extraction, storage, and retrieval. No inference lock-in, no markup on tokens you could have bought directly.

Plug In Anywhere

Engram launches with three integration paths:
  • The MCP server works with Claude.ai web, Claude Desktop, Claude Code, Cursor, Windsurf, Codex, ChatGPT, and OpenClaw out of the box.
  • The REST API provides standard HTTP endpoints for ingest, query, memory management, and usage stats.
  • Official SDKs are available for TypeScript (@lumetra/engram on npm), Python (lumetra-engram on PyPI), and Go.

Pricing

Engram launches with usage-based pricing. No per-seat, no per-project surcharges:
  • Free is for evaluation and hobby projects.
  • $29 per month covers indie developers and small teams.
  • $99 per month covers production workloads.
  • Enterprise is custom and includes dedicated support.

Paid tiers meter only on memories stored and retrievals served. There are no per-call inference fees that scale with your success, and no per-token surcharges layered on top of the model you already pay for.

A Year Behind Closed Doors

Engram spent the past year in invitation-only beta with design partner NeonBay (neonbay.ai), running in production while the team hardened the retrieval pipeline, ingest path, and recall quality. Today's launch opens that same system to every developer.

Quotes

"MCP changed the math on memory. Once a client speaks MCP, adding long-term memory is a config change instead of a rewrite. We built Engram MCP-native from day one because we think that's where the ecosystem is going."
Ben Meyerson, Co-Founder, Lumetra

"Most memory products are black boxes. You hand over your data and trust that the right thing comes back. Engram is built so every recall points to a stored memory and a graph edge. If you don't like an answer, you can see exactly where it came from."
Jacob Davis, Co-Founder, Lumetra

About Lumetra

Lumetra builds memory infrastructure for AI agents. Founded in 2025 by Ben Meyerson and Jacob Davis (previously on AWS IoT at Amazon Web Services), the company is headquartered in Seattle, WA. Engram is its first product.

Start free: https://lumetra.io
Documentation: https://lumetra.io/docs
LongMemEval methodology and results:https://lumetra.io/engram-on-longmemeval

Contact
Lumetra, LLC
contact@lumetra.io


Source: Lumetra, LLC

Read Full Story - Lumetra Launches Engram, an MCP-Native Memory Layer Scoring 91.6% on LongMemEval | More news from this source

Press release distribution by PRLog
Report this content

If you believe this article contains misleading, harmful, or spam content, please let us know.

Report this article

More News

View More

Recent Quotes

View More
Symbol Price Change (%)
AMZN  264.14
-3.08 (-1.15%)
AAPL  300.23
+2.02 (0.68%)
AMD  424.10
-25.60 (-5.69%)
BAC  49.77
-0.08 (-0.16%)
GOOG  393.32
-3.85 (-0.97%)
META  614.23
-4.20 (-0.68%)
MSFT  421.92
+12.49 (3.05%)
NVDA  225.32
-10.42 (-4.42%)
ORCL  192.95
-2.66 (-1.36%)
TSLA  422.24
-21.06 (-4.75%)
Stock Quote API & Stock News API supplied by www.cloudquote.io
Quotes delayed at least 20 minutes.
By accessing this page, you agree to the Privacy Policy and Terms Of Service.