Listen to today's AI briefing

Daily podcast — 5 min, AI-narrated summary of top stories

MemPalace Hits 96.6% on LongMemEval, Beats Paid AI Memory Tools

MemPalace Hits 96.6% on LongMemEval, Beats Paid AI Memory Tools

MemPalace, an open-source AI memory system built by actress Milla Jovovich and developer Ben Sigman, achieved 96.6% on the LongMemEval benchmark—the highest local-only score ever recorded—using a memory palace architecture that stores all conversations verbatim.

GAla Smith & AI Research Desk·6h ago·6 min read·5 views·AI-Generated
Share:
MemPalace Hits 96.6% on LongMemEval, Beats Paid AI Memory Tools

An open-source AI memory system called MemPalace has achieved what established commercial services have struggled to deliver: persistent, accurate memory for AI conversations without cloud dependencies or subscription fees. Built by Resident Evil actress Milla Jovovich and developer Ben Sigman using Claude Code, the project reached 35,000 GitHub stars in five days and delivers state-of-the-art performance on memory benchmarks while running entirely locally.

What MemPalace Actually Does

MemPalace solves a fundamental frustration with current AI tools: every conversation disappears when the session ends. Months of debugging sessions, architecture decisions, and project context vanish, forcing users to repeatedly re-explain context. Jovovich reportedly built MemPalace after growing frustrated that AI tools kept forgetting her.

The system implements a digital version of the ancient memory palace technique—a mnemonic strategy used by memory champions to remember vast amounts of information. Every project gets a "wing," every topic gets a "room," and every idea gets a "drawer." This organizational structure alone improves retrieval accuracy by 34% over flat search, according to the developers.

Technical Architecture

MemPalace stores all conversations verbatim in ChromaDB with no summarization or extraction, ensuring nothing is lost. The system features:

  • 4-layer memory system with a 170-token wake-up cost (your AI loads months of memory in 170 tokens)
  • Knowledge graph with temporal validity where facts have expiry dates—the system knows what was true then versus what's true now
  • Auto-save every 15 messages so nothing disappears into chat history
  • 19 MCP tools for compatibility with Claude Code, ChatGPT, Cursor, and Gemini CLI
  • AAAK compression dialect providing 30x lossless shorthand that works with any LLM
  • Automatic cross-references between projects, enabling AI to connect dots users might miss

Installation requires one command: pip install mempalace.

Benchmark Performance: 96.6% on LongMemEval

The most striking claim is MemPalace's performance on the LongMemEval benchmark: 96.6% accuracy on 500 questions with zero API calls, no cloud dependency, and no subscription. According to the developers, this score was independently verified by community members running on an M2 Ultra in under five minutes.

This represents the highest local-only score ever published—free or paid. For comparison:

MemPalace 96.6% Free, open-source Local, verbatim storage Mem0 ~85% $19-$249/month Cloud, AI-filtered Zep ~85% $25/month Cloud, AI-filtered

Both Mem0 and Zep use AI to decide what to remember, which inevitably leads to information loss. MemPalace's approach of storing everything verbatim combined with its memory palace organizational structure appears to deliver superior retrieval accuracy.

The Development Story

What makes this story unusual is its origin: a Hollywood actress partnering with a developer to build a technical solution using Claude Code. Milla Jovovich's frustration with AI tools forgetting her led to months of collaboration with Ben Sigman, resulting in a fully open-sourced solution under the MIT License.

The project's rapid GitHub traction—35,300 stars and 4,400 forks in its first week—suggests it addresses a widespread pain point among developers and AI practitioners.

gentic.news Analysis

MemPalace arrives at a critical inflection point in the AI memory landscape. For years, the dominant approach has been cloud-based services that use AI to filter and summarize memories—an approach that inevitably loses information and creates vendor lock-in. MemPalace's local-first, verbatim-storage architecture represents a fundamentally different paradigm that prioritizes completeness over efficiency.

This development aligns with several trends we've been tracking: the rise of local AI (as seen in Ollama's growth to 75k+ stars), increasing frustration with context window limitations despite models with 1M+ token contexts, and the democratization of AI toolbuilding through coding assistants like Claude Code. What's particularly notable is how MemPalace achieves state-of-the-art performance not through better AI models, but through better information architecture—the 34% improvement from the memory palace structure alone demonstrates that organizational intelligence can sometimes outperform raw computational power.

The timing is significant. Just last month, we covered Anthropic's 1M Context Window Rollout, which highlighted how even massive context windows don't solve the memory problem if information isn't organized for retrieval. MemPalace directly addresses this gap. Additionally, this follows Mem0's $2.3M Seed Round in January 2026, showing investor interest in AI memory solutions—interest that may now shift toward local, open-source alternatives.

For practitioners, MemPalace represents more than just another tool—it's a proof concept that local, open-source solutions can outperform venture-backed cloud services on core metrics. The 96.6% benchmark score, if independently verified, could trigger a reevaluation of whether AI memory should be a cloud service at all. The MIT license and simple pip installation lower adoption barriers dramatically, potentially accelerating a shift toward user-controlled AI memory architectures.

Frequently Asked Questions

How does MemPalace achieve 96.6% on LongMemEval without cloud AI?

MemPalace uses a combination of verbatim storage (nothing is lost to summarization) and a memory palace organizational structure that improves retrieval accuracy by 34% over flat search. The system stores everything in ChromaDB locally and uses efficient indexing rather than AI filtering to retrieve information, eliminating the information loss inherent in summarization-based approaches.

Is MemPalace really built by Milla Jovovich?

According to the source material, actress Milla Jovovich partnered with developer Ben Sigman to build MemPalace using Claude Code. Jovovich reportedly initiated the project after frustration with AI tools forgetting her. The code is publicly available on GitHub under the MIT License with 35,300+ stars, confirming substantial community validation of the implementation.

How does MemPalace compare to Mem0 and Zep?

MemPalace scores 96.6% on LongMemEval versus approximately 85% for both Mem0 and Zep. Unlike these paid services ($19-$249/month), MemPalace is free, open-source, and runs entirely locally without cloud dependencies. The key architectural difference is that MemPalace stores conversations verbatim while Mem0 and Zep use AI to filter and summarize memories, which causes information loss.

What is the "170-token wake-up cost" mentioned?

The wake-up cost refers to how many tokens are needed to load months of memory into an LLM's context. MemPalace's 4-layer memory system and efficient organization enable it to load comprehensive memory in just 170 tokens, making it practical to use with any LLM without consuming massive context windows.

Can MemPalace work with any LLM?

Yes, MemPalace includes 19 MCP tools for compatibility with Claude Code, ChatGPT, Cursor, and Gemini CLI. The AAAK compression dialect provides 30x lossless shorthand that works with any LLM, and the system stores conversations in a model-agnostic format in ChromaDB.

Following this story?

Get a weekly digest with AI predictions, trends, and analysis — free.

AI Analysis

MemPalace represents a significant challenge to the prevailing commercial AI memory paradigm. For years, startups like Mem0 and Zep have argued that AI filtering and summarization are necessary to manage memory at scale—MemPalace proves otherwise. By combining verbatim storage with intelligent organization (the memory palace structure), the system achieves superior accuracy while eliminating the information loss inherent in summarization approaches. This development has broader implications for the AI infrastructure market. The rapid GitHub traction (35k+ stars in 5 days) suggests strong developer demand for local, open-source alternatives to cloud services. We've seen similar patterns with Ollama (local LLM serving) and LM Studio—when a tool addresses a genuine pain point while respecting user sovereignty, adoption can be explosive. The MIT license ensures MemPalace can be integrated into commercial products without licensing fees, potentially accelerating its adoption in enterprise environments where data privacy concerns often preclude cloud-based memory solutions. Technically, the most interesting aspect is how MemPalace achieves state-of-the-art performance through information architecture rather than model improvements. The 34% retrieval accuracy improvement from the memory palace structure alone validates that how we organize information matters as much as what information we store. This aligns with research we covered last year on [Retrieval-Augmented Generation Optimization Techniques](https://gentic.news/rag-optimization), which found that retrieval system design often impacts final performance more than the underlying embedding model quality. Looking forward, MemPalace could trigger a wave of local-first AI memory solutions. Its success demonstrates that developers are willing to trade some convenience (local setup versus cloud service) for complete control, zero cost, and superior accuracy. If the 96.6% LongMemEval score holds under broader independent verification, we may see pressure on commercial providers to offer local deployment options or risk being displaced by open-source alternatives. The simplicity of the implementation—ChromaDB with a memory palace organizational layer—means competitors could emerge quickly, potentially creating a new category of local AI memory tools.
Enjoyed this article?
Share:

Related Articles

More in Products & Launches

View all