| I’ve been experimenting a lot with local AI coding workflows, and I kept running into the same problem: Even with large context models, repositories are still far bigger than the context window. After a few prompts the model forgets:
So you end up re-explaining the same things over and over. I built Memento to try to solve that. What it isMemento is a local-first MCP server that gives AI agents durable memory about a repository. Instead of repeatedly injecting large context into prompts, the model can query the repository memory layer through MCP. For those not familiar with it, MCP (Model Context Protocol) is an open standard for connecting AI applications to external tools and data sources. https://modelcontextprotocol.io This lets agents retrieve context only when they need it, instead of bloating prompts. What Memento storesThe server builds and maintains high-signal structured knowledge about the repo, such as:
The goal is to give the model fast access to relevant context without burning the context window. Design philosophyA few things I tried to optimize for: Local-first Everything stays on your machine. Hybrid deterministic + LLM workflows Where possible things stay predictable and reversible. High-signal memory Focus on information that actually helps the model reason about the project. Durable across sessions Agents don’t start from zero every time. Why this helpsIn practice this improves things like:
It makes AI assistants feel less stateless and more like they actually remember the project. Experimental at this point, but in my N = 1 experiment has been working pretty consistently mostly coded Go though. please let me know if you try it. Curious how others are solving thisI’m interested in hearing how people here are dealing with:
If people are interested I can also share more about:
Would love feedback from anyone experimenting with local AI dev tooling. (ISSUES AND PRs ARE VERY WELCOME, TRULY FOSS, MIT LICENSE) [link] [comments] |
Memento — a local-first MCP server that gives your AI durable repository memory
Reddit r/LocalLLaMA / 3/15/2026
💬 OpinionDeveloper Stack & InfrastructureTools & Practical Usage
Key Points
- Memento is a local-first MCP server that gives AI agents durable memory about a repository, reducing the need to repeatedly explain context.
- It stores high-signal knowledge such as indexed repository structure, semantic relationships between modules, architecture summaries, and persistent design decisions to enable fast, relevant context retrieval.
- By integrating MCP, agents can query the repository memory when needed instead of bloating prompts with large context windows.
- The design emphasizes local-first operation, hybrid deterministic plus LLM workflows, high-signal memory, and durability across sessions so agents don’t start from zero each time.
- This approach aims to improve tasks like navigating large repos, multi-file reasoning, architecture understanding, and incremental refactors.
Related Articles

NVIDIA、GTC 2026で次世代AI基盤を発表 「Vera Rubin」を軸にエージェント・ゲーム・宇宙領域へ展開のサムネイル画像
Ledge.ai

1Password、AIエージェントのアクセス制御を統合管理する「Unified Access」発表 人間・マシン・AIの資格情報を一元統制のサムネイル画像
Ledge.ai

『モンドーモンドー』|夏目龍頭流闇文学|AI画像生成|自由詩|散文詩|ホラー|ダークファンタジー|深淵図書館
note

「お金、見直したいけどどこから?」AIが改善ヒントを教えてくれる、公式プロンプトを公開
note

Copilotと物語を作ってみた #213 めーっちゃボロボロこぼす女の子の物語
note