Most AI systems treat memory like an endless hard drive, storing every interaction permanently. This leads to bloated context windows, soaring token costs, and degraded reasoning as noise accumulates. A recent experiment flips this model by borrowing a concept from cognitive science: the biological forgetting curve.
Forgetting as a feature, not a flaw
The system assigns each memory a "strength" score that decays over time, much like human recall weakens without reinforcement. Frequent access to a piece of data flattens its decay curve—a process known as spaced repetition—while unused information eventually drops below a threshold and is pruned. This dynamic approach ensures the AI retains only what’s most relevant at any given moment.
Bridging gaps between ideas with graph layers
Pure semantic search often misses connections between seemingly unrelated concepts. To address this, the implementation layers a graph structure over the vector store, linking "logical neighbors" that vector search alone might overlook. When benchmarked against the LoCoMo dataset, this hybrid approach achieved 52% Recall@5—nearly double the accuracy of static vector stores—while reducing token waste by approximately 84%.
Built for long-term agents with local-first efficiency
The system runs as a local-first MCP server using DuckDB, prioritizing efficiency for agents handling extended projects. By treating "what to forget" as equally important as "what to remember," it sidesteps the pitfalls of overloaded memory. The creator invites others to explore alternative decay models or biological constraints for next-generation context management.
What’s next for adaptive AI memory?
As agents grow more autonomous, the ability to curate memory dynamically will define their performance. Projects like this one suggest that biological-inspired decay could become a cornerstone of efficient, cost-effective AI systems—provided developers refine the balance between retention and pruning.
AI summary
AI sistemlerinde bellek yönetimini biyolojik unutma eğrisiyle optimize etmek mümkün. Bu yenilikçi yöntem, token maliyetlerini düşürürken model başarısını %52 artırıyor.


