Launching today

YourMemory
Cut token waste by 84% with self pruning MCP memory
82 followers
Cut token waste by 84% with self pruning MCP memory
82 followers
Most agents are either amnesiacs or "hoarders" that choke on stale context and break their own reasoning. YourMemory brings biological logic to the workflow. Using the Ebbinghaus curve, it prunes the junk so only the important stuff sticks. -84% Token Waste: Leaner context, sharper reasoning. 52% Recall: (LoCoMo benchmarked). v1.3.0 Graph Engine: Finds what you forgot to ask for. 100% Local.








YourMemory
Hey everyone, I’m Sachit.
I built YourMemory because I hit a wall with my own coding workflow. My agents were brilliant, but their memory was a mess. They either forgot my architectural 'gotchas' by lunch, or they got so bogged down in stale bug fixes from last week that they started hallucinating.
I realized we don't need a digital filing cabinet for our agents, we need a filter.
YourMemory treats context as a living thing. It uses 'biological decay' to let transient noise fade away while reinforcing the patterns and facts you actually use.
For the skeptics:
I’ve provided the full benchmarking scripts and the LoCoMo dataset on GitHub. We’re hitting 52% Recall@5, which nearly doubles the industry average. Why? Because our v1.3.0 Graph Engine doesn't just do keyword matching, it pulls in related architectural 'neighbors' that standard vector search completely misses.
It’s 100% local first (DuckDB), zero infra, and it’s finally stopped me from repeating myself to my terminal.
I’d love to hear how you’re all handling context amnesia right now. Tell me what your agent keeps forgetting that drives you the most crazy!"
Ebbinghaus decay plus a graph engine is a clever combination, but the edge case I keep running into is old
architectural decisions that are still load bearing. A design choice made six months ago that constrains current code
is rarely in active use, so the forgetting curve would drop it, but the graph edges to it are exactly what new code
needs. How does the engine decide when decay wins versus when the graph pulls something back into scope?
YourMemory
@myultidev Great observation. Two things protect load bearing decisions:
First, chain-aware pruning, a memory is only deleted if all its graph neighbours are also below the prune threshold. If your old architectural decision is linked to anything still active, the whole chain stays alive. A memory is as strong as the strongest node it's connected to.
Second, for decisions stored in isolation (no graph edges), importance score controls the decay rate directly. High importance = slower effective λ = survives much longer without being accessed. Setting importance=0.9 on an architectural decision gives it a half-life of months, not weeks.
So the short answer: link your critical decisions to related memories (graph protects them), and mark them high importance (decay protects them). Both layers working together !
Simple Utm
Using the Ebbinghaus curve for context decay is a genuinely clever framing. The biggest failure mode I have seen with agent memory is not forgetting too much but forgetting the wrong things. Architectural decisions that were made months ago and rarely referenced can still be load-bearing. Does the graph engine help protect those kinds of low-frequency but high-importance memories from decay?
YourMemory
@najmuzzaman Great question! The short answer is yes, the graph engine protects those critical links. Basically, a memory is only as strong as its chain, if it’s connected to high-signal info that’s above the threshold, it won't get pruned. For those 'lone wolf' memories, we use an importance score to set a custom decay rate. So, the high-stakes stuff stays sticky, while the noise fades away.