Are we over-engineering AI memory? (Markdown vs. Vector DBs for small datasets)
Hey makers!
Lately, I’ve been looking closely at how independent builders and small teams are managing AI knowledge bases. It feels like the default "industry standard" is to immediately reach for a complex RAG pipeline and a heavy, paid Vector Database.
But I'm starting to wonder if we are over-engineering this for 90% of standard use cases.
Vector DBs are incredibly powerful for massive scale, but for smaller or non-massive datasets, they can be expensive, complex to query, and act as complete black boxes. If a search returns a weird chunk, diagnosing it is often a nightmare.
Lately, I've been experimenting with a dead-simple alternative for smaller projects, and I'd love to know if anyone else is doing this:
Dump raw data into a local folder.
Use a strict markdown schema (like a CLAUDE.md or similar instruction file) to tell the LLM how to summarize and map the data.
Let the AI build a flat, interlinked wiki overnight with directional backlinks (e.g., Supports [[Topic A]] or Contradicts [[Topic B]]).
I've found this to be 100% human-readable, highly portable across AI tools, virtually free to host, and shockingly accurate because agents can actually traverse the file logic rather than relying on pure mathematical similarity scores.
I want to hear from the builders and technical founders here:
Are Vector databases actually a mechanical necessity for smaller, non-massive datasets?
Or are we defaulting to them because it's the trendy stack right now?
If you aren't using a Vector DB for your AI features, how are you handling knowledge retrieval?
Let's talk shop in the comments!

Replies
I think the real issue isn’t vector DB vs markdown.
It’s that memory is treated as read-only and per-agent.
Once you have multiple agents, everything fragments fast.
What worked better for me is shared memory that agents can both read and write to.
At that point, storage matters less than how memory flows.
Curious if anyone here tried cross-agent memory instead 👀
Do you ever miss semantic search capabilities?
What's the failure mode when the model misinterprets links?
Are you using embeddings at all in this workflow?
Is this more of a prototyping solution or production ready
What kind of datasets has this worked best for?
How do you debug when retrieval feels off?
Does this approach work across different LLMs consistently?
How much manual curation is still required?
What's the biggest limitation you've hit so far?