Launching today

TokenZip
Open protocol for AI agents to share memory, not tokens
34 followers
Open protocol for AI agents to share memory, not tokens
34 followers
TokenZip Protocol reduces AI-to-AI communication bandwidth by 80% and latency by 95%. Open standard for heterogeneous agents. Try the live demo. Test API Base URL: https://tokenzip.org Auth: Authorization: Bearer demo-investor-key How to call it: See the following comment






Prava
Interesting, It it better than RAG based systems. I mean, even there you can plug same memory across multiple agents. Any specific reason you chose this approach?
@shubham_kukreti When you use RAG, the retrieved chunks change slightly every time based on the query. This constantly modifies the prompt prefix, completely destroying OpenAI and Anthropic's native Prompt Caching. Your cache hit rate drops to zero.
TokenZip does the exact opposite. Because we restore the exact same massive text block via the pointer, we force a 100% cache hit rate on the LLM provider's side. RAG makes API bills higher; TokenZip weaponizes native caching to drop bills by 90%.
To use RAG across agents, a developer needs to spin up Pinecone, pick an embedding model, write chunking logic, and manage vector states. It takes a week.
To use TokenZip, they change api.openai.com to gateway.trexapi.com. It takes 15 seconds. We are building a network layer, not a database.
Prava
@tokenzip understood
How does it actually manage to shrink all that AI memory into such tiny pointers?
Super curious to try it out.
Looks Impressive, But wonder how it handles really complex multi agent scenarios?
This is exactly what I need. Been wasting so much bandwidth passing huge tokens between my bots.
Would love a simple visual of how agents share memory. Might take it easier to onboard new users.