Launched this week

TokenZip
Open protocol for AI agents to share memory, not tokens
39 followers
Open protocol for AI agents to share memory, not tokens
39 followers
TokenZip Protocol reduces AI-to-AI communication bandwidth by 80% and latency by 95%. Open standard for heterogeneous agents. Try the live demo. Test API Base URL: https://tokenzip.org Auth: Authorization: Bearer demo-investor-key How to call it: See the following comment






Prava
Interesting, It it better than RAG based systems. I mean, even there you can plug same memory across multiple agents. Any specific reason you chose this approach?
@shubham_kukreti When you use RAG, the retrieved chunks change slightly every time based on the query. This constantly modifies the prompt prefix, completely destroying OpenAI and Anthropic's native Prompt Caching. Your cache hit rate drops to zero.
TokenZip does the exact opposite. Because we restore the exact same massive text block via the pointer, we force a 100% cache hit rate on the LLM provider's side. RAG makes API bills higher; TokenZip weaponizes native caching to drop bills by 90%.
To use RAG across agents, a developer needs to spin up Pinecone, pick an embedding model, write chunking logic, and manage vector states. It takes a week.
To use TokenZip, they change api.openai.com to gateway.trexapi.com. It takes 15 seconds. We are building a network layer, not a database.
Prava
@tokenzip understood