TwoTrim AI - Make Every Token Count, Save on LLM API Bills
byโข
TwoTrim is a research-backed, lossless token compression engine that reduces LLM input tokens by 20โ60% with no change in output quality.
It is stateless, secure, and drop-in compatible with OpenAI, Anthropic, and Gemini.
Built from a published study across 50,000+ prompts, TwoTrim delivers predictable savings with zero deployment cost and zero data storage.
Our goal is simple: help AI builders scale without scaling their API bills.



Replies
Maker Town
Maker Town
Hereโs our research summary ๐
We evaluated TwoTrim across 50,000 diverse prompts spanning reasoning, chat, structured tasks, analysis, and JSON workflows.
Some of the key findings:
25.96% average token reduction with perfect output fidelity
62.1% of prompts achieved 20โ40% savings
Mathematically guaranteed invertibility, ensuring outputs remain identical to uncompressed prompts
Model-agnostic design, validated across OpenAI, Anthropic, Gemini and others
We built TwoTrim as a research project first, then turned it into a tool for builders once the numbers consistently held up.
Read the full paper here:
๐ https://www.twotrim.com/resources/research
Happy to answer anything about methodology, benchmarks, or implementation!