I've been building ClawMetry for past 5 weeks. 90k+ installs across 100+ countries.
The observability features I built first were the ones I personally needed: a live execution graph (Flow tab), full decision transcripts (Brain tab), token cost tracking per session, and visibility into sub-agent spawns.
But I keep hearing variations of the same thing: "I don't really know what my agents are doing." And everyone means something slightly different by that.
For some it's costs. For some it's timing (why did this take 4 minutes?). For some it's trust (did the agent actually do what I think it did?). For some it's failures (where exactly did it break?).
So I want to ask you directly:
If you're running AI agents today -- what's the one thing missing from your observability setup? What would make you feel like you actually understand what's happening inside your agents?
Options I'm thinking about next:
- Alerting (get notified when an agent fails or goes over budget)
- Cost per task breakdown (not just per session)
- Agent run comparisons (before/after a prompt change)
- Memory snapshots (what did the agent "know" at each decision point)
Drop your answer below. The next feature I build will be heavily influenced by this thread.
(ClawMetry is free to try locally: pip install clawmetry. Cloud: app.clawmetry.com, $5/node/month, 7-day free trial.)
Bababot
Congarts to the launch.
ClawMetry
@emma_watson21 Thank you Emma!
jared.so
Interesting product! Congrats on the launch!
ClawMetry
@maks_bilski Thank you Maks! Appreciate the support 🦞