Nicolas Grenié

Basalt Agents - Evaluate AI workflows and reach 99% AI quality.

Basalt, #1 AI observability tool for teams, is launching its brand new Agent Builder : prototype, test, and deploy complex AI flows composed of multiple prompts, and run them through a dataset of scenarios.

Add a comment

Replies

Best
Vladimir Lugovsky

Great launch! How does Basalt prevent overfitting on the evaluation dataset — e.g. if prompts start optimizing too much to test cases and lose generality?

Francois de Fitte

@vladimir_lugovsky great question !! To avoid overfitting, we recommend creating a dynamic dataset, meaning that you continuously enrich it with new test cases from your logs (something you can do from Basalt or programmatically) !

Steffan Bankier

Happy to discover the product! No more excel sheets for evals ;))

Francois de Fitte

@steffanb exactly !! Thanks :)

Ning Dong

Congrats on the launch! Curious if there's any example workflow with multimodal (video, speech) results?

Petter Magnusson

This looks solid! The evaluation-first approach is exactly what's needed. Curious how you handle workflows that need human approval gates between steps, not just eval metrics, but hard stops for review before continuing?