Ashr is a test and evaluation platform purpose-built for AI agents. Part of YC W2026, it was founded by Shreyas Kaps (Fortune 100 AI agent experience) and Rohan Kulkarni (CTO, ex-Berkeley AI startup exit). Since agents cannot be unit tested like traditional APIs — inputs are unstructured, outputs are probabilistic, and failure modes are creative — Ashr generates synthetic but authentic user stories that flow through your product.
The platform works across voice, text, image, file generation, and multimodal interactions, catching errors that would take hours of manual testing. It includes prompt versioning with inline diffs and pass-rate tracking per version, full test timelines showing every speaker turn, tool call, and response, plus side-by-side comparison of expected vs. actual results.
Teams integrate via SDK and can run evaluations both pre-production and post-production. Users at UC Berkeley and Stanford are already on the platform. Ashr fills the critical gap of systematic, repeatable testing for probabilistic AI systems.
Free trial available
Teams building multi-modal AI agents
Ashr tests AI agents before deployment while Respan monitors them in production. Together they provide full lifecycle coverage from pre-production testing to production observability.
Top companies in Observability, Prompts & Evals you can use instead of Ashr.
Companies from adjacent layers in the AI stack that work well with Ashr.
Last verified: March 27, 2026