How they work
After a test run completes, the Insights agent reviews the full set of results and looks for patterns across individual test cases. Rather than reviewing each failure one by one, the agent groups failures into themes and highlights the most impactful areas for improvement. The inputs to Test Run Insights include:- Test case results — pass/fail outcomes and scores across all samples in the run
- Evaluation reasoning — the rationale behind each score from LLM judges or other evaluators
- Version comparisons — when multiple versions are tested, differences in performance across versions
Test Run Insights are currently available to select design partners. Reach out to your Freeplay contact to learn more.
Related resources
- AI Insights Overview — How Freeplay’s AI Insights work across the platform
- Testing — Run batch evaluations against datasets

