Analyze and evaluate automation potential for support tickets
Are you sure you want to reset the application? This will delete all test runs and results. This action cannot be undone.
Evaluate support ticket resolution workflows, tool usage, and scoring.
Standard mode for testing exact matches or content validation.
Upload one or multiple JSON test suite files
(( file.testCount )) test(s)
System prompt: (( file.systemPrompt.length > 100 ? file.systemPrompt.substring(0, 100) + '...' : file.systemPrompt ))
Hint for the LLM about the primary language of the tickets.
Runs only the first N tests specified in the test suite.
Runs a random sample of N tests from the test suite.
Controls randomness: Lower values (e.g., 0.2) make output more focused/deterministic, higher values (e.g., 0.8) make it more creative/diverse and potentially better at analyzing sparse details.
Add optional delay (ms) between requests to avoid overloading the API
Set the maximum number of tests to run in parallel.
No test cases yet. Click "Add Test Case" to create one.
Upload a previously downloaded JSON result file to view it below.
(( formatCompactRuntime(selectedRun.summary.runtime_seconds) ))
((selectedRun.summary.total_tests))
(( Math.round(selectedRun.summary.success_rate * 100) ))% ( ((selectedRun.summary.successful_tests)) )
(( formatScoreAsPercentage(selectedRun.summary?.score_average) ))
| Tool Name | Usage Count | Usage % |
|---|---|---|
| (( tool )) | (( count )) | (( calculateToolPercentage(count, selectedRun.summary.tests_completed) ))% |
| Test Name | Category | Status | Expected | Actual |
|---|---|---|---|---|
| ((result.test_name || 'Unnamed Test')) (Iter: ((result.iteration || 1))) | ((result.category || 'No Category')) | (( result.success ? 'Pass' : 'Fail' )) |
(( result.validation ? JSON.stringify(result.validation.expected, null, 2) : 'N/A' )) |
((result.output || 'No output')) Error: (( result.error )) |
Tool: (( result.tools_assessed.join(', ') )) Tool: None Assessed
Click to view details
Error: (( result.error ))
(( cleanInputText(selectedRun.results[selectedTestIndex].input) ))
Input prompt not available for this test case.
(( agentResponse.steps && agentResponse.steps[0] ? agentResponse.steps[0] : 'None listed' ))
None listed
None Assessed
((selectedRun.results[selectedTestIndex].output))
Error: (( selectedRun.results[selectedTestIndex].error ))
No response output or error recorded for this test.
No required steps defined for this test.