Once a Test Run is complete, the Portal gives you three layers of information:Documentation Index
Fetch the complete documentation index at: https://docs.trymaitai.com/llms.txt
Use this file to discover all available pages before exploring further.
- Details: what configuration was used + timestamps
- Results: macro metrics and score distribution
- Requests: per-request scores, criteria breakdowns, and response comparisons
Run details (what was executed)
On the Test Run page, the Details section includes:- Date created / completed
- Status (e.g.
COMPLETED,ERROR, etc.) - Description
- Configuration
- The UI lets you open a read-only config panel to inspect the run’s config (including the model).
Results (macro view)
The Results section includes:- # Requests: total requests in the run
- Completed % / Error %: how many requests completed successfully vs errored
- Response time percentiles
- The UI shows a headline percentile and exposes additional percentiles (e.g. p50/p90/p95) in a tooltip.
Score distribution (1–5)
The score distribution shows the run broken into five buckets:- 1 — Poor
- 2 — Fair
- 3 — Good
- 4 — Great
- 5 — Perfect
Requests table (debugging view)
The requests table is where you debug individual cases. You’ll see:- Status for each request execution
- Tags associated with the underlying test request
- Score
- Hover to see the criteria breakdown (criteria name + 1–5 star value + optional description)
- Time (per-request response time, when available)
- Actions
- View: open the original request for deeper inspection
- Responses: compare the baseline/original response against the test-run response side-by-side
- Compare: open a comparison modal for the same request across multiple runs
Comparing runs
Maitai supports two comparison workflows in the Portal:- Compare Runs (test-set level): select multiple completed runs from the Test Set page and open a table that shows request-by-request scores across runs. You can also open a “Responses” comparison to view multiple run outputs side-by-side.
- Compare (single request across runs): from a Test Run’s requests table, open a modal that lists how that specific request performed across different runs.