Generate a performance dashboard for an evaluator — KPIs, timeline trends, quality heatmap, top failed criteria, participant rankings, and a filterable transcription table.
The Evaluator Analysis page aggregates all transcriptions processed with a selected evaluator and organizes the results into KPI summaries, trend charts, a weekly quality heatmap, criteria failure breakdowns, participant rankings, and a filterable transcription detail table.
Each analysis generation consumes one unit of your account’s analytics_evaluator quota. Loading a previously cached result for the same evaluator ID does not consume quota.
Select an evaluator from the searchable selector at the top of the page and click Generate Analysis. The selector shows each evaluator’s name, ID, and criteria count.
Navigating directly to a URL that includes an evaluator ID (e.g. from a bookmark or shared link) auto-loads the last cached result — no quota consumed.
The Generate Analysis button is disabled when no evaluator is selected or when your quota is exhausted. If you click it without a selection, a warning is shown: “You must select an evaluator to generate the analysis.”
Two side-by-side period cards — This Month and Previous Month — each showing:
Calls — total count
Average Score — numeric
Pass Rate — percentage
A delta summary below the cards shows the score change (green upward arrow for improvement, red downward for decline) and the pass rate change between periods.
State
Display
One period has no data
Amber warning on that card
Both periods have no data
”Data in both months is needed to show the comparison”
Three tiles grouping calls by duration using thresholds dynamically calculated from the dataset:
Tier
Threshold
Short
Below p25
Medium
Between p25 and p75
Long
Above p75
Each tile shows the call count and average score for that tier.
Duration thresholds are dynamically calculated using the p25 and p75 percentiles of your dataset — they adapt to the actual distribution of your calls, not a fixed value.
A bar chart grouping all transcriptions into 20-point score ranges: 0–20 · 20–40 · 40–60 · 60–80 · 80–100. Shows how scores are distributed across the full dataset.
A ranked list of the 5 most frequently failing criteria across all evaluated transcriptions.
Column
Detail
Rank
1–5 with a gradient badge
Criterion
Name — truncated with a tooltip showing the full text
Evaluator
Color-coded tag
Fail count
Absolute number of failures
Fail rate
Percentage of calls where this criterion failed, with a progress bar
These are the highest-priority areas for coaching and training. Use this list to identify where your team or candidates need the most focused improvement.
This section is only rendered when criteria failure data is present.