Skip to main content

Metrics

Metrics

1. Average Run Time

Shows the mean duration of all test runs in scope. Indicates the typical time your pipeline requires to complete. Use it as a baseline for tracking daily performance.

2. Fastest Run

Displays the shortest single run duration in the selected window. The “Best yet” badge marks a new record relative to previous data in your project.

3. Speed Improvement

The percentage decrease in average run time compared to the previous period. A higher positive value means runs are faster than before.

Tags

A table of all run-level tags used in the selected time period and environment. Use it to compare stability across tag categories such as regression, smoke, or release-candidate. The table header shows the total number of tags found in the selected period. Use the search box to find a specific tag. Pagination controls appear when the tag list exceeds one page.
ColumnDescription
TagTag name with a color indicator and link icon
RunsNumber of test runs containing this tag
PassedTotal passed test count (green)
FailedTotal failed test count (red)
FlakyTotal flaky test count (yellow)
Pass RateOverall pass percentage
Sort by any column to surface the least stable or most-used tags. Click a tag row to open the Test Runs list filtered to that tag.

Speed by Branch Performance

Branch performance speed comparison A column chart that compares average test run time across branches. A baseline helps you see which branches are above or below the target. Hover to view each branch’s average duration and test run count. Test execution efficiency trends over time This is an area chart of average run duration per day. Highlights gradual drifts or sudden regressions in runtime. Hover to see the day’s average and test run count.

Test Run Speed Distribution

Test run speed distribution categories Stacked bars that bucket daily runs into Fast, Normal, and Slow groups based on duration thresholds. Reveals whether slow runs are isolated or common on a given day.