History
The History tab displays the complete execution record for a single test case, aggregated from all test runs on the current active branch. This view isolates a test's performance to track its reliability over time.
The history is filtered to show only test runs from the current active branch. Test runs from other branches are excluded.
What you see
1. Test Metrics
This section displays key performance indicators (KPIs) for the test case, scoped to the current branch.
- Stability: Represents how reliable a test case is over time. This is the primary indicator of test health. A 100% stability score is achieved only if the test case has passed in all tracked executions on this branch.
Stability is calculated as (Passed Runs ÷ Total Runs) × 100
-
Total Runs: The total number of executions tracked on this branch. This number provides context for all other metrics.
-
Passed / Failed / Flaky / Skipped: The specific counts for each outcome.
2. Last Status Tiles
These summary cards provide direct links to the most recent runs for each outcome, allowing for quick comparison.
Each tile displays:
-
The Run # (e.g., Run #251).
-
The timestamp of when that run occurred (e.g., "1 week 1 day ago").
-
A (current) label if the run you are currently viewing matches that status.
The tiles link directly to the test run report for:
-
Last Passed: The most recent run where this test case passed successfully.
-
Last Failed: The most recent run where this test case failed.
-
Last Flaky: The most recent run where this test case was marked as flaky.
3. Execution History Table
This table lists every execution of the test case on the current branch in chronological order.
| Column | Description | Purpose |
|---|---|---|
| Executed At | The timestamp of the test execution. | Shows when each run occurred, helping to correlate failures with deployments or other events. |
| Run # | The unique test run identifier. | Provides a specific ID for each execution. |
| Status | The outcome (e.g., Passed, Failed, Flaky, and Skipped) shown with a colored badge. | Allows for quick visual scanning of the test's pass/fail/flaky pattern over time. |
| Duration | The total execution time for the test. | Helps spot performance regressions or identify tests that are becoming slower. |
| Retries | The number of retry attempts made during that run. | A key indicator of flakiness. A run that passes after 1 or more retries is often flaky. |
| Run Location | A direct link to the specific build in your CI provider. | Provides a one-click path to the original CI job for deeper log analysis. |
| Actions | A View Test button to navigate to that specific test run's detail page. | Allows you to load the full test run report for any historical execution. |
The rows expand to show Error Details for failures or Console Logs if they were captured during execution.
How stability differs by outcome
Stability provides an at-a-glance measure of a test's reliability on the current branch. The percentage reflects the entire history, not just the most recent run.
-
100% Stability: The test case has passed in every single execution on this branch. This indicates perfect reliability in this context.
-
< 100% Stability: The test's history includes at least one Failed or Flaky run. Even if the most recent run passed, any previous inconsistency will lower the stability score.
Why it matters
-
Performance Tracking: This view allows you to monitor a test's reliability and stability trends over time.
-
Root Cause Analysis: You can quickly distinguish if a failure is a new regression (a stable test that just failed) or a recurring flaky test (a test with a sub-100% stability history).
-
Pattern Recognition: Analyze retry patterns, duration changes, and error messages across multiple runs to find the root cause of instability.
-
Context Access: Use the table to access historical error details, console logs, and direct links to CI builds for faster debugging.