Test Runs
Summary

Summary

Use Summary to group failed, flaky, and skipped tests by cause. Filters and sorting help you review faster.

KPI Tiles

Detailed Analysis

1. Failed

A failed test ran and ended with an error or unmet assertion. Evidence shows the failing step or message.

  • Assertion Failure: The expected value did not match the actual value.
  • Element Not Found: The locator did not resolve to an element.
  • Timeout Issues: An action or wait exceeded the set time.
  • Network Issues: A request failed or returned an unexpected status.
  • Other Failures: Errors that do not fit the above, for example, script errors or setup issues.

2. Flaky

A test is categorized as Flaky when the outcome is inconsistent across attempts or recent runs without a code change. It often passes on retry.

  • Timing-related: Order, race, or wait sensitivity. Often passes on retry.
  • Environment-dependent: Fails only in a specific environment or runner.
  • Network-dependent: Intermittent remote call or service instability.
  • Intermittent Assertion: Non-deterministic data or state causes occasional mismatches.
  • Other Flaky: Unstable for reasons outside the above buckets.

3. Skipped

This test did not run due to a skip directive, configuration, or runtime condition. No assertions executed.

  • Manually Skipped: Explicitly skipped in code or via tag.
  • Configuration Skipped: Disabled by config, project, or reporter settings.
  • Conditional Skipped: Skipped due to an evaluated condition at runtime.

Detailed Analysis

Detailed Analysis

A single-table view of every test in the run, designed for fast triage at a glance.

  • Shows status, spec file, duration, retries, failure cluster, and AI category in one place.

  • History preview shows the current run and up to 10 past executions to confirm new vs recurring issues.

  • Provides token search and filter chips.

  • Sort by duration or status to surface slow or failing tests first.

  • One-click access to evidence (error text, screenshots, console) and to the test case details page.

Use this view to move from a high-level signal to the exact test that needs action in seconds.

1. Search and Tokens

To filter, either select from the Failed/Flaky/Skipped cards or, if you require a more specific sorting, use these short tokens.

s: status (passed, failed, flaky, skipped)
c: cluster (assertion-failure, timeout, network-error, …)
@ tag (smoke, regression, e2e)
b: browser (chrome, firefox, safari, edge)

2. Duration Sort

Switch between Default, High to Low, and Low to High to spot slow tests and quick wins.

3. Context Carry-Over

When you choose a group in Summary, the same context is applied here, so the table lists only matching tests.