Use the Summary page in Leapwork Performance

This article explains what the Summary tab in Run results is for in Leapwork Performance.

What the Summary page is for

The Summary tab is the high-level view of a run. It gives you a quick overview of the run outcome, highlights key performance metrics, and helps you decide what to investigate next.

Use the Summary page when you want to:

  • review the overall outcome of a run without inspecting every timestamp

  • see the main performance signals in one place

  • understand whether the run behaved as expected under load

  • identify which areas may need deeper analysis in other result views

What you can see on the Summary page

The Summary tab provides a high-level overview of the run and includes:

  • run status

  • start time

  • total duration

  • peak virtual users

  • average response time

  • throughput

  • error rate

  • peak load

  • visual summaries such as cards that help you understand the overall run outcome more quickly

  • AI-generated run interpretation for supported test types

  • aggregated sequence-level metrics

Screenshot 2026-04-09 201340-20260409-144341.png

AI report detail

The AI Report detail section gives you a short written interpretation of the run.

At the top of the section, you can select the test type. The available options shown in the current UI are:

  • Performance test

  • Load test

  • Soak test

  • Stress test

  • Baseline test

The AI report answers key questions about the run, including:

  • whether performance degraded with load

  • where errors increased

  • which sequence or endpoint was the bottleneck

Screenshot 2026-04-09 201403-20260409-144403.png

Aggregated sequence metrics

The Aggregated sequence metrics table gives you a sequence-level view of performance so you can compare how different sequences behaved during the run.

The table includes columns such as:

  • Sequence

  • Total executions

  • Avg duration (ms)

  • average duration at specific virtual-user levels

  • p95 (ms)

  • Error rate

The table also includes a Configure VUs labels option so the virtual-user-specific average-duration columns can be labeled for the load levels you want to review.

Use this section when you want to:

  • identify which sequences ran most often

  • compare average and p95 duration across sequences

  • check whether a specific sequence becomes slower at higher virtual-user levels

  • confirm whether errors are concentrated in a particular sequence

Screenshot 2026-04-09 201628-20260409-144628.png

When to use Summary instead of other result views

Use Summary when you want the clearest overall picture of a single run.

Use Detailed view when you need to inspect how the system behaved at a specific point in time during the run.

Use Trend analysis when you want to compare comparable runs over time and understand whether performance is improving, stable, or regressing.

Typical workflow

  1. Open Results for the run you want to review.

  2. Open the Summary tab.

  3. Review the top-level run context and KPI values.

  4. Read the AI Report detail section for a quick interpretation of the result.

  5. Review Aggregated sequence metrics to see which sequences contributed most to duration or errors.

  6. Use the Summary page to decide whether you need deeper investigation in Detailed view or Trend analysis.