Summary Tab

Last modified on November 20, 2023

The Summary tab of the test results screen provides a brief overview of the test run and accumulated results. This tab is active by default on the screen.

Summary Chart

The chart at the top shows how various metrics changed during the test execution. It also helps you see the correlation between the metrics and the number of simulated VUs and errors. The chart displays data for one UI or API test that virtual users simulated:

Results: The Summary chart

Click the image to enlarge it.

Move the mouse pointer over the chart to see metric values at certain time points.

Metrics
Metric Description

Gross Duration  

Total step execution time (includes think times of individual test events).

Net Duration  

Step execution time excluding test events’ think times. In other words:

Net Duration = Gross Duration – Think Time – Non-Blocking Think Time

Response Time  

The total time during the test run (on all iterations of the selected test) that the browser spent waiting for the response from the server. This includes time periods spent on DNS resolving, setting up a connection to the tested server (if needed), processing redirects, and transferring data over the network.

Response time

3rd-Party Response Time  

The response time taken for requests to domains different from the domain of the navigated URL. Examples of these third-party domains include Google Analytics, font storages, Facebook, weather info providers, and so on.

This value helps you understand how much time your pages spent on requesting third-party sites.

Think Time  

The sum of all Think Times in the simulated test. This value doesn’t include non-blocking Think Times, that is, the Think Time values that didn’t cause pauses in the network activity (see below).

Non-Blocking Think Time  

This metric is a sum of Think Time values that don’t pause the test playback.

A typical test has multiple think times in it. For example, a recorded UI test has Think Times for almost every recorded event. Depending on the length and location of these Think Times in the test, they might pause the test execution or not. For example, a Think Time at the beginning of the test step, most probably, doesn’t block the test run, because at this time the browser is still sending requests and loading resources requested by the current page. However, a long Think Time at the end of the test step will most likely pause the test run, because all the resources, most probably, have been downloaded and the test engine spends time idling.

Front-End Processing Time  

The time spent on client-side processing. This includes time for running JavaScript client code, various actions with DOM, and so on.

Note: Currently, this time also includes LoadNinja overhead (like time spent on JavaScript validations).

Working with the chat

  • Zoom in or zoom out chart regions. The easiest way to do this is to drag the mouse pointer within the chart – this will select a chart region and will zoom in on it.

    You can also zoom in or out by using special buttons:

    Zoom in and zoom out buttons
    Note: The “zoom-in by dragging” mode works if the setting is selected (it is selected by default).
  • Show or hide metrics quickly:

    Show and hide metrics
  • The and buttons specify how the chart handles mouse drags:

    Mouse drag modes
  • Export chart data using these menu items:

    Download chart image
    Menu items Description
    Download SVG
    Download PNG

    Save the current chart view to an SVG or PNG file (that is, if you zoom in on some part, the images will have that visible part only, not the entire chart).

    Download CSV 

    Saves all chart data to a .csv file. You can then analyze and process the exported data in the way you need.

Summary per Script

This panel contains statistics on how successfully virtual users performed script runs.

Column Description
Script The name of the script the virtual users performed.
Load Distribution The share of the load you set for this script in scenario settings.
Expected Iterations For iteration-based tests. The number of iterations each virtual user was supposed to run.
Total Successful Steps The number of test steps that the virtual users completed successfully.
Failed The number of test steps that the virtual users failed.

Scenario information

This panel shows the scenario settings you specified for this load test run.

See Also

Test Results

Highlight search results