Reports

OpenText Core Performance Engineering's new report interface provides statistics and performance data for a test run.

Report overview

OpenText Core Performance Engineering's Report page displays information about your load test run. Open the report in one of the following ways:

  • Go to the Load tests page, select a test, and go to the Runs Runs pane button pane. Select a run and click Report.

  • Open the Results page, select a row and click the Report button "".

By default, the Summary and Transactions report sections are displayed. For details, see Report sections.

Back to top

Set the report time frame

The time frame for which report results are displayed is shown above the report.

Report time frame banner

You can change the time range using one of these options:

  • Drag, stretch, or shrink the time slider.

  • Change the start and end time counters.

Back to top

Report sections

You can choose which sections to display in the report by selecting the checkbox for the section. In most tables, you can sort the lists by any column.

The following table shows the sections available in the report.

Section Description
Summary

The report summary provides an overview of the test run including duration, status, and important performance statistics such as average throughput.

Top 10 Transactions

The following data is displayed for the 10 transactions with the highest number of passes:

Passed and Failed Transactions

Passed and Failed Transaction graph example

This graph shows the following data throughout the test duration:

  • Passed transactions per second
  • Running Vusers
  • Failed transactions per second

The average and maximum values for these metrics are displayed in a table under the graph.

Transactions

The following data is displayed for the transactions in your test's scripts:

Errors per Second

Errors per second graph example

This graph shows the errors per second.

The average and maximum values for these metrics are displayed in a table under the graph.

Errors Summary

A list of errors encountered by Vusers during the load test run.

Throughput and Hits per second

Throughput and Hits per second graph example

This graph shows the following data throughout the test duration:

  • Hits per second
  • Running Vusers
  • Throughput

The average and maximum values for these metrics are displayed in a table under the graph.

Scripts

This grid lists all the scripts that were included in the test definition, and their configuration.

Distribution This table shows how the test Vusers were distributed over load generators during the test run.
Monitors

This table shows resource usage for server and application Monitors that were configured to work with OpenText Core Performance Engineering.

For each metric, the table shows its average and maximum value.

LG health monitoring

This table shows system health and resource usage for both on-premises and cloud-based load generators.

The system health metrics include CPU, Disk (on-premises only), and Memory.

HTTP Responses

The table lists the HTTP responses received during the test run and shows the total number received for each response code.

Note: The average number of HTTP responses is an estimate based on the selected time frame.

OPLG Scripts

The table lists which scripts ran on which on-premises load generators.

The OPLG Scripts table shows:

  • Load Generator: The name of the on-premises load generator.

  • Script: The name of the script.

  • Vusers: The number of Vusers that ran the script.

OPLG Errors

The table lists the top three errors for each script that occurred on on-premises load generators during the test run.

Note: This section is only available for scripts run on on-premises load generators version 3.3 or later.

Pause Scheduling

The table lists the pause and resume schedule actions that occurred during the test run. For information about pausing a test, see Pause scheduling.

This table is only visible when Pause Scheduling was enabled in the Run configuration settings.

Back to top

Percentiles

A percentile is a statistical tool that marks the value into which a percentage of participants in a group falls. In the context of test runs, if you check against the 90th percentile, you are looking to see which tests performed better than 90% of tests that participated in the run.

The following algorithms are commonly used to calculate percentiles:

Algorithm Description
Average percentile over time This method calculates the average percentile over time—not the percentile for the duration of the test. The percentile is calculated per interval (by default 5 seconds). The dashboard and report show an aggregate of the percentile goals.
Optimal percentile (T-Digest) An optimal (not average) percentile, based on the T-Digest algorithm introduced by Ted Dunning as described in the T-Digest abstract. This algorithm is capable of calculating percentile goals for large raw data sets, making it ideal for calculating percentiles for the large data sets generated by OpenText Core Performance Engineering.

OpenText Core Performance Engineering algorithm usage

For the Report:

  • Tenants created before May 2020 (version 2020.05):

    • Tests created before version 2020.10 (prior to October 2020), by default use the Average percentile algorithm to calculate the percentile.
    • Tests created with version 2020.10 or later (from October 2020), by default use the Optimal percentile (T-Digest) algorithm.
    • You can turn on the Optimal percentile (T-Digest) switch as described in Optimal percentile (T-Digest).
  • Tenants created in May 2020 and later (from version 2020.05) by default use the Optimal percentile (T-Digest) algorithm to calculate percentiles in all tests.

Note: For the Dashboard, OpenText Core Performance Engineering uses the Average percentile over time algorithm.

Percentile calculation

Typically, percentiles are calculated on raw data, based on a simple process which stores all values in a sorted array. To find a percentile, you access the corresponding array component. For example, to retrieve the 50th percentile in a sorted array, you find the value of my_array[count(my_array) * 0.5]. This approach is not scalable because the sorted array grows linearly with the number of values in the data set. Therefore, percentiles for large data sets are commonly calculated using algorithms that find approximate percentiles based on the aggregation of the raw data.

OpenText Core Performance Engineering uses the Optimal Percentile method, based on an algorithm known as T-Digest. This method was introduced by Ted Dunning in the Computing Accurate Quantiles using T-Digest abstract.

Recalculate the percentile

After a test run, you can recalculate the percentile for the transaction response time.

In the Transactions or Top 10 Transactions sections, expand the Percentile dropdown, select a new percentile from 1 through 99, and click Apply. The values in the grid change accordingly. For details, see Report sections.

Differences between percentile calculations between LoadRunner family products

If you have similar tests running in OpenText Core Performance Engineering and LRP or LRE, you may notice a difference in the response times, as shown by the percentile. This differences can be a result of the following:

  • LRP and LRE calculate the percentiles based on raw data. To do a fair comparison, make sure that you run your test in OpenText Core Performance Engineering with the Optimal T-Digest algorithm.
  • Think time setting should be consistent in your test runs. If tests in LRP or LRE include think time, while the same test excludes think time in OpenText Core Performance Engineering, you may see differences in the calculated percentile.
  • Even if your test on LRC runs with the Optimal percentile, for small sets of TRT (transaction response time) values you may encounter discrepancies. Typically this happens when T-Digest outputs results outside of the actual dataset, by performing average between values in the dataset, whereas other algorithms may also output values from inside the dataset.

     

    The following table shows the percentile calculation based on T-Digest using a small set of values:

    TRT values Percentile Value
    1,2,3,4,5,6,7,8,9,10 90th 9.5
    1,2,3,4,5,6,7,8,9,10 95th 10
    1,2,3,4,5,6,7,8,9,100 90th 54.5
    1,1,2,3 50th 1.5

Back to top

Add a dashboard graph custom snapshot

Save a snapshot of a customized graph in the dashboard and display it in the report for the run.

Add a snapshot from the Dashboard

  1. Go to the new dashboard for a specific run.
  2. Customize a graph with the data you want to view.
  3. Click the Add to report button Add to report button in the toolbar of the graph to save the snapshot. The snapshot is added as a section in the report.

For details on working in the dashboard, see Dashboard.

Note: You can add a maximum of 10 snapshots to a report.

Configure a snapshot in the report

Use the following actions to configure a snapshot in the report:

Action Task
Rename a custom snapshot
  1. Double-click the snapshot name.
  2. Edit the text.
Exclude a snapshot from the report Deselect the custom snapshot in the sections list.
Remove a custom snapshot Click the X at the top of a custom snapshot in the report viewer.

Back to top

Add notes

When viewing a report, you can add notes to the individual report sections.

To add a note to a specific report section, click ADD NOTE. In the dialog box that opens, enter your note and save it.

Back to top

Create report templates

The Reports window lets you save and load report templates. This capability lets you configure a report layout and save it as a template for future use. Using templates prevents you from having to reconfigure the same layout for each run report.

The report templates are stored as assets of OpenText Core Performance Engineering. To view the templates that are available for your tenant, go to Assets > Templates. For details, see Templates and Report template demos.

The template name is shown above the list of the report sections. If there are differences between your current configuration and the template, they are indicated by an asterisk and a slight notation above the ellipsis.

Template list menu

To save a layout as a template:

  1. Configure your report to show the sections that interest you.
  2. Click the Report templates button Report templates button on the top of the sections list.
  3. In the Template management menu, select Save as.

To load a template for your report:

  1. Click the Report templates button Report templates buttonon the top of the sections list.
  2. In the Template management menu, select Select template.

  3. Choose a template, and click Select.

    Choose template dialog box

Template actions

Click the Template Management buttonTemplate Management  button to perform one of the following actions:

  • To discard the differences between your layout and only use the layout of the chosen template, select Discard changes.

  • To overwrite the selected template by applying the current layout to the selected template, select Save changes.

  • To save the current layout as a new template, select Save as.

Back to top

Update template with transaction changes

You can modify an existing report template to reflect transaction changes in your test scenario. This includes:

  • Removing empty metrics from the template. For example, transaction data from the original test that was used to build the template.

  • Adding additional transactions to a template. For example, if you created a snapshot for a dashboard graph for the current run, and you want to include this in the template.

  1. Go to the dashboard for the specific run.

  2. To delete the empty metrics from the original script, click Delete empty metrics above the metrics pane.

    Delete empty matrix

  3. Customize the graph with the data you want to view, and click the Add to report button Add to report button in the graph's toolbar.

  4. Add sections for the new scenario to the template.

    1. Open the Report tab, select the sections that you want to apply to the template, and click Add to template.

      Add to template example

    2. In the Template Management menu (at the top of the Sections list), click the Report templates buttonReport templates button, and select Select template.

    3. Choose a template and click Select to load a template for your report.

    4. In the Template Management menu, click the Report templates buttonReport templates button and select Save changes to apply the current layout to the selected template.

Back to top

Exporting and sharing

For details on exporting report results in different formats and sending them by email, see Run results actions.

Back to top

See also: