Create manual tests

This topic describes how to create manual tests and parameterize their steps with data sets.

For details

Creating tests

This sections describes how to create a manual test. You can create tests from the Quality, Team Backlog, and Requirement modules, either from the Tests page or from a Test Suite.

To create a manual test:

  1. Create a test in one of the following ways:

    • In the Team Backlog or Quality module menus, click the Tests tab. Click + Manual Test. If this option is not visible, expand the dropdown and choose + Manual Test.
    • In the Requirements module menu, click Tests and select a Requirement document. Click + Manual Test. If this option is not visible, expand the dropdown and choose + Manual Test.
    • Inside a Test Suite, select the Planning tab. Click + Manual Test.Click + Manual Test. If this option is not visible, expand the dropdown and choose + Manual Test.
  2. In the Add Manual Test dialog box, assign the following test attributes:

    Field Details
    Name The name of the test.
    Test type The type of test, such as acceptance, end-to-end, regression, sanity, security, and performance.
    Estimated duration (minutes) The time it takes to run the test.
    Backlog coverage The backlog items the test will cover. This helps you track the release quality.
    Application modules The product's application modules. This helps you track product quality, regardless of release.
    Description A textual description of the test.
  3. Click Add & Edit. The test is created, and the Steps tab of the test opens.

  4. Add steps to the test. For details, see Add steps to a test.

  5. Save the test.

    Tip: While you work, save and label versions of your manual test. As you work with the test over time, compare versions to view modifications. For details, see Use versions of test scripts.

Back to top

Add steps to a test

After you have created a test, you add steps that describe how to set up and validate the application.

To add steps to a test:

  1. In the Steps tab of the test, select a script editor view to show test steps as a grid, a list, or plain text. The underlying test steps and syntax remain the same in all views, and you can switch between views as needed.

    View Details
    Grid view View the steps as a grid. Each row of the grid includes an Action column with a setup step, and an Expected result column with a corresponding validation step.
    List view View the steps as a list. Each line of the list includes one step: a setup step, validation step, or call step.
    Text view View the steps in plain text.

    Tip: The admin can set the default script view for new users using the MANUAL_TEST_SCRIPT_DEFAULT_VIEW parameter. For details, see Configuration parameters.

  2. Add steps using the toolbar buttons.

    If you are entering steps manually, make sure to use the correct syntax. For details, see Manual test syntax.

    Button Details
    Add Step

    Add a setup step (an action in the application).

    Enter the step text. For example: Create a new Epic.

    Add Validation Step

    Check something in the application.

    Enter the step text. For example: Verify that the current phase is New.

    During the test run, you specify a pass or fail status for each validation step.

    Add Call Step

    Add steps from an existing test to your script.

    In the Add Tests dialog box, select the test or test that you want to call, and then click Add.

    The step is added as a hyperlink to the original test. Click the View test steps button to display the steps of the test that you called.

    Note: ALM Octane uses the latest version of the called test at run time, and not the latest version at the time of planning.

  3. To apply text formatting, see Format text in the test script.
  4. To add attachments, select a step and click Attach . You can also paste an image directly into the test.

    When viewing the steps in text view, the attachment appears in the line under the associated step.

    You can view the file in the Attachments tab of the test, in the Manual Runner when starting a test run, or in the test run report.

    Note: If you rename a test's attachment after it is already included in a step, you must manually update the link to the attachment in the step.

Back to top

Using parameters in tests

You can use parameters in tests to run a single test several times with different values.

To use parameters you must create a data table, define the parameters, and create at least one data set with values. The number of rows in your data set determines the number of iterations the test will run.

If you create multiple data sets, when you run the test, you will be able to select the data set with which to run the test.

Multiple data sets are useful when you want your test to reflect numerous scenarios. For example:

  • If you want to test log in success and failure, you can create two data sets - one with valid credentials and another with credential that will fail to log in.
  • You can create a different data set for each environment you want to test, like different browsers or different operating systems.

Here's an example showing a list of users and passwords. Each iteration will use a different set of values:

Data tables are shared, so you can use a single data table for multiple tests.

To work with parameters: 

  1. In the Steps tab of the test, click Data Table.

  2. In the pane that opens on the right, click Create a new table.

  3. Enter a name for the table and rename the parameters. You can add or remove columns as required.

    To add or remove rows or columns, click the dropdown arrow next to the row or column name. The arrow is visible when you hover over a cell or column header.

    Parameter names cannot contain empty spaces or angle brackets (< >).

  4. In each row, enter a set of parameter values.

  5. To add data sets, click the menu and then click Add Set. Enter parameter values for this set.

    The first data set in the data table is always the default data set. When a data set is not explicitly selected in the run, the default data set is used.

    The names of the sheets in the import file will be the names of the target data sets. You can rename data sets using the menu commands after import.

  6. Click Save when your table is ready.
  7. You can use the menu options next to the Save button to manage your parameter tables. If you select Remove from test, you disconnect the table from the current test but it is still saved in ALM Octane. If you select Delete table, it is removed from all tests that are using this parameter table. ALM Octane shows you which tests use the table before you delete it.

  8. In the relevant test step, enter parameter names using the following syntax: <parameter name>.

    Tip: You can also quickly access the list of parameters by pressing CTRL-SPACE. Use the keyboard arrows to navigate the list, and Enter to select a parameter.

  9. When you run the test, it runs in multiple iterations. Each iteration uses a different set of parameter values from the table.

    For example, here are execution steps presented to the tester:

Back to top

Export and import data tables or data sets

You can export data tables and data sets to Excel. You can update the Excel file, and then import it back to your manual test.

For each data set, consider the following requirements:

  • Must have a unique name
  • Must not exceed 31 characters
  • Cannot include special characters (for example \ / * [ ] :?)

To export a data table:

  1. In the Steps tab of the test, click Data Table
  2. Select Export table. The data table opens in Excel. Each data set is displayed in a separate Excel sheet.

To export a data set:

  1. In the Steps tab of the test, click Data Table.
  2. Select a Data Set.
  3. Select Export set. The data set opens in Excel.

To import a data table:

  1. In the Steps tab of the test, click Data Table
  2. Select Import table. The Import Data Table dialog box opens.
  3. Select Override the content of the current data table or Create new data table and link to current test.
  4. If you select Override the content of the current data table, click Browse. Locate the data table. Make sure that table name matches the table name in the Excel file that you are importing from. Click Import. ALM Octane reads each sheet and determines whether to override or add data sets.

  5. If you select Create new data table and link to current test, click Import. ALM Octane creates a new data table with at least one data set and links it to the current test.

    Note: If you have unsaved changes and you choose to create a new data table, your changes will be lost.

To import a data set:

  1. In the Steps tab of the test, click Data Table.
  2. Select a Data Set.
  3. Select Import Set. The Import Data Set dialog box opens.
  4. Select whether to Override the content of the selected data set or Create new data set under the current table.
  5. If you select Override the content of the selected data set, click browse. Locate the data table. Click Import. ALM Octane loads the first sheet and copies the values into the selected data set.
  6. If you select Create new data set under the current table, click Import. ALM Octane creates a new data set and places it under the current data table.

Back to top

Copy tests to another workspace

You can copy tests from one workspace to another, within the same shared space.

For details, see Copy items to another workspace.

Back to top

Next steps: