Resolve Failing Test Cases

Introduction

Even when test cases have been carefully designed to be stable and maintainable, test failures can happen. There are several possible uses of the term “test failure,” so let’s distinguish between them:

Debugger

A negative test case

This is a test case that you expect to return an error from the application under test (AUT), such as an invalid password. This type of test case succeeds when it returns the expected error message.

Maintenance Mode

A test case that uncovers a defect in the application

This is actually a successful test case because identifying defects is one of the main goals of software testing. Consider this type of test case for inclusion in your regression test set.

Ranorex Remote

A test case that fails for a reason unrelated to the functionality of the application

This is the meaning of the term “failed test case” as used in this article.

When a test case fails, the first job is to decide whether situation #2 or #3 applies: Did the test case fail due to a defect in the AUT, or is the problem with the test case itself, such as missing or invalid test data, problems with the test environment, or changes in the AUT that are not defects? If it is not immediately clear, you may need to troubleshoot the test case itself before reporting a defect in the application.

It may be tempting to simply re-run a failed test case to see if it passes. But a test case that passes sometimes and fails on other occasions for no discernable reason is a “flaky,” unreliable test case. It’s important to resolve the issue that caused it to fail so that you can have confidence in the results of your automated testing.

Configure test runs to assist debugging

An earlier article in this series, “Build Maintainable Tests” described best practices for designing test cases that make them more stable and less likely to fail. These included eliminating dependencies between test cases as much as possible, ensuring that your test environment is stable, and removing tests that you expect to fail (such as ones for unresolved defects) from the test run. It is also helpful to configure your test cases to take a screenshot when a failure occurs.

In addition to these recommendations, be sure to configure the test run to handle failures appropriately. Only allow a failing test to stop the entire test run if that makes sense for the situation – for example, if the application fails to launch, or smoke tests fail. Ranorex Studio’s modular approach to test case design includes several options for continuing after a test case returns an error, including “continue with iteration,” “continue with sibling,” and “continue with parent.” You can also automatically retry a failed test case. To learn more, read the Ranorex User Guide chapter on the Ranorex Test Suite.

It’s also important to manage the size of test run reports by focusing only on true errors and failures. For example, Ranorex supports multiple pre-defined report levels, including “debug,” “information,” “warning,” and “success.” In a large test run, reporting information at this level may result in an excessive amount of data. Consider reporting results only for the “error” and “failure” levels to make it easier to spot true problems that need to be resolved.

Isolate the problem

If many test cases are failing, look for a problem with the environment, test framework, or the AUT.

Environment

Environment

Issues with the environment can include required services not running, or not running in administrative mode if required.

Test Framework

Test Framework

Look for issues with the test framework, such as a licensing error, or a remote agent not configured properly.

System Error

Application Under Test

Verify that the AUT is prepared correctly. This can include issues such as location-specific system settings, the wrong browser version, or even a different system language. Or, there could be a pending O/S update that blocks the user interface.

If most test cases in your test run have succeeded, then suspect issues with the individual failing test case(s). There may be an error message that points to the cause. If not, don’t just assume that the test case failed “accidentally” and re-run it. All test failures happen for a reason. A test case that appears to succeed or fail for no discernable reason is a “flaky” test. To get to the root of the problem, refer to the probable-cause checklist below.

Troubleshoot failed test cases

Work through a probable-cause checklist to troubleshoot each failed test case, asking questions such as the following:

  • Is the test case up-to-date with the AUT? For example, has the test case been updated with any/all changes in UI elements?
  • Is the input data correct and available to the test?
  • Are all parameters set correctly?
  • Are the expected results valid? Does the test case expect a single valid result, but the application returns multiple valid results?
  • Does the test case have any dependencies on earlier test cases that might have caused the problem? To avoid this situation, make test cases as modular and independent of each other as described in the blog article Build Maintainable Tests.
  • Did the teardown of the most recent test run work correctly? Is the AUT in the correct state, for example, with all browser windows closed? Has all the data entered during the last test run been deleted or reset?
  • Is there a timing issue? A study of flaky tests done by the University of Illinois at Urbana-Champaign found that flaky tests are often caused by asynchronous waits: the test fails because the AUT doesn’t return the expected result fast enough. In this case, it may be necessary to add a wait time to the test case step so that it doesn’t fail unnecessarily. For more information on how this works in Ranorex, refer to the user guide chapter Waiting for UI Elements.

Use your debugging tools

Make use of the tools available to you that may help resolve failing test cases. For example, Ranorex Studio provides several tools to assist in troubleshooting failed test cases, including the following:

Debugger

Debugger

This tool allows you to set breakpoints and step through a failed test case, examining the value of variables and expressions for each statement.

Maintenance Mode

Maintenance Mode

This tool allows you to identify and repair failing test cases directly from the test run report. Learn more in the following article: maintenance mode.

Ranorex Remote

Ranorex Remote

This is a great tool for troubleshooting test failures that occur on virtual machines. Use the Ranorex Remote Agent to update a run configuration to perform only the steps necessary to reach the point just before the failure occurred, so that the AUT is in the correct state. Then, connect to the virtual machine and troubleshoot the failed test case, as described in the blog article How to Reconstruct Failed Test Cases in CI Systems.

Taking the time to resolve your failed test cases, and to learn from the failures, will help make your entire test suite more reliable.