Even when test cases have been carefully designed to be stable and maintainable, test failures can happen. There are several possible uses of the term “test failure,” so let’s distinguish between them:
A negative test case
This is a test case that you expect to return an error from the application under test (AUT), such as an invalid password. This type of test case succeeds when it returns the expected error message.
A test case that uncovers a defect in the application
This is actually a successful test case because identifying defects is one of the main goals of software testing. Consider this type of test case for inclusion in your regression test set.
A test case that fails for a reason unrelated to the functionality of the application
This is the meaning of the term “failed test case” as used in this article.
When a test case fails, the first job is to decide whether situation #2 or #3 applies: Did the test case fail due to a defect in the AUT, or is the problem with the test case itself, such as missing or invalid test data, problems with the test environment, or changes in the AUT that are not defects? If it is not immediately clear, you may need to troubleshoot the test case itself before reporting a defect in the application.
It may be tempting to simply re-run a failed test case to see if it passes. But a test case that passes sometimes and fails on other occasions for no discernable reason is a “flaky,” unreliable test case. It’s important to resolve the issue that caused it to fail so that you can have confidence in the results of your automated testing.
Configure test runs to assist debugging
An earlier article in this series, “Build Maintainable Tests” described best practices for designing test cases that make them more stable and less likely to fail. These included eliminating dependencies between test cases as much as possible, ensuring that your test environment is stable, and removing tests that you expect to fail (such as ones for unresolved defects) from the test run. It is also helpful to configure your test cases to take a screenshot when a failure occurs.
In addition to these recommendations, be sure to configure the test run to handle failures appropriately. Only allow a failing test to stop the entire test run if that makes sense for the situation – for example, if the application fails to launch, or smoke tests fail. Ranorex Studio’s modular approach to test case design includes several options for continuing after a test case returns an error, including “continue with iteration,” “continue with
It’s also important to manage the size of test run reports by focusing only on true errors and failures. For example, Ranorex supports multiple pre-defined report levels, including “debug,” “information,” “warning,” and “success.” In a large test run, reporting information at this level may result in an excessive amount of data. Consider reporting results only for the “error” and “failure” levels to make it easier to spot true problems that need to be resolved.
Isolate the problem
If many test cases are failing, look for a problem with the environment, test framework, or the AUT.
Issues with the environment can include required services not running, or not running in administrative mode if required.
Look for issues with the test framework, such as a licensing error, or a remote agent not configured properly.
Application Under Test
Verify that the AUT is prepared correctly. This can include issues such as location-specific system settings, the wrong browser version, or even a different system language. Or, there could be a pending O/S update that blocks the user interface.
Troubleshoot failed test cases
Work through a probable-cause checklist to troubleshoot each failed test case, asking questions such as the following:
- Is the test case up-to-date with the AUT? For example, has the test case been updated with any/all changes in UI elements?
- Is the input data correct and available to the test?
- Are all parameters set correctly?
- Are the expected results valid? Does the test case expect a single valid result, but the application returns multiple valid results?
- Does the test case have any dependencies on earlier test cases that might have caused the problem? To avoid this situation, make test cases as modular and independent of each other as described in the blog article Build Maintainable Tests.
- Did the teardown of the most recent test run work correctly? Is the AUT in the correct state, for example, with all browser windows closed? Has all the data entered during the last test run been deleted or reset?
- Is there a timing issue? A study of flaky tests done by the University of Illinois at Urbana-Champaign found that flaky tests are often caused by asynchronous waits: the test fails because the AUT doesn’t return the expected result fast enough. In this case, it may be necessary to add a wait time to the test case step so that it doesn’t fail unnecessarily. For more information on how this works in Ranorex, refer to the user guide chapter Waiting for UI Elements.
Use your debugging tools
Make use of the tools available to you that may help resolve failing test cases. For example, Ranorex Studio provides several tools to assist in troubleshooting failed test cases, including the following:
This tool allows you to set breakpoints and step through a failed test case, examining the value of variables and expressions for each statement.
This tool allows you to identify and repair failing test cases directly from the test run report. Learn more in the following article: maintenance mode.
This is a great tool for troubleshooting test failures that occur on virtual machines. Use the Ranorex Remote Agent to update a run configuration to perform only the steps necessary to reach the point just before the failure occurred, so that the AUT is in the correct state. Then, connect to the virtual machine and troubleshoot the failed test case, as described in the blog article How to Reconstruct Failed Test Cases in CI Systems.
Taking the time to resolve your failed test cases, and to learn from the failures, will help make your entire test suite more reliable.