Background[edit]

As software is updated or changed, or reused on a modified target, emergence of new faults and/or re-emergence of old faults is quite common.


Sometimes re-emergence occurs because a fix gets lost through poor revision control practices (or simple human error in revision control). Often, a fix for a problem will be "fragile" in that it fixes the problem in the narrow case where it was first observed but not in more general cases which may arise over the lifetime of the software. Frequently, a fix for a problem in one area inadvertently causes a software bug in another area.


It may happen that when a feature is redesigned some of the same mistakes that were made in the original implementation of the feature also occur in the redesign. In most software development situations, it is considered good coding practice, when a bug is located and fixed, to record a test that exposes the bug and re-run that test regularly after subsequent changes to the program.[5]


Although this may be done through manual testing procedures using programming techniques, it is often done using automated testing tools.[6] Such a test suite contains software tools that allow the testing environment to execute all the regression test cases automatically; many projects have automated Continuous integration systems to re-run all regression tests at specified intervals and report any failures (which could imply a regression or an out-of-date test).[7]


Common strategies are to run such a system after every successful compile (for small projects), every night, or once a week. Those strategies can be automated by an external tool.


Regression testing is an integral part of the extreme programming software development method. In this method, design documents are replaced by extensive, repeatable, and automated testing of the entire software package throughout each stage of the software development process. Regression testing is done after functional testing has concluded, to verify that the other functionalities are working.


In the corporate world, regression testing has traditionally been performed by a software quality assurance team after the development team has completed work. However, defects found at this stage are the most costly to fix. This problem is being addressed by the rise of unit testing. Although developers have always written test cases as part of the development cycle, these test cases have generally been either functional tests or unit tests that verify only intended outcomes. Developer testing compels a developer to focus on unit testing and to include both positive and negative test cases.[8]

General prioritization – Prioritize test cases that will be beneficial on subsequent versions

Version-specific prioritization – Prioritize test cases with respect to a particular version of the software.

Benefits and drawbacks[edit]

Regression testing is performed when changes are made to the existing functionality of the software or if there is a bug fix in the software. Regression testing can be achieved through multiple approaches; if a test all approach is followed, it provides certainty that the changes made to the software have not affected the existing functionalities, which are unaltered.[10]


In agile software development—where the software development life cycles are very short, resources are scarce, and changes to the software are very frequent—regression testing might introduce a lot of unnecessary overhead.[10]


In a software development environment which tends to use black box components from a third party, performing regression testing can be tricky, as any change in the third-party component may interfere with the rest of the system (and performing regression testing on a third-party component is difficult, because it is an unknown entity).[10]

Quality control

Test-driven development

Microsoft regression testing recommendations