2015 IEEE 26th International Symposium on Software Reliability Engineering (ISSRE) (2015)
Gaithersbury, MD, USA
Nov. 2, 2015 to Nov. 5, 2015
Rahul Gopinath , Department of EECS, Oregon State University
Amin Alipour , Department of EECS, Oregon State University
Iftekhar Ahmed , Department of EECS, Oregon State University
Carlos Jensen , Department of EECS, Oregon State University
Alex Groce , Department of EECS, Oregon State University
Mutation analysis is considered the best method for measuring the adequacy of test suites. However, the number of test runs required for a full mutation analysis grows faster than project size, which is not feasible for real-world software projects, which often have more than a million lines of code. It is for projects of this size, however, that developers most need a method for evaluating the efficacy of a test suite. Various strategies have been proposed to deal with the explosion of mutants. However, these strategies at best reduce the number of mutants required to a fraction of overall mutants, which still grows with program size. Running, e.g., 5% of all mutants of a 2MLOC program usually requires analyzing over 100,000 mutants. Similarly, while various approaches have been proposed to tackle equivalent mutants, none completely eliminate the problem, and the fraction of equivalent mutants remaining is hard to estimate, often requiring manual analysis of equivalence. In this paper, we provide both theoretical analysis and empirical evidence that a small constant sample of mutants yields statistically similar results to running a full mutation analysis, regardless of the size of the program or similarity between mutants. We show that a similar approach, using a constant sample of inputs can estimate the degree of stubbornness in mutants remaining to a high degree of statistical confidence, and provide a mutation analysis framework for Python that incorporates the analysis of stubbornness of mutants.
R. Gopinath, A. Alipour, I. Ahmed, C. Jensen and A. Groce, "How hard does mutation analysis have to be, anyway?," 2015 IEEE 26th International Symposium on Software Reliability Engineering (ISSRE), Gaithersbury, MD, USA, 2015, pp. 216-227.