Abstract | ||
---|---|---|
Much research on software testing and test techniques relies on experimental studies based on mutation testing. In this paper we reveal that such studies are vulnerable to a potential threat to validity, leading to possible Type I errors; incorrectly rejecting the Null Hypothesis. Our findings indicate that Type I errors occur, for arbitrary experiments that fail to take countermeasures, approximately 62% of the time. Clearly, a Type I error would potentially compromise any scientific conclusion. We show that the problem derives from such studies’ combined use of both subsuming and subsumed mutants. We collected articles published in the last two years at three leading software engineering conferences. Of those that use mutation-based test assessment, we found that 68% are vulnerable to this threat to validity. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1145/2931037.2931040 | ISSTA |
Field | DocType | Citations |
Computer science,Null hypothesis,Theoretical computer science,Artificial intelligence,Type I and type II errors,Machine learning,Reliability engineering,Software testing | Conference | 29 |
PageRank | References | Authors |
0.73 | 48 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Mike Papadakis | 1 | 1114 | 52.77 |
Christopher Henard | 2 | 383 | 10.88 |
Mark Harman | 3 | 10264 | 389.82 |
Yue Jia | 4 | 2234 | 66.02 |
Yves Le Traon | 5 | 3922 | 190.39 |