iDFlakies: A Framework for Detecting

and Partially Classifying Flaky Tests

The findings from this work are now integrated into the Illinois Dataset of Flaky Tests (IDoFT). Please use the information there for our most up-to-date findings.


Scripts and tools for setting up and running the iDFlakies framework: https://github.com/iDFlakies/iDFlakies

Paper: http://mir.cs.illinois.edu/winglam/publications/2019/LamETAL19iDFlakies.pdf

If you use any of this work, please cite our corresponding paper:

@inproceedings{LamETAL19iDFlakies,
author = "Wing Lam and Reed Oei and August Shi and Darko Marinov and Tao Xie",
title = "{iDF}lakies: {A} framework for detecting and partially classifying flaky tests",
booktitle = "ICST 2019: 12th IEEE International Conference on Software Testing, Verification and Validation",
month = "April",
year = "2019",
address = "Xi'an, China",
pages = "312--322"
}


Details about all of the subjects that we investigated for order-dependent and non-order-dependent tests:


Details of the database generated from the framework:


Docker images of projects with flaky tests:


CSV input files for the framework:


Logs and dataset of all runs by the framework:


JSON files summarizing the order-dependent tests found for each module and the orders in which they passed/failed:

We thank Angello Astorga, Liia Butler, and Owolabi Legunsen for their discussions about flaky tests. This work was partially supported by National Science Foundation grants CCF1421503, CNS-1513939, CNS-1564274, CNS-1646305, CNS1740916, CCF-1763788, CCF-1816615, and OAC-1839010. We acknowledge support for research on flaky tests and test quality from Facebook, Google, Huawei, and Microsoft.

Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation (NSF).