Show simple item record

dc.contributor.authorMahmood, Zaheed
dc.contributor.authorBowes, David
dc.contributor.authorHall, Tracy
dc.contributor.authorLane, Peter
dc.contributor.authorPetric, Jean
dc.date.accessioned2018-07-31T12:31:35Z
dc.date.available2018-07-31T12:31:35Z
dc.date.issued2018-07-01
dc.identifier.citationMahmood , Z , Bowes , D , Hall , T , Lane , P & Petric , J 2018 , ' Reproducibility and Replicability of Software Defect Prediction Studies ' , Information and Software Technology , vol. 99 , pp. 148-163 . https://doi.org/10.1016/j.infsof.2018.02.003
dc.identifier.issn0950-5849
dc.identifier.urihttp://hdl.handle.net/2299/20303
dc.descriptionThis document is the Accepted Manuscript version of the following article: Zaheed Mahmood, David Bowes, Tracy Hall, Peter C. R. Lane, and Jean Petric, 'Reproducibility and replicability of software defect prediction studies', Information and Software Tecnology, Vol. 99: 148-163, July 2018. Under embargo until 10 February 2019. The final, published version is available online at doi: https://doi.org/10.1016/j.infsof.2018.02.003
dc.description.abstractContext: Replications are an important part of scientific disciplines. Replications test the credibility of original studies and can separate true results from those that are unreliable. Objective: In this paper we investigate the replication of defect prediction studies and identify the characteristics of replicated studies. We further assess how defect prediction replications are performed and the consistency of replication findings. Method: Our analysis is based on tracking the replication of 208 defect prediction studies identified by a highly cited Systematic Literature Review (SLR) [1]. We identify how often each of these 208 studies has been replicated and determine the type of replication carried out. We identify quality, citation counts, publication venue, impact factor, and data availability from all 208 SLR defect prediction papers to see if any of these factors are associated with the frequency with which they are replicated. Results: Only 13 (6%) of the 208 studies are replicated. Replication seems related to original papers appearing in the Transactions of Software Engineering (TSE) journal. The number of citations an original paper had was also an indicator of replications. In addition, studies conducted using closed source data seems to have more replications than those based on open source data. Where a paper has been replicated, 11 (38%) out of 29 studies revealed different results to the original study. Conclusion: Very few defect prediction studies are replicated. The lack of replication means that it remains unclear how reliable defect prediction is. We provide practical steps for improving the state of replication.en
dc.format.extent16
dc.format.extent426284
dc.language.isoeng
dc.relation.ispartofInformation and Software Technology
dc.subjectReplication
dc.subjectReproducibility
dc.subjectSoftware defect prediction
dc.subjectSoftware
dc.subjectInformation Systems
dc.subjectComputer Science Applications
dc.titleReproducibility and Replicability of Software Defect Prediction Studiesen
dc.contributor.institutionCentre for Computer Science and Informatics Research
dc.contributor.institutionSchool of Computer Science
dc.description.statusPeer reviewed
dc.date.embargoedUntil2019-02-10
dc.identifier.urlhttp://www.scopus.com/inward/record.url?scp=85043273528&partnerID=8YFLogxK
rioxxterms.versionofrecord10.1016/j.infsof.2018.02.003
rioxxterms.typeJournal Article/Review
herts.preservation.rarelyaccessedtrue


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record