DocumentCode :
2372956
Title :
Repeatability of systematic literature reviews
Author :
Kitchenham, B. ; Brereton, P. ; Zhi Li ; Budgen, David ; Burn, Andrew
Author_Institution :
Sch. of Comput. & Math., Keele Univ., Keele, UK
fYear :
2011
fDate :
11-12 April 2011
Firstpage :
46
Lastpage :
55
Abstract :
Background: One of the anticipated benefits of systematic literature reviews (SLRs) is that they can be conducted in an auditable way to produce repeatable results. Aim: This study aims to identify under what conditions SLRs are likely to be stable, with respect to the primary studies selected, when used in software engineering. The conditions we investigate in this report are when novice researchers undertake searches with a common goal. Method: We undertook a participant-observer multi-case study to investigate the repeatability of systematic literature reviews. The "cases" in this study were the early stages, involving identification of relevant literature, of two SLRs of unit testing methods. The SLRs were performed independently by two novice researchers. The SLRs were restricted to the ACM and IEEE digital libraries for the years 1986-2005 so their results could be compared with a published expert literature review of unit testing papers. Results: The two SLRs selected very different papers with only six papers out of 32 in common, and both differed substantially from a published secondary study of unit testing papers finding only three of 21 papers. Of the 29 additional papers found by the novice researchers, only 10 were considered relevant. The 10 additional relevant papers would have had an impact on the results of the published study by adding three new categories to the framework and adding papers to three, otherwise empty, cells. Conclusions: In the case of novice researchers, having broadly the same research question will not necessarily guarantee repeatability with respect to primary studies. Systematic reviews must be careful to report their search process fully or they will not be repeatable. Missing papers can have a significant impact on the stability of the results of a secondary study.
Keywords :
digital libraries; software engineering; software reviews; ACM digital libraries; IEEE digital libraries; SLR; participant-observer multi-case study; repeatability; software engineering; systematic literature reviews; Case Study; Repeatability; Systematic Literature Review;
fLanguage :
English
Publisher :
iet
Conference_Titel :
Evaluation & Assessment in Software Engineering (EASE 2011), 15th Annual Conference on
Conference_Location :
Durham
Electronic_ISBN :
978-1-84919-509-6
Type :
conf
DOI :
10.1049/ic.2011.0006
Filename :
6083161
Link To Document :
بازگشت