Systematic literature reviews in software engineering – A tertiary study

https://doi.org/10.1016/j.infsof.2010.03.006Get rights and content

Abstract

Context

In a previous study, we reported on a systematic literature review (SLR), based on a manual search of 13 journals and conferences undertaken in the period 1st January 2004 to 30th June 2007.

Objective

The aim of this on-going research is to provide an annotated catalogue of SLRs available to software engineering researchers and practitioners. This study updates our previous study using a broad automated search.

Method

We performed a broad automated search to find SLRs published in the time period 1st January 2004 to 30th June 2008. We contrast the number, quality and source of these SLRs with SLRs found in the original study.

Results

Our broad search found an additional 35 SLRs corresponding to 33 unique studies. Of these papers, 17 appeared relevant to the undergraduate educational curriculum and 12 appeared of possible interest to practitioners. The number of SLRs being published is increasing. The quality of papers in conferences and workshops has improved as more researchers use SLR guidelines.

Conclusion

SLRs appear to have gone past the stage of being used solely by innovators but cannot yet be considered a main stream software engineering research methodology. They are addressing a wide range of topics but still have limitations, such as often failing to assess primary study quality.

Introduction

In a series of three papers Kitchenham, Dybå and Jørgensen suggested that software engineers in general, and empirical software engineering researchers in particular, should adopt evidence-based practice as pioneered in the fields of medicine and sociology [1], [2], [3]. They proposed a framework for Evidence-based Software Engineering (EBSE), derived from medical standards, that relies on aggregating best available evidence to address engineering questions posed by practitioners and researchers. The most reliable evidence comes from aggregating all empirical studies on a particular topic. The recommended methodology for aggregating empirical studies is a systematic literature review (SLR) (see for example [4], [5], [6]). Kitchenham adapted the medical guidelines for SLRs to software engineering [7], and later updated them to include insights from sociology research [8].

SLRs are a means of aggregating knowledge about a software engineering topic or research question [5], [6], [7], [8]. The SLR methodology aims to be as unbiased as possible by being auditable and repeatable. SLRs are referred to as secondary studies and the studies they analyse are referred to as primary studies. There are two different types of SLRs:

  • Conventional SLRs aggregate results related to a specific research question e.g. “Is testing technique a more effective at defect detection than testing technique b?” If there are sufficient comparable primary studies with quantitative estimates of the difference between methods, meta-analysis can be used to undertake a formal statistically-based aggregation. However, we have found that meta-analysis is seldom possible for SLRs in software engineering because there are often insufficient primary studies.

  • Mapping studies. These studies aim to find and classify the primary studies in a specific topic area. They have coarser-grained research questions such as “What do we know about topic x?” They may be used to identify available literature prior to undertaking conventional SLRs. They use the same methods for searching and data extraction as conventional SLRs but rely more on tabulating the primary studies in specific categories. An example is the study of software engineering experiments [9] which led to a series of follow-on SLRs including [10], [11]. In addition, some mapping studies are concerned about how academics undertake research in software engineering (e.g. [13]) rather than what we know about a specific software engineering topic. The study reported in this paper is a mapping study.

This distinction between mapping studies and conventional SLRs can be somewhat fuzzy. Some mapping studies (like this one) provide a more detailed review of the topics covered in each primary study including issues such as major outcomes and quality evaluations of primary studies.

We believe secondary studies can play a vital role both in supporting further research efforts and also in providing information about the impact of methods and tools to assist software engineering practitioners and managers [2], [1]. However, these studies need to be readily available to those who would benefit from them. For example, researchers entering a new field would benefit from mapping studies in the area, whereas standards writers would benefit from conventional SLRs evaluating the benefits of specific techniques. Academics would also benefit from mapping studies and conventional SLRs when preparing teaching materials or writing text books. For this reason we believe it is important to catalogue and evaluate all such papers.

We recently published the results of a mapping study aimed at identifying software engineering SLRs [12]. The study is referred to as a tertiary study, because it was a SLR of secondary studies.The goal of the study was to identify how many SLRs had been published, what research topics were being addressed, and the limitations of current SLRs. For that study we used a manual search of a targeted set of 13 conferences and journals during the period January 1st 2004 to 30th June 2007. The sources were selected because they were known to include empirical studies and literature surveys, and had been used as sources for other mapping studies (e.g. [9], [13]). This search identified 20 SLRs of which eight were mapping studies and one a meta-analysis.

In this paper, we report the results of a broad automated search covering the period 1st January 2004 to 30th June 2008, and contrast them with our previous results. In effect we compare three sets of SLRs:

  • Those reported in the original study, covering the time period January 2004 to June 30th 2007 [12].

  • Those found in the time period January 2004 to June 30th 2007 that were found by the broad automated search and were not included in the original study. We discuss the differences between the results of the manual search and the broad automated search in [14].

  • Those found in the time period July 1st 2007 to June 30th 2008.

These are illustrated in Fig. 1. For convenience and to simplify referencing, these sets papers are respectively referred to as T1, T2-1 and T2-2 respectively in the rest of the paper (T for ‘tertiary’). The original study [12] is referred to as T1, this one as T2.

Section 2 reports our methodology. Section 3 reports data we extracted from each SLR. Section 4 answers our research questions. We report the limitations of our study in Section 5 and our conclusions in Section 6.

Section snippets

Method

We applied the basic SLR method as described by Kitchenham and Charters [8]. The main differences between the methods used in this study compared with the method used for the original study were that:

  • We used a broad automated search rather than a restricted manual search process.

  • Three researchers collected quality and classification data. For the papers found in the same time period as the original search, they took the median or mode value (as appropriate) as the consensus value. For the set

Data extraction results

The 33 SLRs that were published in the time period 1st January 2004 to 30th June 2008 (excluding those reported in the original tertiary study) are shown in Table 1. For each review we identify:

  • Whether it posed detailed technical questions (RQ) or was interested primarily in trends in a particular software engineering topic area (SERT) or the way in which software engineers undertake research (RT).

  • The quality score assigned to the study.

  • The year of publication.

  • Whether the study positioned

Discussion of research questions

This section addresses our specific research questions and identifies any changes between SLRs discussed in our original study and SLRs found in this study.

Study limitations

One of the major problems with SLRs is finding all the relevant studies. In this case, we used an automated search of six sources which found most of the papers we found in a previous manual search. However, the search missed three papers that should have been found, since it appears that they were not indexed when the original searches took place. The additional search performed in July 2009 found all papers that used conventional terminology and were mainstream software engineering papers.

Conclusions

The results of this study show two main changes compared with our previous study:

  • The number of SLRs being published appears to be increasing. However, it is still the case that many literature reviews are not performed in accordance with any methodology. Over the time period January 1st 2004 to 30th June 2008, we found 53 SLRs (of varying degrees of quality) but we also found 54 literature reviews that did not use any defined search strategy (see Section 3.2). This set of 54 studies does not

Acknowledgement

This study was funded by the UK Engineering and Physical Sciences Research Council project EPIC/E046983/1.

References (70)

  • T. Dybå et al.

    Evidence-based software engineering for practitioners

    IEEE Software

    (2005)
  • M. Jørgensen, T. Dybå, B.A. Kitchenham, Teaching evidence-based software engineering to university students, in: 11th...
  • Khan S. Khalid et al.

    Systematic Reviews to Support Evidence-based Medicine

    (2003)
  • A. Fink

    Conducting Research Literature Reviews. From the Internet to Paper

    (2005)
  • Mark Petticrew et al.

    Systematic Reviews in the Social Sciences: A Practical Guide

    (2005)
  • B.A. Kitchenham, Procedures for Undertaking Systematic Reviews, Joint Technical Report, Computer Science Department,...
  • B.A. Kitchenham, S. Charters, Guidelines for Performing Systematic Literature Reviews in Software Engineering Technical...
  • D.I.K. Sjøberg et al.

    A survey of controlled experiments in software engineering

    IEEE Transactions on Software Engineering

    (2005)
  • J. Hannay et al.

    The role of deliberate artificial design elements in software engineering experiments

    IEEE Transactions on Software Engineering

    (2008)
  • B. Kitchenham, P. Brereton, M. Turner, M. Niazi, P. Pretorius, D. Budgen, The impact of search procedures for...
  • C. Jefferies, P. Brereton, M. Turner, A systematic literature review to investigate reengineering existing systems for...
  • J. Bailey et al.

    Evidence relating to object-oriented software design: a survey

  • S. MacDonell et al.

    Comparing local and global effort estimation models reflections on a systematic reviews

  • L. Harjumaa, J. Markkula, M. Oivo, How does a Measurement Programme Evolve in Software Organizations? PROFES 2008, LNCS...
  • G.K. Hanssen, F.O. Bjørnson, and H. Westerheim, Tailoring and Introduction of the Rational Unified Process, EuroSPI...
  • P. Mohagheghi et al.

    Quality, productivity and economic benefits of software reuse: a review of industrial studies

    Empirical Software Engineering

    (2007)
  • M.N. Wicks et al.

    A new research agenda for tool integration

    Journal of Systems and Software

    (2007)
  • P. Mohagheghi, V. Dehlen, Where Is the Proof? – A Review of Experiences from Applying MDE in Industry, ECMDA-FA, LNCS...
  • R.C. de Boer, R. Farenhorst, In search of ’architectural knowledge’, in: SHARK ’08: Proceedings of the 3rd...
  • A.P. Freire, R. Goularte, R.P.M. Fortes, Techniques for developing more accessible web applications: a survey towards a...
  • M. Renger, G.L. Kolfschoten, G.-J. de Vreede, Challenges in Collaborative Modeling: A Literature Review, CIAO! 2008 and...
  • Centre for Reviews and Dissemination, What are the Criteria for the Inclusion of Reviews on DARE? 2007,...
  • F. Dong, M. Li, Y. Zhao, J. Li, Y. Yang, Software Multi-Project Resource Scheduling: A Comparative Analysis, ICSP 2008,...
  • M.I. Kamata, A.Y. Yoshida, H. Yoshida, N. Aoji, Figure out the current software requirements engineering – what...
  • A. Trifonova, S.U. Ahmed, L. Jaccheri, SArt: towards innovation at the intersection of software engineering and art,...
  • Cited by (0)

    View full text