You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

8 September 2022

Evidence-Based Software Engineering: A Checklist-Based Approach to Assess the Abstracts of Reviews Self-Identifying as Systematic Reviews

and
Lassonde School of Engineering, York University, 4700 Keele St, Toronto, ON M3J 1P3, Canada
*
Author to whom correspondence should be addressed.
This article belongs to the Topic Software Engineering and Applications

Abstract

A systematic review allows synthesizing the state of knowledge related to a clearly formulated research question as well as understanding the correlations between exposures and outcomes. A systematic review usually leverages explicit, reproducible, and systematic methods that allow reducing the potential bias that may arise when conducting a review. When properly conducted, a systematic review yields reliable findings from which conclusions and decisions can be made. Systematic reviews are increasingly popular and have several stakeholders to whom they allow making recommendations on how to act based on the review findings. They also help support future research prioritization. A systematic review usually has several components. The abstract is one of the most important parts of a review because it usually reflects the content of the review. It may be the only part of the review read by most readers when forming an opinion on a given topic. It may help more motivated readers decide whether the review is worth reading or not. But abstracts are sometimes poorly written and may, therefore, give a misleading and even harmful picture of the review’s contents. To assess the extent to which a review’s abstract is well constructed, we used a checklist-based approach to propose a measure that allows quantifying the systematicity of review abstracts i.e., the extent to which they exhibit good reporting quality. Experiments conducted on 151 reviews published in the software engineering field showed that the abstracts of these reviews had suboptimal systematicity.

1. Introduction

Systematic reviews are well established in the software engineering field [1]. A systematic review relies on explicit and systematic methods to collect and synthesize findings of studies that focus on a given research question [2]. It provides high-quality evidence-based syntheses for efficacy under real-world conditions and allows understanding the correlations between exposures and outcomes [3]. A systematic review also allows making recommendations on how to act based on the review findings and help support future research prioritization [4,5]. The core features of a methodologically sound systematic review include transparency, replicability, and reliance on clear eligibility criteria [6]. Systematic reviews are usually called secondary studies, and the documents they synthesize are usually called primary studies [1]. Systematic reviews are usually resource intensive, e.g., time-consuming, as they usually require that a team of systematic reviewers, i.e., the authors of the review, work over a long period of time (approximately 67.3 weeks) to complete the review [6].
Systematic reviews are crucial for various stakeholders because they allow them to make evidence-based decisions without being overwhelmed by a large volume of research [4,7,8,9,10]. These stakeholders include patients, healthcare providers, policy makers, educators, students, insurers, systematic groups/organizations, researchers, guideline developers, journal editors, and evidence users [4,9,10]. Systematic reviews are increasingly popular and are conducted in various areas such as health care, social sciences, education, and computer science [2,8].
Systematic reviews that are clearly and completely reported allow the reader to understand how they were planned and conducted, to judge the extent to which the results are reliable, to judge the reproducibility and appropriateness of the methods, and to rely on the evidence to make informed recommendations for practice or policy [8,11]. However, systematic reviews are sometimes poorly conducted and reported, exhibit low methodological rigor and reporting quality, and may result in misleading evidence-based decision-making and in the waste of various (non)financial resources [12,13]. To tackle these issues, reporting guidelines have been developed to support various tasks such as review reporting and assessment.
The most widespread reporting guidelines usually consist of a statement paper describing how the guideline has been developed and presenting the guideline in the form of a checklist that often comprises up to thirty items that need to be reported when writing a systematic review [8]. When reporting a review, it is crucial to follow such guidelines to ensure that the review is transparent, complete, beneficial, trustworthy, and unbiased [14]. According to the EQUATOR Network registry, there are at least 529 registered reporting guidelines as of August 2022. These reporting guidelines include CONSORT, STRICTA, ARRIVE and PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analyses) [15,16]. In particular, PRISMA 2020 (the most recent version of PRISMA) was released in 2021 and comprises two checklists: the PRISMA 2020 checklist that consists of 27 items and the PRISMA 2020 for abstracts checklist that consists of 12 items. PRISMA’s recommendations have been widely adopted by several areas including the health and medical area, social sciences, education, and computer science [2]. As of August 2020, PRISMA yielded citations in more than sixty thousand reports [4]. Besides, PRISMA has been endorsed by more than two hundred journals as well as systematic review organizations [4]. In this work, we focus on the PRISMA reporting guideline as it yields a more concise and transparent reporting than other reporting guidelines [17] and is suitable for the systematic reviews reporting.
Several approaches have been proposed to assess the extent to which studies have adhered to reporting guidelines, such as CONSORT, COREQ, TRIPOD, STROBE, PRISMA, and STRICTA. Thabane et al. [10], Whiting et al. [18], Shea et al. [19,20], Pussegoda et al. [12], Jin et al. [21], Blanco et al. [22], Heus et al. [23], Walsh et al. [24], Fernández-Jané et al. [15], Logullo et al. [8], and Sargeant et al. [5] discussed such approaches. Some of these approaches focus on the assessment of the adherence of systematic reviews to reporting guidelines and/or on the assessment of the risk of bias in systematic reviews. Most of these approaches are either qualitative or quantitative. Qualitative approaches usually rely on ratings scales (e.g., low, medium) to assess the adherence to reporting guidelines. Quantitative approaches often rely on frequencies and/or percentages to assess that adherence. Most of the work that has assessed the methodological rigor and the reporting quality of systematic reviews has concluded that adherence to existing reporting guidelines is usually suboptimal [13,15,21,22]. However, none of that work is suitable for the assessment of existing reviews published in the software engineering field. This calls for the proposal of new approaches to assess existing reviews and provide to stakeholders of systematic reviews a quantitative indicator that will allow them to determine whether the review at hand is sufficiently systematic and that its findings are therefore credible enough to support decision-making.
As Page et al. [2] and Andrade [25] pointed out, the abstract is one of the most important components of a review as it usually synthesizes the content of a review. It may be the only part of the review that most readers will read when forming an opinion on a given topic. It may help readers decide whether the review content is worth reading or not. However, abstracts are sometimes poorly written and may therefore give a misleading and even harmful picture of the review’s contents. To assess the reporting quality of review abstracts published in the software engineering field, we proceeded as follows: (1) we introduce the Check-SE 2022 checklist for abstracts, an eleven-item question-based checklist that is derived from the PRISMA 2020 for abstracts checklist [26], and that embodies the best reporting guidelines and/or procedures the abstract of a review published in the SE (software engineering) field should comply with to exhibit good reporting quality; and (2) we introduce a measure that allows assessing compliance with that checklist. That measure quantifies the systematicity of review abstracts by assessing their reporting quality. Experiments conducted on published reviews claiming to be systematic shows that the systematicity of their abstracts is suboptimal. More efforts should therefore be made to increase the systematicity of reviews, starting from their abstracts. We further describe our work in the remainder of this paper.

3. An Approach to Support the Systematicity Assessment

In this section, we introduce the notion of systematicity in the context of reviews, and propose a checklist-based approach to assess that systematicity.

3.1. What Is Systematicity?

As stated above, according to Moher et al. [9,27] and Page et al. [2], a systematic review is a “review that uses explicit, systematic methods to collate and synthesize findings of studies that address a clearly formulated question”. A systematic review, therefore, focuses on a clearly formulated research question and relies on systematic, reproducible, and explicit methods to support the identification, selection, and critical appraisal of relevant research meeting some eligibility criteria as well as the collection and analysis of data retrieved from that research. A systematic review also relies on explicit, reproducible, and systematic methods that allow reducing the potential bias that may arise when identifying, selecting, synthesizing, and summarizing studies.
In that context, we define the systematicity of a review as the extent to which it is systematic, i.e., the extent to which a review exhibits good reporting quality. Based on the literature, we can conclude that the systematicity property is usually akin to several sub-properties, such as the completeness of the reporting, the transparency of the reporting, the trustworthiness in the review findings, the appropriateness of the methods used to conduct the review, the unbiasedness of the reporting, and the reproducibility of the study [2,6,8,14,29,39]. Given a review that was planned, conducted, and reported using a specific reporting guideline, we can also perceive the systematicity as the extent to which systematic reviewers who authored that review have adhered with the target reporting guideline. In this regard, Blanco et al. [22] defined adherence as the “Action(s) taken by authors to ensure that a research report is compliant with the items recommended by the appropriate/relevant reporting guideline. These can take place before or after the first version of the manuscript is published.”
Based on the literature, e.g., Thabane et al. [10], Fernández-Jané et al. [15], Logullo et al. [8], and Oliveira et al. [13], Belur et al. [6], focusing on reporting quality and methodological rigor, we can conclude that the notion of systematicity is usually subjective, variable from one review to another and suboptimal given that each review does not usually abide by the same standards of quality and methodological rigor. The rationale is that some systematic reviewers usually prefer to report the items of reporting guidelines that seem the most important to them and to not report the remaining ones [5]. Besides, as Belur et al. [6] pointed out, conducting a systematic review usually implies making subjective decisions at various points of the review process. These decisions that are usually made by multiple systematic reviewers may hinder the reproducibility of the review if that review does not report with sufficient details how coding and screening decisions were made and how were resolved the potential disagreements that occurred between systematic reviewers involved in the review process. Relying on quantitative assessment to assess the systematicity of reviews will, therefore, foster the objective and consistent comparison of adherence to reporting guidelines across studies [23]. Besides, quantitative assessment will help decision makers to distinguish high-quality reviews from low-quality ones and therefore to avoid accepting right away the findings of a single systematic review without second-guessing their credibility [20].
Similar to the literature, e.g., Shea et al. [19,20], and Whiting et al. [18], we follow in this paper a checklist-based approach and, therefore, rely on a set of eleven items, i.e., on a set of eleven questions, to quantitatively assess the systematicity of the reviews that self-identify as “systematic reviews”. We further describe these questions below.

3.2. A Checklist-Based Approach to Quantitatively Assess Review Abstracts

In this paper, we chose to focus on the systematicity of abstracts because the abstract is one of the most important components of a systematic review given it usually synthesizes the content of the review. In this regard, Page et al. [4] and Kitchenham et al. [26] stated that after reading the information a review abstract provides, readers should be able to decide if they want to access the full content of the review or not. Furthermore, as Andrade [25] pointed out, the abstract is usually the only part of an article, e.g., a review, that a referee reads when choosing to accept the invitation made by an editor to review a manuscript. Besides, the abstract may be the only part of a review that most readers will read when forming an opinion on a research topic.
As Page et al. [4] and Kitchenham et al. [26] pointed out, an abstract should, therefore, report the main information regarding the key objective(s) or question(s) that the review deals with. It should also specify the methods, results, and implications of the review findings. Besides, some readers, may only have access to a review abstract, e.g., in the case of journals that require a payment/subscription to access their articles. Before requesting the full content of a review, it is therefore critical for such readers to find in the abstract sufficient information about the main objective(s) or question(s) that the review addresses, methods, findings, and implications of the findings for future practice, policy, and future research. Besides, the keywords that are part of a review abstract are usually used to index the systematic review in bibliographic databases. It is, therefore, recommended to properly write abstracts so that they report keywords that accurately specify the review question, e.g., population, interventions, outcomes. However, as Andrade [25] pointed out, abstracts are sometimes poorly written, may lack crucial information, and may convey a biased and misleading overview of the review they summarize.
Similar to the literature, e.g., Shea et al. [19,20], Whiting et al. [18], to assess the systematicity of the selected review abstracts, we analyzed them using a checklist-based approach. The latter relies on a set of eleven questions. We have inferred these questions from a thorough analysis of the literature focusing on existing reporting guidelines and procedures. In our work, we focused on:
(a)
The analysis of reporting guidelines and procedures, including the PRISMA reporting guideline that was discussed in Page et al. [2,4,11], Moher et al. [9] and Kitchenham et al. [26]. That analysis also focused on the reporting procedures discussed by Kitchenham [32], Petticrew and Roberts [33], Kitchenham [34], Pickering and Byrne [35], Wohlin [36], and Okoli [30].
(b)
The analysis of references explaining how to properly write abstracts. These include: [25], the references describing the PRISMA 2020 for abstracts checklist (e.g., [2,4]) and the reference describing revisions that Kitchenham et al. [26] made to that abstract’s checklist in the context of the software engineering field.
From our analysis, we decided to mainly focus on PRISMA because it is a well-established reporting guideline that federates several existing reporting procedures used in several research areas, including computing. The PRISMA statement embeds several good reporting practices used in several disciplines including software engineering. The PRISMA 2020 checklist focuses on the adequate reporting of the whole systematic review, while the PRISMA 2020 for abstracts checklist focuses on the adequate reporting of the abstract of a review. Kitchenham et al. [26] further explained how to apply PRISMA 2020 in the software engineering field. In particular, Kitchenham et al. [26] provided supplementary explanations and examples of the SEGRESS (Software Engineering Guidelines for REporting Secondary Studies) items. These explanations of the SEGRESS items mainly focus on quantitative and qualitative systematic reviews. These guidelines derive from the PRISMA 2020 statement (and some of its extensions), the guidelines used to carry out qualitative reviews, and the guidelines used to carry out realist syntheses. Kitchenham et al. [26] revised some of the existing PRISMA 2020 explanations to make them suitable for software engineering research.
We also decided to focus on PRISMA 2020 because it has been endorsed by more than 200 journals, and it embeds more recent reporting practices and, therefore, reflects the advances in methods used to plan and conduct systematic reviews [2,4,11,31].
Our analysis allowed us to propose the Check-SE 2022 checklist for abstracts, a modified question-based version of the PRISMA 2020 checklist for abstracts. The Check-SE 2022 checklist for abstracts contains eleven items and is derived from the PRISMA 2020 for abstracts checklist and from Kitchenham et al. [26], and that applies to the reporting of reviews’ abstracts published in the SE (software engineering) field. The Check-SE 2022 checklist for abstracts embodies the best reporting procedures and/or guidelines that the reviews’ abstracts published in the SE (software engineering) field should comply with to exhibit good reporting quality.
Table A1 in Appendix A reports the PRISMA 2020 for abstracts checklist. Our first ten questions only consider the first ten items in the PRISMA 2022 for abstracts checklist. Thus, our questions do not take into account the 11th and the 12th item of the PRISMA 2022 for abstracts checklist, as, according to Kitchenham et al. [26], these two items do not usually apply to the software engineering field. Besides, in the software engineering community, researchers usually perform quality assessment instead of risk of bias evaluations [26] as the fifth item of the PRISMA 2020 for abstracts checklist recommends. Our fifth question does not, therefore, focus on risk of bias as the PRISMA 2020 for abstracts checklists recommends, but on the quality assessment. Besides, we have noticed that some software engineering journals usually recommend having well-formed structured abstracts by using keywords that embody the review structure, e.g., context, background, methods, results, conclusion. We, therefore, consider that recommendation as an additional reporting guideline item, and our 11th question addresses that item by attempting to determine whether existing reviews comprise structured abstracts. This is consistent with Andrade [25] who stated that journals usually require abstracts to have a formal structure consisting of the following sections: background, methods, results, and conclusions. Some journals may recommend using additional sections, e.g., objectives, limitations, when reporting abstracts.
To assess the systematicity of a review abstract, we extracted data from the abstract to populate the form that Table 1 depicts. This allowed us to answer the eleven questions that our checklist-based approach comprises. The first row of Table 1 specifies the review reference information, e.g., names of the review authors, the publication year, and the title of the review. The second column of Table 1 shows the questions we created to assess the systematicity of reviews abstracts. The third column of Table 1 indicates the sentence(s)/keyword(s) of a reference abstract’s that allows answering a given question. The third column of Table 1 indicates the confidence, i.e., certainty, in the answers to each question, using a 3-point scale that is widely adopted in the literature (e.g., [1,20]).
Table 1. Question-Driven Data Extraction Form to Assess the Systematicity of Review Abstracts.
That confidence can take one of the following values: Certain (i.e., 1), Partially certain (i.e., 0.5), or Not certain (i.e., 0). That confidence therefore quantifies the extent to which the abstract review provides information that allows answering the so-called question.
According to Kitchenham et al. [26], the gray cells in Table 1 correspond to items in the PRISMA 2020 for abstracts checklist that are crucial in the Software Engineering field, and that every review published in this field should report. The other items (unshaded cells) seem less crucial and may not be reported, especially if the lack of space prevents systematic reviewers from reporting them. The rationale is that abstracts in software engineering journals are usually short, between 100 and 250 words. Software engineering journals therefore have more stringent limits regarding the length of abstracts than medical journals [26] for which systematic reviewers are offered more space to report all the items of the PRISMA 2020 for abstracts checklist.
In Table 1, we refer to the overall confidence in the systematicity of the abstract of a review as OConfidence. We compute the OConfidence value of a review as a percentage whose maximal value is 100% and whose minimal value is 0%. For a given review abstract Absi, we therefore compute the OConfidence as the sum of the confidence in the answer to each individual question multiplied by one hundred and divided by the total number of questions in our checklist. Equation (1) formalizes that reasoning. In that equation, |Questions| represents the number eleven, i.e., the total number questions reported in Table 1. Recall from above that for a given review abstract Absi, the confidence in the answer Ak provided for a given question Qk, can be one of the following: 1 (certain), 0.5 (partially certain) or 0 (not certain). Here, k designates the identifier of a question and can vary from one to eleven.
O C o n f i d e n c e   ( A b s i ) = 100   k = 1 | Q u e s t i o n s |   c o n f i d e n c e ( A k ,   Q k ) | Q u e s t i o n s |  
Similar to the literature, e.g., Weaver et al. [40], Boaye Belle et al. [41], we also used a n-level scale to interpret the values of the overall confidence based on their respective ranges. That scale varies from Invalid (i.e., 1) to Valid (i.e., 10), with n = 10 in their case. We adopted that scale in a more quantitative manner, considering that the confidence and relevance are expressed in percentage and vary from 0% to 100% instead of 1 to 10. Table 2 therefore specifies the various interpretations of the OConfidence measure based on its value range. Table 2 maps a confidence specified in percentage to a confidence expressed using a level. The latter allows indicating to the review stakeholders the level of systematicity that a given review yields. For instance, a review’s abstract whose OConfidence is equal to 29% (i.e., belonging to the 20–40 range) has a Low-level confidence i.e., exhibits a low systematicity.
Table 2. Interpretation of the Overall Confidence, from Boaye Belle et al. [41].

4. Empirical Evaluation

In the remainder of this section, we present the experiments we have carried out to validate our approach.

4.1. Dataset

We have run our experiments on a dataset consisting of reviews that self-identified as systematic reviews. To obtain these references, we searched the following well established databases that are popular in the software engineering field: Google Scholar, IEEE Software and ACM Digital Library. We used the following query and specifying a year range varying from 2012 and 2022:
  • IEEE (using the Command Search, and searching in metadata): (((“All Metadata”: ”systematic review”) OR (“All Metadata”: ”PRISMA review”)) AND (“All Metadata”: ”software engineering”))
  • Google scholar (using the Publish or Perish tool, and searching in the titles): (“systematic review” OR “PRISMA review”) AND (“software engineering”)
  • ACM (using the Advanced Search option and searching in the titles): (“systematic review” OR “PRISMA review”) AND (“software engineering”).
Searching the three databases using that query identified 267 reviews. We imported these references into Endnote (a reference manager) by creating an EndNote library of them. We then eliminated the duplicates in the so-obtained EndNote reference list. After that, we analyzed the content of the corresponding reviews and decided to include in our dataset the reviews that met the following inclusion criteria:
-
It was written in English
-
It was a conference paper, a journal paper, or a workshop paper
-
It was peer-reviewed
-
It was published between 2012 and 2022
-
It self-identified as a systematic review
-
It discussed software engineering concepts. If the reference deals with interdisciplinary concepts i.e., focuses on several fields, then software engineering should be part of these fields.
We did not include in our dataset the reviews that met the following exclusion criteria:
-
It was the reference of a magazine, a chapter of the book, a book, conference Protocol, or a tutorial
-
It was the reference of a paper that does not present results
-
It was the reference of a paper that cannot be downloaded and/or is not available
-
It was not related to software engineering
-
It was the reference of a very short paper (i.e., a paper that has less than four pages).
Applying these inclusion and exclusion criteria allowed us to include 151 reviews in our dataset. Eight of these reviews have complied with PRISMA; 143 of them complied with other reporting procedures/methods such as the ones discussed in [32,34,42,43].

4.2. Research Questions

Our experiments focused on the dataset described above, and aimed at answering the following research questions:
  • (RQ1): To what extent are the abstracts of reviews (that self-identify as systematic reviews) systematic? Here we rely on the answers to the eleven questions specified above to assess the systematicity of the reviews that are included in our dataset.
  • (RQ2): To what extent does the publication year influence the systematicity of the analyzed review abstracts?
  • (RQ3): Does the use of PRISMA increase the systematicity of the review i.e., their reporting quality?
  • (RQ4): What are the items that were appropriately reported in the review abstracts i.e., what are the questions to which most of the abstracts of the analyzed reviews were able to provide answers? Investigating this question will help us determine the minimum set of core abstract items to which systematic literature reviews commonly comply with in the software engineering field.
  • (RQ5): Does the systematicity of the review abstracts influence the number of citations of these reviews? The rationale is that the more a review abstract is well-reported, the more the user may feel like the review findings are credible, may be motivated to read the full content of the review to eventually exploit its results, and may decide to cite the review as a relevant related work for instance.
  • (RQ6): How do systematic reviewers usually structure reviews?
When discussing the adherence of reviews to reporting guidelines, Logullo et al. [8] stressed that some researchers estimate that a 100% adherence to the checklist items of a reporting guideline would mean that a review constitutes “a quality paper”. However, in the currently published research, it is virtually impossible to find a review that yields a 100% adherence with the target reporting guideline. Hence, that “top quality” may not be reachable and reviews adhering to, say, 80% of the items should be considered as “well reported”. A threshold equal to 80% seems to be an optimal value and involves the coverage of most items in the reporting guideline checklist. Based on that reasoning, we therefore considered that if the overall confidence in the systematicity of an analyzed review abstract is at least 80% (i.e., High), then that review abstract exhibits an optimal systematicity. Otherwise, that abstract exhibits a suboptimal systematicity.

4.3. Data Collection and Analysis Process

To answer the six research questions specified above, we analyzed the reviews included in our dataset and, for each analyzed review abstract, we populated the question-driven data extraction form that Table 1 reports. Similar to Fernandez-Jané et al. [15], we tested our question-based data extraction form with pilot data coming from three systematic reviews. This allows ensuring the usability of the data extraction form and helped solve potential disagreements on its understanding. We also included data of the pilot test in the final analysis. Two researchers then extracted data from the abstracts of the reviews and assessed them. More specifically, one researcher iteratively extracted all the data from the reviews included in the dataset by populating the question-driven data extraction forms (see Table 1) and specifying the individual confidence values for each answer to a given question. A more experienced researcher then validated the extracted data by double-checking the content of the question-driven data extraction forms. After the validation of the extracted data, one researcher then used Equation (1) to compute the overall confidence in the systematicity of each review’s abstract.
To further answer RQ1, RQ2, RQ3, and RQ4, we created Python scripts that allowed aggregating the overall confidence values obtained for all the abstracts of the reviews included in the dataset. This allowed us to create the charts used to answer these four research questions.
To answer RQ5, we used Google Scholar to retrieve the number of citations of each review as of July 2022. We then computed the Spearman correlation between the systematicity of the review abstracts and their number of Google citations.
To answer RQ6, we relied on the answers to questions 11 that we collected from the analyzed review abstracts (when applicable). Recall from above that question 11 focused on identifying the various keywords that the systematic reviewers often use to structure their review abstracts. This allowed us to generate a word cloud depicting the most frequent keywords used to structure review abstracts.

4.4. Results Analysis

(1)
(RQ1): Investigation of the systematicity of the review abstracts
Table 3 shows the distribution of the confidence values that each review abstract yielded. All the confidence values, therefore, vary between 5 and 80 (80 excluded). Sixty-two of the review abstracts yield a confidence varying between 40 and 60 (60 excluded): these, therefore, have a Medium level confidence. Fifty-nine of the review abstracts yield a confidence varying between 20 and 40 (40 excluded): these therefore have a Low-level confidence. Only seven review abstracts have a confidence varying between 60 and 80 (80 excluded): they therefore have a Medium High confidence. None of the review abstracts has a confidence that is at least 80% (i.e., High). This allows us to conclude that all the analyzed review abstracts exhibit a suboptimal systematicity and hence, a suboptimal adherence to Check-SE 2022.
Table 3. Confidence of the Review Abstracts.
(2)
(RQ2): Influence of the Publication Year on the Systematicity of the Reviews’ Abstracts
Figure 1 shows the variations of the OConfidence values that the analyzed review abstracts yield. It, therefore, indicates that the confidence of these reviews varies from one year to another, with highest variations scored in years 2016 and 2017.
Figure 1. Variation of the OConfidence throughout the years.
(3)
(RQ3): Influence of the Well-Established Prisma Reporting Guideline on the Systematicity of the Review Abstracts
Figure 2 shows the variations of the OConfidence values that the analyzed PRISMA-driven review abstracts yield. Based on Figure 2, we can conclude that the PRISMA-driven review abstracts, therefore, yield relatively higher OConfidence values than all the analyzed review abstracts (see Figure 1). Based on the timeline that Figure 2 reports, we can also conclude that the software engineering community has recently started to use the well-established PRISMA reporting guideline, the least recent analyzed PRISMA-driven review dating back 2019.
Figure 2. Variation of the OConfidence throughout the years for the PRISMA-driven reviews.
(4)
(RQ4): Minimal set of items that review abstracts usually report
Figure 3 shows a summary of the answers of the eleven questions that we extracted from the analyzed review abstracts. That Figure allows us to conclude that questions 1, 2, 7, and 10 are those that most of the analyzed abstracts have appropriately answered. Hence, when writing the review abstracts, most of systematic reviewers ensured that:
  • The abstract identifies the review as systematic (question 1)
  • The abstract stated the main objective(s) or question(s) that the review addressed (question 2)
  • The abstract specified the total number of included studies and participants and summarize relevant characteristics of studies (question 7)
  • The abstract provided a general interpretation of the results as well as important implications (question 10).
Figure 3. Summary of the items answered by reviews.
As stated in Section III.B, Kitchenham et al. [26] recommended reporting at least items 2, 4, 6, 8, and 10 of the PRISMA 2020 for abstracts checklist as they are the most critical. But Figure 3 shows that existing reviews do not comply with that requirement, as they mostly focused on answering questions 1, 2, 7, and 10 that correspond to items 1, 2, 7, and 10 in the PRISMA 2020 for abstracts checklist.
(5)
(RQ5): Influence of the systematicity of review’ abstracts on the number of citations of these reviews
We computed the correlation between the OConfidence values of the analyzed review abstracts and the number of Google citations (as of mid-July 2022). The corresponding value of the Spearman coefficient (https://www.socscistatistics.com/tests/spearman/default2.aspx (accessed on 17 July 2022)) rs is 0.01235, with a p-value of 0.97685. By normal standards, the association between the two variables would not be considered statistically significant. We therefore cannot conclude whether there is a relationship between the systematicity of the review abstracts and the number of times these reviews get cited. Hence, there is no evidence that the systematicity of reviews has an impact on the influence, e.g., number of citations, of the review on the research community. Other factors may come into play and are worth investigating as future work. Such factors may include the reputation of the systematic reviewers in the research community, the importance, e.g., impact factor, conference ranking, conference acceptance rate, of the venue in which the review is published.
(6)
(RQ6): On the well-formedness of the review abstracts
Thirty of the analyzed reviews had structured their abstracts by using specific keywords that mirrored the structure of the reviews they summarize. Based on the answers to question 11 that we collected from these thirty reviews, we identified the various keywords that the systematic reviewers usually used to structure their abstracts. This allowed us to generate the word cloud shown in Figure 4, which was generated using the online tool called Free Cloud Generator (https://www.freewordcloudgenerator.com/ (accessed on 19 July 2022)). Based on that figure, we can conclude that the most common structure of abstracts comprised the following sections: Context, Objectives, Method, Results, and Conclusion(s). Some of abstracts also include an additional Background section to improve the well-formedness of review abstracts. These various sections are expected to comprise sufficient information to allow readers to conclude that the review is transparent, reproducible, trustworthy, unbiased, and complete enough.
Figure 4. Word cloud summarizing the structure of abstracts.

5. Discussion

Our experiments allow us to conclude that all the analyzed review abstracts exhibit a sub-optimal systematicity. This is probably because most of the reporting procedures followed by non-PRISMA driven reviews did not necessarily put the emphasis on the reporting quality that review abstracts should exhibit. To increase that systematicity and encourage systematic reviewers to report higher quality abstracts, existing reporting procedures should be revised to encourage reviewers to create more informative and structured abstracts that properly mirror the content of the reviews. That systematicity could also be improved by using well established reporting guidelines such as PRISMA that have been specified by a panel of experts. The reviews included in our dataset indicate that PRISMA has just recently started to be used by systematic reviewers of the software engineering community. Furthermore, our experiments indicate that using PRISMA may yield a relatively higher systematicity when it comes to review abstracts among others.
Besides, providing more examples illustrating how to write good reviews and more particularly good review abstracts in computing-related disciplines, e.g., in artificial intelligence, internet of things, software engineering, could lead to the reporting of higher quality reviews. This is consistent with Kitchenham et al. [26] who proposed some examples illustrating the use of the PRISMA reporting guidelines to write good reviews in the software engineering field and have notably put the emphasis on how to properly write abstracts. This is also consistent with Okoli [30] who provided exemplars of high-quality literature reviews in the Information Systems field.

6. Conclusions

To assess the extent to which a review abstract exhibits good reporting quality, we have introduced the notion of systematicity. The latter embodies a review’s reporting quality. We have then proposed the Check-SE 2022 checklist for abstracts, an eleven-item question-based checklist for abstracts that is essentially derived from the PRISMA 2020 for abstracts checklist and Kitchenham et al. [26]. The Check-SE 2022 checklist for abstracts embodies the best reporting guidelines and/or procedures that the abstracts of reviews published in the SE (software engineering) field should comply with to exhibit good reporting quality. We have also proposed a measure that allows assessing compliance with that checklist. That quantitative measure allows assessing the systematicity of review abstracts. Our experiments on several reviews published in the software engineering field and self-identifying as systematic reviews have shown that the abstracts of these reviews exhibited a suboptimal systematicity.
As future work, we are planning to expand our approach to assess the systematicity of the whole content of each review included in our dataset. This will help us assess the extent to which existing reviews published in the software engineering field have complied with well-established reporting guidelines such as PRISMA 2020. To support that assessment, we will propose additional questions that allow checking the adherence to the target reporting guidelines. Similar to Boaye Belle et al. [44], in future work, we will also implement an NLP (Natural language Processing)-based extraction tool that will automatically analyze systematic reviews to find answers to the proposed questions.
As future work, we will also focus on the assessment of the reviews published in the software engineering field. In this regard, we will compare the systematicity of reviews self-identifying as systematic reviews to the systematicity of reviews that do not self-identify as systematic. This will help us assess the extent to which the use of reporting guidelines/procedures to write systematic reviews increases the quality of reviews published in the software engineering field.
As we have stated above, the notion of systematicity is usually akin to several sub-properties, such as the completeness of the reporting, the transparency of the reporting, the trustworthiness in the review findings, the appropriateness of the methods used to conduct the review, the unbiasedness of the reporting, and the reproducibility of the study [2,6,8,14,29,39]. In future work, we will carry out a systematic review (systematic mapping study) focusing on the systematicity property to further explore these sub-properties. The resulting taxonomy will help us refine our checklist-based approach to create questions that capture the various systematicity sub-properties. This may result in a more advanced confidence measure that quantifies more accurately the systematicity of reviews.

Author Contributions

A.B.B., conceptualization, investigation, methodology, supervision, software, validation, visualization and writing-original draft; Y.Z., investigation, visualization, validation and writing-original draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the LURA (Lassonde undergraduate research Award) program. The grant number is 1. The APC (Article Processing Charges) were funded by the start-up grant of the lead author.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Appendix A. The Two PRISMA 2020 Checklists

Table A1 and Table A2 respectively report the two PRISMA 2020 checklists.
Table A1. PRISMA 2020 12-item checklist for abstracts adapted from [2,4].
Table A1. PRISMA 2020 12-item checklist for abstracts adapted from [2,4].
Abstract Section TopicItem NumberAbstract Checklist Item
TitleTitle1In the abstract, the systematic reviewers should identify the report as a systematic review
BackgroundObjectives2In the abstract, systematic reviewers should explicitly state the objective(s)/question(s) the review addresses.
MethodsEligibility criteria5In the abstract, systematic reviewers should mention the inclusion and exclusion criteria they used in the review
Sources of information4In the abstract, systematic reviewers should indicate all the sources (e.g., databases, registers) they searched or consulted to identify studies and the most recent date of the search/consultation of these sources.
Risk of bias5In the abstract, systematic reviewers should report the different methods they used when assessing risk of bias in the studies they included in the review
Synthesis of results6In the abstract, systematic reviewers should describe the methods they used to report and synthesize their findings
ResultsIncluded studies7In the abstract, systematic reviewers should indicate the total number of studies they included in the review, the participants and provide a summary of the relevant studies’ characteristics
Synthesis of results8In the abstract, systematic reviewers should specify the results for main outcomes, ideally specifying the number of included studies and participants for each.
DiscussionLimitations of evidence9In the abstract, systematic reviewers should make a brief summary of the limitations (e.g., study risk of bias, inconsistency) of the evidence they included in the review
Interpretation10In the abstract, systematic reviewers should make an interpretation of their findings as well as significant implications
OtherFunding11In the abstract, systematic reviewers should specify the main sources of financial support they received to complete the review
Registration12In the abstract, systematic reviewers should specify the review’s register name and registration number
Table A2. An excerpt of the PRISMA 2020 27-item checklist adapted from [2,4].
Table A2. An excerpt of the PRISMA 2020 27-item checklist adapted from [2,4].
Review Section TopicItem NumberChecklist Item
TitleTitle1Systematic reviewers should identify the report as a systematic review
AbstractAbstract2This item is further described through the PRISMA 2020 for Abstracts 12-tems checklist (see Table 2).
IntroductionRationale3Systematic reviewers should state the rationale for the review in the context of existing knowledge
Objectives4Systematic reviewers should explicitly state the objective(s)/question(s) the review addresses.
MethodsEligibility criteria5Systematic reviewers should mention the inclusion and exclusion criteria they used in the review and explain how they grouped studies for the syntheses.
Sources of information6Systematic reviewers should indicate all the sources (e.g., databases, registers, websites, organizations, reference lists) they searched or consulted to identify studies and the most recent date of the search/consultation of these sources.
Strategy used for the search7Systematic reviewers should report the full search strategies used for all databases, registers, and websites, and specify the different filters and limits used during the search.
ResultsStudy selection16aSystematic reviewers should report the results of the search and selection process by preferably relying on a flow diagram (The flow diagram is a template that illustrates the flow of records retrieved and analyzed in the different stages of the search and selection process (Page et al. [4]))
16bSystematic reviewers should specify the studies seem to meet the inclusion criteria, but that they decided to exclude, and provide a rationale for their inclusion
Characteristics of the study17Systematic reviewers should specify each included study and describe its characteristics.
Risk of bias in studies18Systematic reviewers should report assessments of risk of bias for each study they included
Discussion
Other informationProtocol and registration24aSystematic reviewers should specify the review’s registration information (e.g., register name and registration number) or state that they did not register the review
24bSystematic reviewers should specify the location of the review protocol, or state that they did not prepare a protocol
24cSystematic reviewers should indicate any amendments they made to the information they provided at registration/in the protocol.
Support25Systematic reviewers should specify sources of financial or non-financial they received to support the completion of the review, and specify role of the funders/sponsors in the completion of the review
Competing interests26Systematic reviewers should state any relevant competing interests regarding the review
Data, code, and other materials availability27Systematic reviewers should specify whether some additional information (e.g., template data collection forms, data extracted from included studies, data used for all analyses, analytic code) is publicly available and specify its location

References

  1. Budgen, D.; Brereton, P.; Drummond, S.; Williams, N. Reporting systematic reviews: Some lessons from a tertiary study. Inf. Softw. Technol. 2018, 95, 62–74. [Google Scholar] [CrossRef]
  2. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  3. Evans, D. Hierarchy of evidence: A framework for ranking evidence evaluating healthcare interventions. J. Clin. Nurs. 2003, 12, 77–84. [Google Scholar] [CrossRef]
  4. Page, M.J.; Moher, D.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. PRISMA 2020 explanation and elaboration: Updated guidance and exemplars for reporting systematic reviews. BMJ 2021, 372, n160. [Google Scholar] [CrossRef]
  5. Sargeant, J.M.; Reynolds, K.; Winder, C.B.; O’Connor, A.M. Completeness of reporting of systematic reviews in the animal health literature: A meta-research study. Prev. Vet. Med. 2021, 195, 105472. [Google Scholar] [CrossRef] [PubMed]
  6. Belur, J.; Tompson, L.; Thornton, A.; Simon, M. Interrater reliability in systematic review methodology: Exploring variation in coder decision-making. Sociol. Methods Res. 2021, 50, 837–865. [Google Scholar] [CrossRef]
  7. Abdul Qadir, A.; Mujadidi, Z.; Boaye Belle, A. Incidium competition 2022: A preliminary systematic review centered on guidelines for reporting systematic reviews. Appear. STEM Fellowsh. J. 2022; in press. [Google Scholar]
  8. Logullo, P.; MacCarthy, A.; Kirtley, S.; Collins, G.S. Reporting guideline checklists are not quality evaluation forms: They are guidance for writing. Health Sci. Rep. 2020, 3, e165. [Google Scholar] [CrossRef]
  9. Moher, D.; Shamseer, L.; Clarke, M.; Ghersi, D.; Liberati, A.; Petticrew, M.; Shekelle, P.; Stewart, L.A. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst. Rev. 2015, 4, 1. [Google Scholar] [CrossRef]
  10. Thabane, L.; Samaan, Z.; Mbuagbaw, L.; Kosa, S.; Debono, V.B.; Dillenburg, R.; Zhang, S.; Fruci, V.; Dennis, B.; Bawor, M. A systematic scoping review of adherence to reporting guidelines in health care literature. J. Multidiscip. Healthc. 2013, 6, 169. [Google Scholar] [CrossRef]
  11. Page, M.J.; E McKenzie, J.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Moher, D. Updating guidance for reporting systematic reviews: Development of the PRISMA 2020 statement. J. Clin. Epidemiol. 2021, 134, 103–112. [Google Scholar] [CrossRef] [PubMed]
  12. Pussegoda, K.; Turner, L.; Garritty, C.; Mayhew, A.; Skidmore, B.; Stevens, A.; Boutron, I.; Sarkis-Onofre, R.; Bjerre, L.M.; Hróbjartsson, A.; et al. Systematic review adherence to methodological or reporting quality. Syst. Rev. 2017, 6, 131. [Google Scholar] [CrossRef] [PubMed]
  13. Oliveira, N.L.; Botton, C.E.; De Nardi, A.T.; Umpierre, D. Methodological quality and reporting standards in systematic reviews with meta-analysis of physical activity studies: A report from the Strengthening the Evidence in Exercise Sciences Initiative (SEES Initiative). Syst. Rev. 2021, 10, 304. [Google Scholar] [CrossRef]
  14. Bougioukas, K.I.; Liakos, A.; Tsapas, A.; Ntzani, E.; Haidich, A.-B. Preferred reporting items for overviews of systematic reviews including harms checklist: A pilot tool to be used for balanced reporting of benefits and harms. J. Clin. Epidemiol. 2018, 93, 9–24. [Google Scholar] [CrossRef] [PubMed]
  15. Fernández-Jané, C.; Solà-Madurell, M.; Yu, M.; Liang, C.; Fei, Y.; Sitjà-Rabert, M.; Úrrutia, G. Completeness of reporting acupuncture interventions for chronic obstructive pulmonary disease: Review of adherence to the STRICTA statement. F1000Research 2020, 9, 226. [Google Scholar] [CrossRef]
  16. Wiehn, E.; Ricci, C.; Alvarez-Perea, A.; Perkin, M.R.; Jones, C.J.; Akdis, C.; Bousquet, J.; Eigenmann, P.; Grattan, C.; Apfelbacher, C.J.; et al. Adherence to the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) checklist in articles published in EAACI Journals: A bibliographic study. Allergy 2021, 76, 3581–3588. [Google Scholar] [CrossRef] [PubMed]
  17. Ferdinansyah, A.; Purwandari, B. Challenges in combining agile development and CMMI: A systematic literature review. In Proceedings of the 2021 10th International Conference on Software and Computer Applications, Kuala Lumpur, Malaysia, 23–26 February 2021; pp. 63–69. [Google Scholar]
  18. Whiting, P.; Savović, J.; Higgins, J.P.T.; Caldwell, D.M.; Reeves, B.C.; Shea, B.; Davies, P.; Kleijnen, J.; Churchill, R. ROBIS: A new tool to assess risk of bias in systematic reviews was developed. J. Clin. Epidemiol. 2016, 69, 225–234. [Google Scholar] [CrossRef]
  19. Shea, B.J.; Grimshaw, J.M.; Wells, A.G.; Boers, M.; Andersson, N.; Hamel, C.; Porter, A.C.; Tugwell, P.; Moher, D.; Bouter, L.M. Development of AMSTAR: A measurement tool to assess the methodological quality of systematic reviews. BMC Med. Res. Methodol. 2007, 7, 10. [Google Scholar] [CrossRef]
  20. Shea, B.J.; Reeves, B.C.; Wells, G.; Thuku, M.; Hamel, C.; Moran, J.; Moher, D.; Tugwell, P.; Welch, V.; Kristjansson, E.; et al. AMSTAR 2: A critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ 2017, 358, j4008. [Google Scholar] [CrossRef]
  21. Jin, Y.; Sanger, N.; Shams, I.; Luo, C.; Shahid, H.; Li, G.; Bhatt, M.; Zielinski, L.; Bantoto, B.; Wang, M.; et al. Does the medical literature remain inadequately described despite having reporting guidelines for 21 years?–A systematic review of reviews: An update. J. Multidiscip. Healthc. 2018, 11, 495. [Google Scholar] [CrossRef]
  22. Blanco, D.; Altman, D.; Moher, D.; Boutron, I.; Kirkham, J.J.; Cobo, E. Scoping review on interventions to improve adherence to reporting guidelines in health research. BMJ Open 2019, 9, e026589. [Google Scholar] [CrossRef] [Green Version]
  23. Heus, P.; Damen, J.A.A.G.; Pajouheshnia, R.; Scholten, R.J.P.M.; Reitsma, J.B.; Collins, G.; Altman, D.G.; Moons, K.G.M.; Hooft, L. Uniformity in measuring adherence to reporting guidelines: The example of TRIPOD for assessing completeness of reporting of prediction model studies. BMJ Open 2019, 9, e025611. [Google Scholar] [CrossRef]
  24. Walsh, S.; Jones, M.; Bressington, D.; McKenna, L.; Brown, E.; Terhaag, S.; Shrestha, M.; Al-Ghareeb, A.; Gray, R. Adherence to COREQ reporting guidelines for qualitative research: A scientometric study in nursing social science. Int. J. Qual. Methods 2020, 19, 1609406920982145. [Google Scholar] [CrossRef]
  25. Andrade, C. How to write a good abstract for a scientific paper or conference presentation. Indian J. Psychiatry 2011, 53, 172. [Google Scholar] [CrossRef]
  26. Kitchenham, B.A.; Madeyski, L.; Budgen, D. Segress: Software Engineering Guidelines for REporting Secondary Studies. IEEE Trans. Softw. Eng. 2022. [Google Scholar] [CrossRef]
  27. Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Ann. Intern. Med. 2009, 151, 264–269. [Google Scholar] [CrossRef]
  28. Nair, S.; de la Vara, J.L.; Sabetzadeh, M.; Briand, L. Classification, structuring, and assessment of evidence for safety--a systematic literature review. In Proceedings of the 2013 IEEE Sixth International Conference on Software Testing, Verification and Validation, Luxembourg, 18–22 March 2013; pp. 94–103. [Google Scholar]
  29. Moher, D.; Schulz, K.F.; Simera, I.; Altman, D.G. Guidance for developers of health research reporting guidelines. PLoS Med. 2010, 7, e1000217. [Google Scholar] [CrossRef]
  30. Okoli, C. A guide to conducting a standalone systematic literature review. Commun. Assoc. Inf. Syst. 2015, 37, 43. [Google Scholar] [CrossRef]
  31. Veroniki, A.A.; Tsokani, S.; Zevgiti, S.; Pagkalidou, I.; Kontouli, K.; Ambarcioglu, P.; Pandis, N.; Lunny, C.; Nikolakopoulou, A.; Papakonstantinou, T. Do reporting guidelines have an impact? Empirical assessment of changes in reporting before and after the PRISMA extension statement for network meta-analysis. Syst. Rev. 2021, 10, 246. [Google Scholar] [CrossRef]
  32. Kitchenham, B. Procedures for performing systematic reviews. Keele UK Keele Univ. 2004, 33, 1–26. [Google Scholar]
  33. Petticrew, M.; Roberts, H. Systematic Reviews in the Social Sciences: A practical Guide; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  34. Kitchenham, B.; Pretorius, R.; Budgen, D.; Brereton, O.P.; Turner, M.; Niazi, M.; Linkman, S. Systematic literature reviews in software engineering—A tertiary study. Inf. Softw. Technol. 2010, 52, 792–805. [Google Scholar] [CrossRef]
  35. Pickering, C.; Byrne, J. The benefits of publishing systematic quantitative literature reviews for PhD candidates and other early-career researchers. High. Educ. Res. Dev. 2014, 33, 534–548. [Google Scholar] [CrossRef] [Green Version]
  36. Wohlin, C. Guidelines for snowballing in systematic literature studies and a replication in software engineering. In Proceedings of the 18th International Conference on Evaluation and Assessment in Software Engineering, London, UK, 13–14 May 2014; pp. 1–10. [Google Scholar]
  37. Yu, Y.; Shi, Q.; Zheng, P.; Gao, L.; Li, H.; Tao, P.; Gu, B.; Wang, D.; Chen, H. Assessment of the quality of systematic reviews on COVID-19: A comparative study of previous coronavirus outbreaks. J. Med. Virol. 2020, 92, 883–890. [Google Scholar] [CrossRef]
  38. Gates, M.; Gates, A.; Duarte, G.; Cary, M.; Becker, M.; Prediger, B.; Vandermeer, B.; Fernandes, R.; Pieper, D.; Hartling, L. Quality and risk of bias appraisals of systematic reviews are inconsistent across reviewers and centers. J. Clin. Epidemiol. 2020, 125, 9–15. [Google Scholar] [CrossRef] [PubMed]
  39. Rethlefsen, M.L.; Kirtley, S.; Waffenschmidt, S.; Ayala, A.P.; Moher, D.; Page, M.J.; Koffel, J.B. PRISMA-S: An extension to the PRISMA statement for reporting literature searches in systematic reviews. Syst. Rev. 2021, 10, 39. [Google Scholar] [CrossRef]
  40. Weaver, R.; Fenn, J.; Kelly, T. A pragmatic approach to reasoning about the assurance of safety arguments. In Proceedings of the 8th Australian Workshop on Safety Critical Systems and Software, Canberra, Australia, 9–10 October 2003; Volume 33, pp. 57–67. [Google Scholar]
  41. Boaye Belle, A.; Lethbridge, T.C.; Kpodjedo, S.; Adesina, O.O.; Garzón, M.A. A novel approach to measure confidence and uncertainty in assurance cases. In Proceedings of the 2019 IEEE 27th International Requirements Engineering Conference Workshops (REW), Jeju, Korea, 23–27 September 2019; pp. 24–33. [Google Scholar]
  42. Kitchenham, B.; Charters, S. Guidelines for Performing Systematic Literature Reviews in Software Engineering, Version 2.3; Technical Report EBSE 2007-01; Keele University Newcastle, UK; Durham University: Durham, UK, July 2007. [Google Scholar]
  43. Petersen, K.; Feldt, R.; Mujtaba, S.; Mattsson, M. Systematic mapping studies in software engineering. In Proceedings of the 12th International Conference on Evaluation and Assessment in Software Engineering (EASE), Bari, Italy, 12 June 2008; pp. 1–10. [Google Scholar]
  44. Boaye Belle, A.; El Boussaidi, G.; Kpodjedo, S. Combining lexical and structural information to reconstruct software layers. Inf. Softw. Technol. 2016, 74, 1–16. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.