Keywords

1 Introduction and Background

1.1 Systematic Reviews in Healthcare

A systematic review is a summary of primary research on a specific topic to collect information and reduce bias. It “typically involve[s] a detailed and comprehensive plan and search strategy”, “reducing bias by identifying, appraising, and synthesizing all relevant studies” [1]. With ever-increasing advancements in medical technology, practices, and knowledge, medical practitioners need to be well informed of the most recent evidence. Systematic reviews offer medical practitioners a collection of primary studies to assist them with decision-making. Not only are systematic reviews useful for obtaining up-to-date information, but they are also used when developing new clinical practice guidelines [2]. Systematic reviews are relied upon for the prevention of errors, patient satisfaction and safety, improving the quality of care, and saving lives. Deaths from healthcare errors are estimated to be over 250,000 a year, with many likely unreported. Based on these numbers, healthcare errors are the 3rd leading cause of death in the United States [3]. Challenges arise for busy healthcare clinicians to fully access information of similar and relatable trials which may be applicable to quality and safety improvements specific to their areas of practice. Reviews are rapidly added, making it difficult to keep up to date [4]. Many of these reviews may not be fully updated or amended when new data arises. As healthcare advancements rapidly change, healthcare clinicians have less time to research, evaluate, educate, and adopt new evidence-based practices. Inadequate reporting can create poor interpretation and critical analysis of systematic reviews. While they have many drawbacks, they are still more advanced in comparison to traditional “opinion reviews that cite evidence selectively to support a viewpoint” [5]. Finally, concerns with validity exist within these trials as limitations and bias, whether transparent or unrecognized, along with sample populations and targeted responses to questions may lack relevance to current clinical practice. Systematic reviews require extensive evaluation and time to increase the likelihood of being applicable to current practice [5].

1.2 Systematic Review Automation

With evidence-based publications arising at rapid rates, healthcare guidelines are continuously requiring revisions to be up to date. Significant delays in literature review for application in clinical practice make achieving improved patient outcomes difficult to achieve. A proposed solution is developing and adopting automation tools to search, gather, and more effectively and efficiently group and analyze publications [4]. Automation can serve by increasing the speed of data acquisition and generate valid, up-to-date recommendations for best practices. Using artificial intelligence within automation can serve to solve complex problems and accelerate advancements in health-related practice and treatments [6]. In addition, automation may reduce the risk of bias in the grouping of publications or omission of other relevant publications that do not serve the interests of the reviewer. While automation may lead to greater efficiency, concerns arise with “the extent to which an innovation is in line with current values and practices” [4]. Other challenges with automation include the need for high relatability and transparency which requires critical thinking and experience [7].

2 Purpose of Study

The purpose of this study is to conduct a literature review on how automation is being integrated into the healthcare systematic review process. Automated data collection has been gaining attention from both large and small businesses in the past decade. While it has made progress in the healthcare sector, there is still hesitation to fully implement an automated systematic review for medical practices. This study will look at how automating healthcare systematic reviews are perceived by the medical community and understand why there is hesitation in fully automating this process within the medical field. This study uses programs such as Web of Science, Harzing’s Publish or Perish, VOSviewer, CiteSpace, and MAXQDA. The program Zotero was used to create the citations and reference list [8].

3 Methodology

3.1 Data Collection

Data was initially collected from doing a keyword search within the Web of Science database. This search was later refined with a larger data pool using the program Harzing’s Publish or Perish to conduct a keyword search in both the Web of Science and Google Scholar databases. Both databases were used since each has its advantages and disadvantages. Web of Science offers data on the title, authors, source, abstract, and cited references, but only allows 500 articles per data set. Google Scholar does not include data on cited references but offers up to 1000 articles in a data set from a larger amount of articles. Since cited references are required for co-citation analysis, the data from Google Scholar cannot be used. However, the larger data set from Google Scholar is preferable for cluster analysis. The keywords used in the data collection were “automating systematic reviews” and “healthcare systematic reviews”.

3.2 Trend Analysis

The trend analysis was conducted using data from the Web of Science along with analysis tools provided within the program. All data was collected between the years 1980 and 2021.

Figure 1 shows the number of articles published per year that were found using a keyword search of “automating systematic reviews”. This graph shows that interest in this topic first began in 1993, but gained interest in the 2010s. The year 2021 shows a significant decrease in article publication due to the year not being completed. Based on the increasing trend of prior years, it can be predicted that the year 2021 will have an even greater amount of published articles than the years prior.

Fig. 1.
figure 1

Trend analysis of data from the articles on automating systematic reviews [9]

Figure 2 is created with the same data as Fig. 1, but it shows how many articles are written in each field. This graph shows that most articles that have the keyword “automating systematic review” are from the computer science information systems field. There is a significant decrease in articles of over 20 articles between the computer science information systems field and the remaining fields. This graph also shows the field of health care sciences services is ranked 7th for this keyword search with 24 articles. The information from Fig. 2 can suggest that most of the development of automating systematic reviews are occurring within the computer science information systems field and from there, being applied to other fields.

Fig. 2.
figure 2

Trend analysis of data from the articles on automating systematic reviews organized by field [9]

Figure 3 is created with the same data as Fig. 1 and Fig. 2 but has refined the keyword search to only include articles within the medical field. This decreased the articles in the data set from 416 articles to 35 articles. This graph shows the year 2019 was the peak of when articles with the keyword “automating systematic review” were published. This contradicts the trend of increasing published articles in Fig. 1. The lower amount of published articles in 2020 could have been due to resources being put towards the COVID-19 pandemic.

Fig. 3.
figure 3

Trend analysis of data from the articles on automation of systematic reviews in the medical field [9]

Figure 4 is a graph of the number of articles published per year that were found using a keyword search of “healthcare systematic reviews”. This graph shows a trend of an increase in published articles. The year 2020 had the largest amount of published articles. This lowers the likeliness of the explanation for the decrease in articles in the year 2020 for Fig. 3 to be true.

Fig. 4.
figure 4

Trend analysis of data from the articles on healthcare systematic reviews [9]

Figure 5 is a pivot table showing the results of an analysis of the sum of articles per author. Figure 5 was created from data collected using a keyword search of “automating systematic reviews in healthcare” conducted with Harzing’s Publish or Perish from the Google Scholar database. A pivot table is a tool that can convert 65,000 rows of data into a summarized table [10]. The author with the largest number of articles is James Thomas, with nine articles. Most of the articles he has published are on the topic of automation and computer assistance. While not directly in the medical field, his articles show interesting perspectives on automation. One such article is “On the use of computer-assistance to facilitate systematic mapping” [11]. The author with the second largest amount of published articles is Paul Glasziou who began working on the topic of automating systematic reviews when the topic gained popularity. One of his earlier publications is “The automation of systematic reviews” [12].

Fig. 5.
figure 5

Pivot table in Excel of the number of articles per author [13, 14]

4 Results

4.1 Co-citation Analysis

The co-citation analysis was conducted with the program VOSviewer with data extracted from the article search in the Web of Science. Specifically, the data used to make Fig. 3 was used to create Fig. 5. The exported data included the title, authors, source, abstract, and cited references. Co-citation analyses are used as a way to adequately identify the relationships and structure between research areas [15]. The basis of the co-citation analysis is the higher amount of co-citations between authors there is, the stronger the relationship of the authors is [16]. The VOSviewer creates a cluster map with the imported data that shows the articles that we cited the most. The links between the authors show which articles have been cited together. Limits can be set for the minimum number of citations for each article. Another program that can perform a co-citation analysis is CiteSpace. This program uses the same data as the VOSviewer, but a unique output it has is a citation burst. A citation burst is a way that makes “it possible to find the articles that receive particular attention from the related scientific communities in a certain period of time” [17]. Not only is it able to perform co-citation analyses, but it is also able to assist in “identifying betweenness centrality between pivotal points in scientific articles, which indicates the significance of the nodes in a network” [17]. CiteSpace is able to identify co-cited articles by choosing different nodes [17].

VOSviewer Analysis.

Figure 6 is a co-citation cluster map of the data from the keyword search for “automating systematic reviews” within the medical field. The minimum number of co-citations was set to 10. Figure 6 contains 2 clusters, 44 links, and has a total link strength of 244. The article with the largest citations was “Using text mining for study identification in systematic reviews: a systematic review of current approaches” [18] with 22 citations. The article with the second largest citations was “Systematic review automation technologies” [19] with 21 citations. The earliest article within this cluster is from 2009, with most of the other articles from the mid to late 2010’s. This appears to follow the article publication trend from Fig. 4.

Fig. 6.
figure 6

Co-citation analysis for automating systematic reviews in the medical field [20]

CiteSpace Analysis.

Figure 7 is a cluster burst created with CiteSpace using the data from the keyword search “healthcare systematic reviews”. There are 5 bursts within Fig. 7 that show the main keyword per cluster.

Figure 8 is a list of references from the data in Fig. 7 with the strongest citation burst. The top reference is “Grading Quality of Evidence and Strength of Recommendations: A Perspective” [21].

Fig. 7.
figure 7

Citation burst from CiteSpace on data for healthcare systematic reviews [22]

Fig. 8.
figure 8

References from Fig. 8 with the strongest citation bursts [22]

Figure 9 is a cluster burst created with CiteSpace using the data from the keyword search for “automating systematic reviews” within the medical field.

Figure 10 is a list of references from the data in Fig. 9 with the strongest citation burst. The top reference happens to be the same as Fig. 8, “Grading Quality of Evidence and Strength of Recommendations: A Perspective” [21]. The second top reference is “Seventy-Five Trials and Eleven Systematic Reviews a Day: How Will We Ever Keep Up?” [23]. Figure 8 and Fig. 10 show that while there has been a significant increase in articles written for automating systematic reviews for healthcare, the most referenced article is from 2009. Technology has increased significantly since this paper was published; however, Fig. 8 and Fig. 10 suggest that there is still a lack of research and integration on this topic within the medical field.

Fig. 9.
figure 9

Citation burst from CiteSpace of data for automating systematic reviews in the medical field [22]

Fig. 10.
figure 10

References from Fig. 8 with the strongest citation bursts [22]

4.2 Co-authorship Analysis

A co-authorship analysis is used as “a proxy of research collaboration” “that links different sets of talent to produce a research output” [24]. It is a useful analysis to understand where articles are coming from and what social networks have an interest in specific research topics. Co-authorship also allows the “relations between these groups and changes in time” between these groups to be analyzed [25].

Figure 11 is a cluster map of a co-authorship analysis for the data from the keyword search for “automating systematic reviews” within the medical field. Figure 11 contains 4 clusters, 41 links, and has a total link strength of 84. The authors with the most connections are James Thomas, Paul Glasziou, and Guy Tsafnat. These results are confirmed with the trend analysis in Fig. 5 since these authors have the most published articles so they will most likely have the largest co-authorship within this keyword search. Together Fig. 4, Fig. 5, Fig. 11 suggest that more authors have taken interest and begun to publish papers.

Fig. 11.
figure 11

Co-authorship analysis on data for automating systematic reviews in the medical field [20]

4.3 Content Analysis

Comparable with co-citation analysis, content analysis is conducted with the program VOSviewer. However, it uses data collected from Google Scholar using the program Harzing’s Public or Perish [13]. Since content analysis does not require cited references in the data, data from Google Scholar can be used. The advantage to using data from Google Scholar is the ability to export a larger data pool compared to Web of Science. Content analyses are used to organize and interpret the data from imported articles to gain reasonable results and understanding from the set of articles [26].

Figure 12 is a content analysis cluster map produced from the data collected from the keyword search for “systematic review in healthcare”. The data set was collected from Google Scholar and contains 980 articles. The minimum occurrence was set to 20 which resulted in 51 keywords but only the 60% most relevant keywords which is 31. Figure 12 contains 5 clusters, 287 links, and has a link strength of 1918. The keywords with the largest connection are “systematic review” with 692 occurrences, ‘healthcare” with 177 occurrences, and “systematic literature review” with 121 occurrences.

Fig. 12.
figure 12

Cluster analysis on data for automating systematic review in healthcare [20]

Figure 13 is a content analysis cluster map produced from the data collected from the keyword search for “automating systematic review” in the medical field. This data set was exported from the Web of Science and contains 35 articles. Figure 9 has 3 clusters, 78 links, and a total link strength of 1112. The minimum number of articles was set to 20 occurrences and 16 keywords met this condition. The top keywords from this content analysis cluster are systematic review, automation, and study. These keywords had occurrences of 84, 83, and 72, respectively. Both Fig. 12 and Fig. 13 show that the articles within the data sets used focus on automation and systematic reviews.

Fig. 13.
figure 13

Cluster analysis on data for automating systematic reviews in the medical field [20]

4.4 Content Analysis from MAXQDA

MAXQDA performs a similar content analysis as VOSviewer, however, MAXQDA uses a smaller set of data and is able to examine this data in greater detail. MAXQDA also allows the user to manually remove stop words from being analyzed within the data. MAXQDA is also able to visualize results into a word cloud, which can be an effective way to communicate the most common words within articles [27].

Figure 14 is a word cloud created from a data set collected from Web of Science using a keyword search of “healthcare automation”. This word cloud shows that the top three keywords found in the articles in the data set are “systematic”, “automation”, and “review”. Figure 14 shows that recent articles having to do with “healthcare automation” are beginning to focus on systematic reviews.

Fig. 14.
figure 14

Word Cloud from MAXQDA [28]

5 Conclusion

A systematic review of literature is critical for the advancement of healthcare practice, safety, and improved patient outcomes. The healthcare sector has been more hesitant to adopt automation and often continues to rely on traditional methods. Challenges exist in both traditional and advanced technology methods. Shortcomings with traditional methods include significant delays in the collection of relevant and relatable data, a potential bias that may result in omitting or overuse of selected articles and keeping up to date with the rapid additions of research literature. Automation technology is quickly advancing but currently lacks the ability to apply critical thinking of how literature is relevant to current practice, more easily understood by clinician experience, attitudes, and values. Developing and utilizing further advancements in technology, including artificial intelligence may lead to maximizing the full potential of systematic literature reviews. Understanding and addressing concerns of healthcare clinicians such as ensuring technology is well-matched with current practice and values, is vital for acceptance and trust of automation.

6 Future Work

Advancements in artificial intelligence, natural language processing, and machine learning have created more advanced algorithms that can enhance the efficiency and quality of automated systematic reviews. This complex ecosystem can provide interconnections of numerous techniques in artificial intelligence to solve complex problems within the automation process. This may lead to faster activity, more accurate, and less expensive results. Machine learning is a promising type of artificial intelligence that may accelerate quality outcomes [29]. Funding challenges have led to delays and fragmented development. Collaboration of automated tools are critical towards overcoming this barrier. In 2015, the International Collaboration for the Automation of Systematic Reviews was founded to exchange data and develop more efficient toolkits. Members include a wide variety of interdisciplinary professions such as software engineers, librarians, statisticians, experts in artificial intelligence, linguists, and researchers [29]. The NSF is also currently funding the University of Arkansas System with an award for collaborative research on automated knowledge discovery in reliability and healthcare from complex data with covariates [30]. This award has already helped produce publications such as “Flexible methods for reliability estimation using aggregate failure-time data” [31] and “Comparison study on general methods for modeling lifetime data with covariates” [32].