Introduction

Altmetrics has offered non-traditional way of measuring the diverse impact of scientific product. Altmetrics indicators have been widely adopted by various stakeholders for showcasing the general attention and the potential impact of their work. Many previous studies focused on altmetrics data sources like Twitter (Mohammadi et al. 2018; Yu 2017), Mendeley (Aduku et al. 2017; Zahedi and Haustein 2018) and F1000 (Bornmann and Haunschild 2015). Lately, policy document is studied as a potential useful data source for indicating the societal impact of scientific product (Bornmann et al. 2016), because policy usage of scientific product is supposed to reflect the relationship between academic research and policy making. The policy document is defined very broadly by the Altmetric.com company and refers to any policy, guidance, or guidelines document from a governmental or non-governmental organization.Footnote 1 Compared with other altmetrics data sources, policy document is particularly of interest in twofold, i.e. it is target-oriented and it focuses on a relevant part of society. Wooldridge and King (2018) found that altmetric score is highly correlated with the peer review scores of societal impact and news and policy sources are likely to contain the most informative data to assess impact.

Despite the popularity of altmetrics, many fundamental questions are not answered among which data quality is identified as one of the major challenges (Haustein 2016). According to NISO (2016), accuracy, transparency and replicability are three essential dimensions of data quality. Altmetrics databases are the infrastructure of altmetrics research and services. They constantly collect, clean and store altmetrics data in large scale and provide the data in a systematic and usable way. Altmetric.com database is by far the most commonly used altmetrics database.

Data quality of citation databases

Good quality of bibliographic data is fundamental to bibliometrics research and application. Abundant studies have been dedicated to examine the data quality of citation databases. These studies may provide relevant implication for studies of altmetrics data quality. To sum up, there are two perspectives of studying the topic.

The first perspective is to directly check the accuracy of bibliographic data in the database via content analysis. Investigated databases are mainly multidisciplinary databases like Web of Science (WoS), Scopus and Google Scholar (GS). Data errors can be classified into two major categories (Buchanan 2006), i.e. author error and database mapping error. Author error refers to error that is generated by the cited author, for example, the error can occur in the publication title, publication year, volume number or pagination. Many studies have investigated the first category of data error and reported its percentage. The percentage varies according to the subject area and the scope of author error as defined by the researcher. Based on a large-scale study, Moed (2002) estimated the percentage of author error in SCI database to be 7%. Database mapping error refers to error that is generated by citation databases, for example, transcription error, or cited-article omitted from a cited-article list. However, only a few studies have looked at the second category of data error, because in the past it is difficult to determine the rate of clerical errors introduced in the citation database (Garfield 1974). Franceschini et al. (2014) analyzed the omitted citations and found the omitted-citation rate of Web of Science to be 6%. Many weird errors are discovered in Scopus (Franceschini et al. 2016a). The database mapping error is further classified into errors in the transcription of author names or article title, incomplete cited-article list, omitted cited-article list, wrong or missing DOI, and errors concerning online-first papers (Franceschini et al. 2016b).

The second perspective is to measure the accuracy of bibliographic data by comparing different bibliographic databases. Meho and Yang (2007) compared the data quality of three citation databases, i.e. WoS, Scopus and Google Scholar (GS) and found GS has low accuracy and thus low reliability. Calver et al. (2017) studied the discrepancy of searching result between Scopus, Google Scholar, WoS and WoSCC, and found that retrieval result using the same keyword is different for different databases. Therefore, to improve the consistency across different databases will help users reduce the cost of finding proper database and improve the retrieval efficiency. Donner (2017) studied the influence of inaccurate document type on the citation index and found that data of accurate and inaccurate document type would lead to different citation counts. In addition, several comprehensive studies (Archambault et al. 2009; Harzing and Alakangas 2016; Wang and Waltman 2016) had tapped into the data quality of Scopus, GS and WoS.

Meanwhile, the underlying reasons of the data error and corresponding improving strategies are discussed. It is suggested that authors should be more rigorous and self-disciplined in citing behavior for reducing data error (Zhao 2009). Also, editors should put more efforts to check the authenticity and validity of references (Chen 2014). In database level, dictionary database structure can be used to automatically check the data with programs to reduce human cost (Su 2001). Although not all discovered errors are corrected (Franceschini et al. 2016c), bibliographic databases like Web of Science has made use of these relevant studies to improve their data quality (Prins et al. 2016).

Data quality of altmetrics databases

Similar to the situation of citation indicators, data quality is of critical importance for establishing credibility of altmetrics indicators. Without careful examination of the data quality, the application of altmetrics databases will be severely criticized. Accuracy is above all the basis of data reliability.

Research of altmetrics data quality is still in preliminary stage. Most previous studies compare across different altmetrics aggregators to check the coverage of different altmetrics data sources as collected and maintained by these aggregators (Chamberlain 2013; Peters et al. 2014; Zahedi et al. 2014). It is found that factors such as timing, platform and publication identifier will influence the reported altmetrics data (Meschede and Siebenlist 2018). Even for the same set of papers, different major aggregators provide different metrics (Ortega 2018, 2019). The differences pose challenges to the reliability of altmetrics. Zahedi and Costas (2018) investigated the underlying reasons and found that a range of different methodological, technical and reporting choices have determined the final counts, yet no universe recommendations could be made for data aggregators and users because each practice has its own cons and pros.

As regards the direct checking of the accuracy or completeness of altmetrics data, Zahedi et al. (2014) used a small sample to compare the metadata of publications that were presented in Mendeley and indexed in Web of Science and found that journal and article titles are the most erroneous. Unlike citations, Mendeley reader count is fluctuate and may decrease (Bar-Ilan 2014).

Studies of policy document altmetrics

Policy document usage of academic research is strong evidence of societal impact. According to Newson et al. (2018), two parallel streams of research are contributing to provide insights about the influence of research on policy, i.e. research impact assessments that starts from looking at the research, and research use assessments that starts from looking at the policy document. For research impact assessments, commonly used data sources are interviews, surveys, policy documents, focus groups and direct observation. Case study is the most popular method. Many policy assessment studies use forward tracing approaches to examine single policy document to corroborate claimed impacts. However, these qualitative studies are limited by the data scale and high cost.

Policy document altmetrics has provided the opportunity to measure the usage of scientific product in policy document in a more systematic way. Several studies have explored the usability of policy document altmetrics. Using policy document altmetrics data provided by Altmetric.com company, Bornmann et al. (2016) find that only 1.2% of climate change publications have at least one policy mention. In another study, Haunschild and Bornmann (2017) find that less than 0.5% of papers indexed in Web of Science are mentioned at least once in policy-related documents. The percentage is lower than expected, because Khazragui and Hudson (2015) argue that policy tends to be based upon a large body of work. Nevertheless, Tattersall and Carroll (2018) used policy document altmetrics data collected by Altmetric.com company to evaluate the research from University of Sheffield. It is found that altmetric explorer database has offered important and highly accessible data on the policy impact, but the data must be used with caution because data errors are found.

Research questions

While policy document altmetrics can indicate the societal impact of scientific product, the accuracy of policy document altmetrics data is still unclear. The study is dedicated to measure the accuracy of policy document altmetrics data in Altmetric.com database, in order to inform the potential limitation of research or application that are based on the Altmetric.com policy document data and provide reference for improving the data quality in the future. To be specific, the research questions are:

  1. 1.

    Is there error in policy document altmetrics data? If so, what types of errors may occur?

  2. 2.

    What is the distribution of different categories of errors in policy document altmetrics data?

The result is compared with that of traditional citation databases. Potential underlying reasons of the data errors are discussed.

Methodology

Data source

As one of the major altmetrics data aggregators, Altmetric.com company has started to collect policy document mentions of scientific product in January 2013.Footnote 2 Data were retrieved on June 8th, 2018. In total, there are 1,390,350 policy document mentions of 1,072,397 unique scientific product, from 80,476 unique policy documents that are released by 79 unique source platforms. The top 10 source platforms are listed in Table 1. The length of policy document and the number of scientific products mentioned by the policy document have shown great difference. For example, the longest policy document (entitled with Dietary Reference Intakes: The Essential Guide to Nutrient Requirements) has 1345 pages and mentioned 3731 unique scientific products.

Table 1 Productivity distribution of policy document source platforms (Top 10)

Sampling strategy

To measure the accuracy of policy document mentions, we begin by analyzing the general production process of altmetrics data, as is illustrated in Fig. 1. This can help identify the positions where potential errors may happen. Compared with traditional bibliometrics data, the production of altmetrics data has two additional steps, i.e. production of altmetrics source data (e.g. policy document) by the altmetrics source platform (e.g. policy document website), and production of the altmetrics data in altmetrics database (e.g. Altmetric database) accomplished by altmetrics companies (e.g. Altmetric.com company). Because the quality issue of bibliometrics data has been well studied, in this study, we focus only on these two additional steps of producing altmetrics data. Both steps could potentially introduce data errors. Errors found in these two steps are defined as two categories of data error.

Fig. 1
figure 1

Production process of altmetrics data

Type A error: the error found about mentioned publications in the policy documents, by using the metadata of source publications as standards to compare with.

Type B error: the error found about either the recorded policy documents or the mentioned publications in the altmetrics database, by using the policy documents downloaded from the source website and publications in the source journal as standards to compare with. It must be noted that type B errors are not necessarily produced by the altmetrics database because, for example, third-party data may be used and related errors should not be attributable to the altmetrics database.

According to Altmetric.com company, each policy document source platform has its unique data structure and therefore they have created a personalized crawler for each source platform. Except for that, all records are supposed to share the same data extraction and matching technique that the company has used to automatically collect, merge and visualize altmetrics data. In total, there are 79 unique policy document source platforms. 30 policy documents from each source platform are randomly selected. For policy document source platform of which the total number of captured policy documents is no higher than 30, all policy documents are selected. For each policy document, one record of policy document mention is randomly selected. Each record has the relevant data of a policy document mentioning a scientific product. The final sample is a dataset of 2079 records.

Process of coding a single record

The process of coding a single record is demonstrated in Fig. 2. In general, there are three steps as described below.

Fig. 2
figure 2

Process of coding a single record of policy document mention

  1. 1.

    Click the URL of policy document that is collected and recorded by the Altmetric.com database. If the URL doesn’t work, the title of the policy document is used to search for the policy document, in order to examine the existence of such policy document.

  2. 2.

    After we find the original file of the policy document (mostly in pdf format), we download and open the file to search for the mentioned scientific product as recorded by the Altmetric.com database, in order to examine whether the scientific product is indeed mentioned in the policy document.

  3. 3.

    After we find the location of scientific product (mostly papers) mentioned in the policy document, we cross check the bibliographic information of the scientific product from three sources, i.e. the policy document file, the Altmetric.com detail page and the original file of scientific product. The original file of scientific product is retrieved from full-text databases or open access repositories, and is used as benchmark with which any inconsistency is deemed as error. The error is further classified into different categories. For a certain number of records, two categories of errors could simultaneously occur and are coded. Very few records have more than two categories of errors, only two categories of errors that are supposed to be more important are coded.

Process of coding the whole sample dataset

The coding table is established via three steps.

  1. 1.

    100 records are randomly selected from the sample dataset and are coded by four coders independently. The aim is twofold, to define the possible categories of errors and to standardize the coding process. Based on the coding practice, the initial coding table is created.

  2. 2.

    Another 100 records are randomly selected from the sample dataset and are coded by the four coders using the initial coding table and the standardized coding process. The aim is also twofold, to improve the usability of the coding table and to calculate the coding consistency. A few new categories of error are added to the initial coding table. The rate of coding consistency between the four coders is 86%. The records of coding inconsistency are discussed to understand the underlying reason. Based on the discussion and new discovery of categories of error, the final coding table is obtained (Table 6 and Table 7 in the “Appendix”).

    Type A error is discovered by comparing the original file of scientific product and the information recorded in the policy document. Type B error is discovered by comparing the original file of scientific product, the policy document file and the Altmetric.com detail page.

  3. 3.

    the total sample dataset of 2079 records are coded by two coders using the final coding table. The rate of coding consistency between the two coders is 98%. Records of inconsistent coding are discussed and coded with the help of a third coder.

Result

The general distribution of data errors is shown in Table 2. For the investigated dataset, 73% of records have data errors. The percentage is substantially high. 8% of records have type A error while 70% of records have type B error. 4% of records have both type A error and type B error.

Table 2 Distribution of the first level code data errors

Type A error

Table 3 shows the distribution of type A error. As shown in Table 3, code 1.1 error has the highest percentage. In every twenty policy document mentions of scientific product, one mention will wrongly record the author’s name of the mentioned scientific product.

Table 3 Distribution of the second level code data errors (type A error)

(Code 1.1) Author error of scientific product

A relatively frequent type A error concerns the wrong/missing author’s name of mentioned publication in the policy document. The author’s name of mentioned publication may be misspelled in several ways. Figure 3 (in the “Appendix”) exemplifies that the author “Seager M.” is misspelled as “Seagar M.” in the policy document. In other cases, author whose name is not in English but in other language such as German or Russian may be spelled wrong. In a few cases, author’s middle name is omitted.

(Code 1.2) Title error of scientific product

The second relatively frequent type A error concerns the title of mentioned publication. Figure 4 (in the “Appendix”) exemplifies that in the original publication the title is “Thrive and Survive” while in the policy document it becomes “Succeed”.

(Code 1.3) Publication date error of scientific product

Publication date could also be wrongly recorded by the policy document. Figure 5 (in the “Appendix”) shows that the mentioned publication is in fact published in 1999 but is written as 2009 in the policy document.

(Code 1.4) Source (Journal) error of scientific product

The least commonly seen type A error concern the source of mentioned publication. Figure 6 (in the “Appendix”) shows that the journal Progress in Development Studies is put as Progress of Development Studies in the policy document.

Type B error

For type B error, as shown in Table 4, the code 2.3 transcription error is highlighted and takes up 52% of total erroneous records. It is the major reason of the astonishingly high percentage of type B error. 13% of policy document mentions are false positive where the mentioned scientific product according to Altmetric database is in fact not mentioned by the policy document. The situation of policy document link error is also poor, 5% of the links are no longer accessible.

Table 4 Distribution of the second level code data errors (type B error)

(Code 2.1.1) Policy document link is updated or expired

In a few cases the policy document link is updated or expired but the link is not updated in the Altmetric database. Figure 7 (in the “Appendix”) exemplifies that some policy document is withdrawn and some policy document is updated to another location, as a result, the policy document cannot be visited via the link provided by the Altmetric database.

(Code 2.1.2) Policy document page is not accessible

The most frequent situation where policy documents are not available is just because the page cannot be found any more. Figure 8 (in the “Appendix”) shows that the policy document page cannot be found from the source website.

(Code 2.1.3) Policy document page refers to multiple policy documents

In a few cases, the link from the Altmetric database leads to the page where multiple policy document can be found, but readers are unable to judge which one to look at, as shown in Fig. 9 (in the “Appendix”).

(Code 2.2.1) Policy document has mentioned itself

This is a relatively severe kind of type B error. To our observation, it is usually because the policy document is also a scientific product and there is a suggested citation format for the policy document. The suggested citation is wrongly recognized as a policy document mention. Figure 10 (in the “Appendix”) has exemplified the case. At the end of the policy document entitled with Setting of new MRLs and import tolerances for fluopyram in various crops there is its suggested citation and it is falsely counted as a policy document mention.

(Code 2.2.2) Policy document has not mentioned the scientific product

The relatively severe error is that the policy document in fact has not mentioned the scientific product at all. As an example, Fig. 11 (in the “Appendix”) shows that according to Altmetric database, the policy document has mentioned the publication entitled with Disability living allowance, but this publication cannot be found in the original policy document, instead it merely happens to appear in the full-text. We also have observed other reasons. For example, Fig. 12 (in the “Appendix”) shows that the DOI in the policy document is wrongly detected due to the line break and therefore is mistaken as another irrelevant publication. The DOI of mentioned publication is https://doi.org/10.1016/j.nrl.2016.08.006 but in Altmetric database it is recorded as 10.1016/j. Nevertheless, underlying reason for most of the false positive records are unknown. The supposed mentioned publication just cannot be found in the original policy document by all means.

(Code 2.3.1) Title error of the mentioned scientific product found in the altmetric database

Error can be found in title of the mentioned scientific product in the altmetric database. Figure 13 (in the “Appendix”) exemplifies that part of the title of the publication is missing. The missing part is usually the part after colon, or in a few cases merely the latter part of a long title.

(Code 2.3.2) Author error of the mentioned scientific product found in the altmetric database

Author error of the mentioned scientific product is the predominant type B error. It is the major reason for the high percentage of type B error. Figure 14 (in the “Appendix”) exemplifies the situation where some author of mentioned publication is omitted. The mentioned publication is co-authored by Podgursky M. and Tongrut R., but in Altmetric database Tongrut R. is omitted. Figure 15 (in the “Appendix”) shows that the author Müller-Lissner is misspelled as Muller-Lissner. This is due to the failure of handling non-English letters. Figure 16 (in the “Appendix”) shows that authors of the publication are listed twice in Altmetric database record. The authors are listed in full name once and listed in abbreviated name again.

(Code 2.3.3) Source (Journal) error of the mentioned scientific product found in the altmetric database

We see a few records that source title of the mentioned scientific product is inaccurately recorded in Altmetric database. Figure 17 (in the “Appendix”) shows that the journal of mentioned publication is Gender and Development but in Altmetric database it is written as Gender and Development. This is probably due to improper encoding or decoding by the program.

(Code 2.3.4) Publication date error/inconsistency of the mentioned scientific product found in the altmetric database

Publication date errors are also found in Altmetric database. Figure 18 (in the “Appendix”) shows that the mentioned publication was published in 2009 but the publication date is omitted in Altmetric database. Figure 19 (in the “Appendix”) shows that the mentioned publication was published in 2013 but in Altmetric database the publication date is recorded as 2012.

(Code 2.3.5) Title error of the policy document found in the altmetric database

A relatively small percentage of type B error concerns the title of policy document. Figure 20 (in the “Appendix”) shows that the title of policy document recorded in the Altmetric database is inconsistent with the original title.

Discussion and conclusion

Comparison of data accuracy between policy document mention and citation data

The data quality of policy document altmetrics is compared with that of citation data. Authors of citing publication may produce errors to data of cited publication, similarly, authors of policy document may produce errors to data of mentioned scientific product. Citation database uses program to automate the identification and extraction of citation data, similarly, altmetrics database uses program to crawl and collect data of policy document, but altmetrics database usually gets bibliographic data directly from the third-party citation database instead of collecting the bibliographic data on their own. In case of Altmetric database, CrossRef is the major bibliographic data provider.

For type A error, Table 5 has compared the error rate caused by author of policy document and author of citing publication. It is found that percentage of data errors of mentioned scientific product in policy document are all slightly higher than that of cited publications in citing publications, indicating no significant difference.

Table 5 Comparison of type A errors introduced by authors of policy document and citing publications

For type B error, because sub-categories of error in altmetrics database are different from that in citation database, they are not directly comparable. However, the transcription error (52%) of altmetrics database is obviously much higher than that of citation database which is 3.5% according to Franceschini et al. (2016b).

Possible reasons and influence of the data errors

To begin with, errors in this study are defined from user’s perspective. Suppose an end user of altmetric database is browsing the database, the user would probably be unable to figure out the underlying reasons of the inconsistencies and regard them as errors. There could be other perspectives, for example, from researcher’s perspective, the first online date as the publication date is sometimes more accurate in reflecting the velocity of data accumulation (Fang and Costas 2020). Or, from altmetric database’s perspective, bibliographic data are correct as long as they are in consistency with the data provided from the third party, because it is not their duty to check the accuracy of the bibliographic data. Based on the definition of error from user’s perspective, we further discuss the possible reasons and their influences.

Type A errors are from the authors of the policy documents. Therefore, the underlying reason for type A error is the incaution of authors of policy document. Omission and misspell of title, author or publication date in the references could be attributable to this reason. Interestingly, sometimes altmetrics databases are able to eliminate such kind of error, because Altmetric database doesn’t extract bibliographic data of the mentioned publication from the policy document, instead they use third-party citation database which extract bibliographic data from the source publications. For this reason, in practical usage of altmetrics data, type A error seems to be not a big problem,

Type B errors are further classified into three categories, error code 2.1 (policy document link error) is related to the provenance of policy document, error code 2.2 (false positive policy document mention) is related to the false positive mentions, error code 2.3 (transcription error) is related to the metadata quality of the mentioned records.

For error code 2.1, the underlying reason may be the lack of updating the old records. Due to the dynamic nature of policy document websites, the policy document could be revised and thus updated to a new URL. In other cases, the policy document could be retracted or the entire website section is no longer active. Policy document link error (percentage is 5%) has corroborated that altmetrics data are highly influenced by the stability of source web platform. Once the platform has removed the link or retracted the policy document, the policy altmetrics data are gone forever. Theoretically, Altmetric database could update periodically and revisit the policy document link to check whether it is still available or linking to the right policy document, but it may not be economically feasible for the company. Therefore, this type of error is difficult to get rid of.

Moreover, according to the comments from experts of Altmetric.com company, Altmetric database has chosen to keep records of which the policy document link is no longer accessible, because policymakers sometimes remove policy documents for very good reasons (e.g. clinical care guidelines issued by UK’s NICE that go out of date). Those citations are kept in the database because researchers typically want to know if they’d been cited previously, even if those documents are no longer available. In a way, it’s akin to citation databases indexing citations from print journals that aren’t available online. There’s certainly an element of trust that the citation found previously was correct. As a result, error code 2.1 would not be a severe problem but will cause inconvenience if content analysis of policy document is to be conducted.

For error code 2.2, the underlying reason may be bugs in the program for processing the data. Some of the bugs are minor and easy to get fixed, for example bugs that are related to encoding or decoding. These bugs make the special letter like “&” not properly presented. Some of the bugs are severe and could be easily fixed, for example, the omission or duplication of authors, the error is probably related to the underlying data merging process. Meanwhile, some of the bugs are severe and not easy to find out, for example, the DOI could be wrongly extracted from the policy document file when there is line break in the DOI string. The program is also unaware of the fact that some policy document gives its suggested citation at the end of the text and wrongly identifies that as a policy document mention. It can be avoided by excluding self-mentions. Error code 2.2 could bring very misleading results, because many related altmetrics analysis are based on the count of policy document mentions. The percentage is 12.6% and cannot be neglected.

For error code 2.3, the underlying reason may be attributable to the third-party bibliographic data providers. Therefore, Altmetric database could report these errors to the bibliographic data providers and help them improve the metadata quality. In error code 2.3, publication date error could be difficult to define, because a scientific product could have multiple publication date (Haustein et al. 2015). However, in this study, publication date in the published journal is used as standard publication date, because it is commonly used for selecting data for scientometric analysis. Consequently, many records in altmetrics database that use first online mention date as publication date are deemed erroneous, although they may be more accurately regarded as inconsistency. The error code 2.3 is trivial or can even be neglected for conducting altmetrics analysis, because researchers commonly use bibliographic data from traditional citation database, only unique identifiers (e.g. DOI) provided by Altmetric database are used to match for corresponding altmetrics data. But error code 2.3 could cause potential inaccuracy if metadata are used to retrieve in the Altmetric database and later used directly for analysis.

Further remarks

Results of the study have made it clear that altmetrics data, especially policy document altmetrics data, are far from being perfect. Future studies based on the data need to be aware of this limitation. Although the overall error rate in Altmetric database is much higher than that of traditional citation database, majority of the errors are produced either by the policy document website or the third-party bibliographic data provider. These errors are trivial and can even be neglected in conducting altmetrics research. Moreover, in June 2019 we did a random but not systematic re-check of the erroneous records, and found that most errors related to bibliographic information, such as omission of publication date, duplication of authors and misspell of journal title, seem to be gone.

The error code 2.2 (false positive policy document mention) is perhaps the only error type that could be directly improved by the Altmetric database. The error type would have negative influence on altmetrics research. In June 2019, the random check showed the false positive policy document mentions were still there. For code 2.2.1 error, it can be avoided by removing self-citing policy document. For code 2.2.2 error, more studies are needed to find out the underlying reasons. However, considering the high complexity of extracting mentions of publications from various sources and formats of policy documents, Altmetric database has achieved quite excellent performance.

Due to high labor cost of manual coding work, the study is based on analysis of a relatively small sample of the policy document altmetrics data. This also needs to be taken into consideration in interpreting the results of the study.