Abstract
Average journal impact factor (JIF) percentile is a novel bibliometric indicator introduced by Thomson Reuters. It’s of great significance to study the characteristics of its data distribution and relationship with other bibliometric indicators, in order to assess its usefulness as a new bibliometric indicator. The research began by analyzing the meaning of average JIF percentile, and compared its statistical difference with impact factor. Based upon factor analysis, the paper used multivariate regression and quantile regression to study the relationship between average JIF percentile and other bibliometric indicators. Results showed that average JIF percentile had changed the statistical characteristic of impact factor, e.g. improved the relative value of impact factor, having smaller variation coefficient and distribution closer to normal distribution. Because it’s non-parametric transformation, it cannot be used to measure the relative gap between journals; Average JIF percentile had the highest regression coefficient with journal impact, followed by timeliness and lastly the citable items; The lower the average JIF percentile, the higher the elastic coefficient of journal impact; When average JIF percentile was extremely high or extremely low, citable items were not correlated with the average JIF percentile at all; When average JIF percentile was low, elastic coefficient of timeliness was even higher; Average JIF percentile was not a proper indicator for multivariate journal evaluation; Average JIF percentile had both the advantages and disadvantages of impact factor, and thus had the same limitation in applying as the impact factor.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
On June 18th, 2016, Thomson Reuters issued the latest JCR report 2015 in which two novel bibliometric indicators were launched. One was “average journal impact factor (JIF) percentile”, the other one was “normalized eigenfactor”. Average JIF percentile converted JIF ranking into percentile to provide more intuitive transverse comparison between journals (Thomson Reuters 2016). Analysis of average JIF percentile would help understand characteristics of the indicator and provide guidance to journal evaluation, which was of significant value.
It had been a long tradition since percentile was introduced into bibliometric indicators. Early in 1930s, Bradford (1985) introduced the well-known Bradford’s law, in which scientific journals were sorted by number of publications in descending order, and then divided into core section, relevant section and irrelevant section, with each section having the same number of publications, interestingly the ratio between number of journals in each section was found to be 1:n:n2. Many other scholars continued to conduct various studies on the distribution law of journal level bibliometric indicators. Hintze and Nelson (1998) proposed that using percentile to normalize number of citations can better evaluate the comprehensive impact of journal from different disciplines, document type and publication time. Bornmann and Mutz (2011) divided percentile into 6 grades: top 1 %, 99–95 %, 95–90 %, 90–75 %, 75–50 %, 50–0 %, which was called percentage rank 6 (PR6), and intended to use it as indicator of evaluating journal. Later, he developed the idea and, together with Leydesdorff, proposed integrated impact indicators (I3) indicator (Leydesdorff and Bornmann 2011) which used percentile function beyond six grades. Leydesdorff (2012) then investigated the relationship between “integrated impact indicators” (I3) and the first 10 and 25 % of citation frequency to see whether it can surrogate impact factor. Campanario (2014) analyzed publications of Spanish university and introduced h core indicators based on citation frequency of articles, and proposed evaluation methods using 10, 20 or 30 % percentile.
The basic reason of adapting percentile rank of a certain indicator to evaluate journals was due to the data distribution features of bibliometric indicators as well as the aim of better comparison between journals. Vinkler (2008) found that citation distribution was right skewed and not normal distribution. Adler et al. (2009) thought, according to power law, citation distribution was usually right skewed, and as a result arithmetic mean only represented the citation value of highly cited articles. Percentiles have several advantages in general (Bornmann et al. 2013), they are not strongly influenced by extreme values, do not rely on the choice of a specific probability density function, can be calculated even if the data are skewed and once calculated, they also permit the classification of publications into meaningful citation impact classes (Rousseau 2012; Bornmann 2013). But percentiles have deficiencies such as overvaluing or devaluing an object because of percentile ranking schemes, ignoring precise citation variation among those ranked next to each other, and inconsistency caused by additional papers or citations (Zhou and Zhong 2012) The average JIF percentile is an indicator that intends to combine the advantages of percentiles in many respects and has kept the precision of traditional JIF.
Because average JIF percentile was a completely novel indicator, what are the characteristics of the indicator like? Is it different from impact factor? And what’s the relationship between the indicator and other established bibliometric indicators? These questions remained to be further explored. The paper started by analyzing the meaning of average JIF percentile and adopted statistical methods to compare its difference with impact factor. Factor analysis was used to further analyze the relationship between average JIF percentile and other bibliometric indicators.
Data and processing
The research used data from JCR2015. In order to improve the robustness, economics discipline with comparatively large number of journals was used to do the study. Mainly 12 indicators were published in JCR2015: average journal impact factor percentile, total cites, journal impact factor, impact factor without journal self cites, 5-year impact factor, immediacy index, eigenfactor score, article influence score, normalized eigenfactor, cited half-life, citing half-life and citable items. Use the uppercasing in the name of the indicators as they appear in the JCR.
There were 333 economics journals in JCR2015. 51 journals were eliminated due to missing data so 282 journals remained. Table 1 shows descriptive statistics of each indicator.
Comparative study of average journal impact factor percentile and impact factor
According to Thomson Reuters, the definition of average JIF percentile (JIFP) is as follows.
In formula (1), JIFP was average JIF percentile, N was the number of journals in the discipline, R was the rank position of impact factor sorted in descending order. 0.5 was added to get rid of value 0 when calculating average JIF percentile. Obviously, transformation from impact factor to average JIF percentile was non-parametric transformation. Average JIF percentile can reflect the ranking of journals, but cannot be used to reflect the relative gap between journals. For example, journal A has impact factor 5.0 and average JIF percentile 80; journal B has impact factor 4.0 and average JIF percentile 40. It can be claimed that impact of journal B is 0.8(4.0/5.0) of that of journal A, but it’s absolutely wrong to claim according to the average JIF percentile that impact of journal B is 0.5(40/80) of that of journal A. Average JIF percentile originated from impact factor. They have exactly the same meaning but of different representation form. Moreover, journal ranking based on these two indicators were totally the same. Scatter plot of them were shown in Fig. 1.
Because most journals have low impact factor value, the ranking transformation rendered journals with relatively low impact factor value obtain higher average JIF percentile. In general, average JIF percentile had higher value than impact factor.
After normalization of average JIF percentile and impact factor, namely divided respectively by the maximum value and multiplied by 100, the result is shown in Fig. 2. When impact factor was of high value, it had good degree of differentiation, while average JIF percentile had lower distinctive capability but was generally in uniform distribution.
To further compare the statistical characteristics of impact factor and average JIF percentile, we used minimum to maximum ratio, median to maximum ratio, coefficient of variation, skewness, kurtosis, and Jargue–Bera test. Results are shown in Table 2.
Observing the statistical characteristics, average JIF percentile had higher minimum to maximum ratio, median to maximum ratio than impact factor while had lower coefficient of variation, which indicated that average JIF percentile, after data transformation, had both improved relative value of impact factor and decreased data fluctuation. Observing the data distribution, although both indicators didn’t obey normal distribution, average JIF percentile was closer to normal distribution and can serve better to compare journals.
According to Table 2, average JIF percentile, compared with IF, had higher minimum maximum ratio and median maximum ratio, but lower coefficient of variation, which indicated that after data transformation, the average JIF percentile had improved relative value of impact factor and decreased data fluctuation. Although neither indicators have normal distribution, average JIF percentile can serve better in comparing journals because it’s more like normal distribution.
Relationship between average JIF percentile and other indicators
Multivariate regression analysis was commonly used to analyze the relationship between average JIF percentile and other bibliometric indicators. Because bibliometric indicators had high mutual correlation, traditional regression method inevitably had multicollinearity problem that would make the variation of regression coefficient increase and t test value decrease etc. To solve this problem, the paper first used factor analysis to extract key factors from bibliometric indicators other than average JIF percentile. Since key factors had no correlation between each other, the multicollinearity problem was completely solved.
Prior to factor analysis, KMO and Bartlett test must be conducted to decide whether factor analysis was applicable. Results showed that KMO test value was 0.811, which was higher than empirical value 0.5. Bartlett test value was 7769.165 and concomitant probability was 0.0000. The hypothesis of not using factor analysis was rejected, indicating that factor analysis was applicable.
Factor analysis showed that 3 factors had eigenvalue over 1. Variance contribution of the first factor was 53.94 %, variance contribution of the second factor was 15.68 % and variance contribution of the third factor was 10.89 %. The accumulated variance contribution was 83.59 %, which meant these three factors explained 80.52 % of the total information and meanwhile mutually were not correlated.
Rotated matrix is illustrated in Table 3. The first factor (F1) reflected total cites, impact factor, impact factor without journal self cites, 5-year impact factor, immediacy index, eigenfactor value, paper influence score, normalized eigenfactor, which mainly explained the journal impact; The second factor (F2) mainly reflected citable items; The third factor (F3) mainly reflected cited half-life and citing half-life. The regression results were as follows:
Generally speaking, as transformation of impact factor, average JIF percentile had no high goodness of fit with other bibliometric indicators. R 2 value was 0.476. Impact got the highest regression coefficient. Elastic coefficient was 1.086 which meant that with each 1 % increase of journal impact, the average JIF percentile will increase 1.086 %; it’s followed by regression coefficient of half-life which was −0.708, indicating that with every 1 % decrease of journal half-life, the average JIF percentile would increase 0.708 %. It’s normal that regression coefficient was negative because the third factor was immediacy indicator. Longer half-life meant the journal less timely and impact decreasing. Citable items had the lowest regression coefficient with the elastic coefficient 0.381, which meant that on every 1 % increase of citable items, the average JIF percentile would increase by 0.381 %.
Because the average JIF percentile emphasized more on ranking position, in order to analyze the relationship between average JIF percentile with other bibliometric indicators at different quantile value, quantile regression was used to do the study. In total there were 9 quantiles as shown in Table 4.
To more clearly analyze the relationship between the average JIF percentiles at different quantile value with other bibliometric indicators, the results were mapped in Fig. 3. It was found that regression coefficient of impact was gradually decreasing, i.e. when average JIF percentile was low, regression coefficient of impact was even higher.
The relationship between regression coefficient of citable items and average JIF percentile was in horizontal curve. But when average JIF percentile was very high or very low, citable items had nothing to do with average JIF percentile. This was in accordance with reality, because for journals with very high average JIF percentile, they relied on quality of articles to accumulate impact.
When average JIF percentile was at the lowest 10 quantiles, it had the highest regression coefficient with journal half-life. In other situations, it was relatively stable.
Conclusions and discussions
Firstly, average JIF percentile had changed the statistical characteristics of impact factor. Compared with impact factor, average JIF percentile improved the relative value of impact factor, had smaller coefficient of variation and the data distribution was closer to normal distribution, which was good for horizontal comparison among journals. But because average JIF percentile was kind of transformation based on impact factor ranking, it’s in fact non-parametric transformation, the average JIF percentile can only reflect ranking order while being unable to measure the relative difference of impact between journals.
Secondly, average JIF percentile had medium correlation with other bibliometric indicators. Average JIF percentile had the highest regression coefficient with journal impact indicators that included total cites, impact factor, impact factor without journal self cites, 5-year impact factor, immediacy index, eigenfactor value, article influence score, normalized eigenfactor. It’s followed by journal timeliness indicator that included cited half-life and citing half-life. The lowest regression coefficient was with citable items.
Thirdly, quantile regression analysis on journal impact, timeliness, citable items and average JIF percentile showed that the lower average JIF percentile, the higher that elastic coefficient of journal impact; When the average JIF percentile was very high or very low, citable items had nothing to do with average JIF percentile, because for journals with very high average JIF percentile, they relied mainly on quality of articles to achieve impact; When average JIF percentile was very low, the elastic coefficient of journal timeliness was even higher.
Fourthly, average JIF percentile was not a proper indicator for journal multi-attribute evaluation. Because average JIF percentile was in fact a ranking indicator of non-parametric nature, it’s improper for evaluating indicator system but could be used for direct comparison among the standing of journals being more intuitive.
Fifthly, average JIF percentile had the same advantages and disadvantages of impact factor. Thus, it had the same limitation in application fields as the impact factor.
References
Adler, R., Ewing, J., & Taylor, P. (2009). Citation statistics. Statistical Science, 24(1), 1–26.
Bornmann, L. (2013). The problem of percentile rank scores used with small reference sets. JASIST, 64(3), 650.
Bornmann, L., Leydesdorff, L., & Mutz, R. (2013). The use of percentiles and percentile rank classes in the analysis of bibliometric data: Opportunities and limits. Journal of Informetrics, 7(1), 158–165.
Bornmann, L., & Mutz, R. (2011). Further steps towards an ideal method of measuring citation performance: The avoidance of citation (ratio) averages in field-normalization. Journal of Informetrics, 5(1), 228–230.
Bradford, S. C. (1985). Sources of information on specific subjects. Journal of information Science, 10(4), 173–180.
Campanario, J. M. (2014). Analysis of the distribution of cited journals according to their positions in the h-core of citing journal listed in Journal Citation Reports. Journal of Informetrics, 8(3), 534–545.
Hintze, J. L., & Nelson, R. D. (1998). Violin plots: A box plot-density trace synergism. The American Statistician, 52(2), 181–184.
Leydesdorff, L. (2012). Alternatives to the journal impact factor: I3 and the top-10 % (or top-25 %?) of the most-highly cited papers. Scientometrics, 92(2), 355–365.
Leydesdorff, L., & Bornmann, L. (2011). Integrated impact indicators compared with impact factors: An alternative research design with policy implications. Journal of the American Society for Information Science and Technology, 62(11), 2133–2146.
Rousseau, R. (2012). Basic properties of both percentile rank scores and the I3 indicator. Journal of the American Society for Information Science and Technology, 63(2), 416–420.
Thomson Reuters. JCR Reports. (2016). http://ipscience-help.thomsonreuters.com/incitesLiveJCR/glossaryAZgroup/g8/9586-TRS.html. Accessed 19 Mar 2016.
Vinkler, P. (2008). Introducing the current contribution index for characterizing the recent, relevant impact of journals. Scientometrics, 79(2), 409–420.
Zhou, P., & Zhong, Y. (2012). The citation-based indicator and combined impact indicator—new options for measuring impact. Journal of Informetrics, 6(4), 631–638.
Acknowledgments
This research was supported by China Scholarship Council (No. 201506270024) and the National Natural Science Fund of China (NSFC) (No. 71303179). The authors would thank anonymous reviewers for their helpful comments.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Yu, L., Yu, H. Does the average JIF percentile make a difference?. Scientometrics 109, 1979–1987 (2016). https://doi.org/10.1007/s11192-016-2156-2
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11192-016-2156-2