Introduction

Collaboration has prompted scientific development and boosted scientific discoveries (Newman 2004; Wuchty et al. 2007; Dong et al. 2017), especially in multidisciplinary research which integrates knowledge from different scientific fields (Lawrence 2007). With the number of coauthors of each paper increasing, it is urgent for scientific community to evolve science’s credit system, which is significant for hiring, funding and promotion decisions (Kennedy 2003; Zeng et al. 2017; Allen et al. 2014). For single-author papers, the credit allocation is simple. The sole author obtains the whole credit. However, the situation is more complicate for multi-authors papers because of the difficulty to quantitatively account for the discrepancy of contribution among coauthors (Sekercioglu 2008; Greene 2007; Kaur et al. 2013). The situation can be even worse when coauthors are from different fields applying different credit allocation criteria in multidisciplinary research (Lehmann et al. 2006). It is necessary for scientific community to develop a deeper understanding of the mechanism of credit allocation which could help to improve this situation and provide a perspective to measure each researcher’s scientific impact (Radicchi et al. 2009; Stallings et al. 2013; Sinatra et al. 2016).

Nowadays, there are four main categories of scientific credit allocation methods in scientometrics. The first is to regard each author as a sole author and assign each author full credit. Naturally, this way leads to inflated scientific impact for multi-authors papers (Garfield 1972; Hirsch 2005 ). The second is to make an equal fraction of credit to co-authors of a paper (Van Hooydonk 1997; Hirsch 2007). This method neglects the fact that not all coauthors contribute equally to a paper and dilutes the credit share of the main contributor. The third is to consider the order or the role of coauthors in a paper. Based on this idea, there are four different ways to assign credit: geometric (Egghe et al. 2000), arithmetic (Trueba and Guerrero 2004), harmonic (Hagen 2008) and network-based (Kim and Diesner 2014). Different credit allocation methods are used based on different disciplines (Tscharntke et al. 2007). Methods depending on author list are adopted by some fields but are not applied to others. For example, in most fields of physics, the first author or the corresponding authors get the most of credit share; while in mathematics and experimental particle physics fields, the method based on author order does not work since author names are listed alphabetically. Considering this complex situation, there is a trend to allocate authors’ credit share by specified contribution declaration required by journals (Foulkes and Neylon 1996; Greene 1999). The fourth one is to treat credit allocation as a collective process (Radicchi et al. 2009; Shen and Barabási 2014; Bao and Zhai 2017). The existing collective credit allocation method (CCA method) proposed by Shen, H. W., & Barabási, A. L. allocates coauthors’ credit based on the citation process by the scientific community. This model didn’t consider the aging effect with the evolving system, which has been improve by Bao and Zhai (2017). The CCA method depends much on citing papers. However, the CCA method does not take citing papers’ quality into consideration. They ignore the fact that papers have different scientific impact. Thus the CCA method can be easily manipulated by adding low impact citing papers. The CCA method should be improved this limitation to enhance its robustness.

The collective credit allocation method can modify h index, since one of disadvantages of h index when ranking scientists is that multiple co-authorship is not taken into account (Batista et al. 2006; Bornmann and Daniel 2007; Burrell 2007; Hirsch 2007). h index is proposed by Hirsch in 2005 to evaluate the scientific impact of a scientist, which is defined as the highest number of publications of a scientist that have been cited h or more times (Hirsch 2005). In order to address the problem of not considering multiple co-authorship, a series of works modifying h index have been done. The works include two aspects, the first is considering the number of coauthors. Schreiber, Egghe and Galam defined \(h_{m}\), fractional h and gh based on fractionally credit allocation in multi authored papers (Schreiber 2008a, b, 2009; Egghe 2008; Galam 2011). The second is considering different roles of coauthors (Tscharntke et al. 2007; Liu and Fang 2012; Ancheyta 2015; Ausloos 2015; Crispo 2015; Vavrycuk 2018; Hirsch 2019). The CCA method and the NCCA method have put forward new ways to allocate credit in multi-authored papers, which provides a perspective to define h index.

In this article, we propose a nonlinear credit allocation method (NCCA method), considering citing papers’ scientific impact rather than recognizing citing papers as the same weight, based on the state-of-the-art collective credit allocation method. Firstly, we get some basic statistics of the NCCA method. Then we further studying the robustness of the NCCA method, find that it is tolerant of random perturbations as well as the malicious manipulation in both Nobel-prize papers and ordinary papers. Finally, the CCA method and the NCCA method are applied to modify h index.

Data

In this paper, the database we use contains all papers published in the journals of American Physical Society (APS), which includes Physical Review series and Reviews of Modern Physics. There are over 450000 papers and over 236000 authors, ranging from year 1893 to year 2010 in the database.

Lacking of unique author identifier in APS database, a name disambiguation process is needed to assign each paper to its true authors. In this paper, we make use of the APS database with author names disambiguated by Sinatra et al. (2016). In addition, we gather Nobel authors and papers from two different papers (Shen and Barabási 2014; Bao and Zhai 2017). As this work studies the credit allocation for multi-author papers, we delete Nobel-prize papers with only one author or more than one laureates.

Method

The collective allocation method (CCA method)

The existing collective credit allocation method (CCA method) can capture the coauthors’ contribution to a paper based on a credit allocation matrix and the co-citation strength of all identified co-cited papers (Shen and Barabási 2014). One of the basic assumptions of their method is that papers citing a target paper contribute equally to each co-cited paper, which makes the method sensitive to the malicious manipulation. For example, there is a paper written by author1 and author2. By the CCA method, we find that the credit of author1 is higher than that of author2 at first (Fig 1a). The credit rank of author1 and author2 can be manipulated by adding only two papers (\(d_{6}\) and \(d_{7}\)) (Fig 1b). It is not difficult to add two papers since these two papers are not required as high quality papers. They only need to be published and cite the target paper \(p_{0}\) and other papers written by author2 not by author1 in the meantime. Additionally, authors of these papers can be anybody.

Fig. 1
figure 1

The collective credit allocation method (CCA method). a A case of the CCA method. The credit share of author \(a_{1}\) is 0.6 which is higher than that of author \(a_{2}\). b A manipulation case of the CCA method. Paper \(d_{6}\) and papers \(d_{7}\) are added manipulatively. Paper \(p_{5}\) and papers \(p_{6}\) are written by author \(a_{2}\), not by author \(a_{1}\). After adding papers, the credit share of author \(a_{2}\) is 0.54 which is higher than that of author \(a_{1}\). In both cases, the target paper \(p_{0}\) has two authors \(a_{1}\) and \(a_{2}\) colored in blue and pink, respectively; \(d_{l}\) (\(1\leqslant l\leqslant 7\)) are citing papers of \(p_{0}\); papers \(p_{j}\) (\(0\leqslant j\leqslant 6\)) were co-cited by papers that cite \(p_{0}\); A is the credit allocation matrix obtained by author list of co-cited papers; S represents the co-citation strength between co-cited papers and target paper; C is the final credit share matrix. (Color figure online)

Fig. 2
figure 2

The maliciously manipulative behavior. a The schematic diagram of the maliciously manipulative behavior. The target paper \(p_{0}\) has two authors \(a_{1}\) and \(a_{2}\). If the credit share of \(a_{1}\) in paper \(p_{0}\) is higher than that of \(a_{2}\) at first according to the CCA method, this credit allocation situation can be changed by adding spamming papers which cite \(p_{0}\) and papers written by \(a_{2}\) but not by \(a_{1}\). b The proportion of Nobel-prize papers affected by the malicious manipulation. The affected rate of Nobel-prize papers means the proportion of papers which the credit share of the original second largest one exceed the original largest one by adding a certain amount of spamming papers. In 17 Nobel-prize papers, for 53%, the original second highest credit share author only needs to add less than 10 papers to make her or his credit share higher than the credit share of the original highest one

We try to test the sensitivity of the CCA method in Nobel papers dataset. First we find out author with the second highest credit share of a given Nobel-prize paper by the CCA method, then we add some papers which cite not only the given Nobel-prize paper but also other three papers in dataset written by that author but not by the author with the highest credit share. We count the number of added papers needed to make the credit share of the second highest author exceed the highest one (Fig. 2a). The result (Fig. 2b) shows that for 53% of 17 Nobel-prize papers, the original second highest credit share author only needs less than 10 added papers to make her or his credit share higher than the credit share of the original highest one. This indicates that if we don’t measure the scientific impact of citing papers, adding low scientific impact citing papers can easily manipulate the credit share of coauthors. Based on this, we propose a nonlinear collective credit method considering citing papers’ scientific impact.

The nonlinear collective allocation method (NCCA method)

We put forward a nonlinear collective allocation method (NCCA method) by integrating the scientific impact of citing papers. We consider a paper \(p_{0}\) with M coauthors \(a_{i}\)(1\(\le \) i \(\le \) m). In order to calculate the credit share of each author, we first find all papers that cite \(p_{0}\) in the dataset, forming a set \(D \equiv \{d_{1},\,d_{2},. . .,\,d_{l}\}\). Then we find all co-cited papers \(P \equiv \{p_{0},\,p_{1},\ldots ,p_{n}\}\), representing the complete set of papers cited by papers in the set D. The relevance of each co-cited paper \(p_{j}\) (0 \(\le \) j \(\le \) n) to the target paper \(p_{0}\) is characterized by its co-citation strength \(s_{j}\) between \(p_{0}\) and \(p_{j}\), defined as the sum of the \(f_{i}\) of papers by which \(p_{0}\) and \(p_{j}\) are cited together in set D. Similar to the basic idea of the well-known PageRank algorithm, we can measure a paper’s scentific impact by measuring the scientific impact of papers that cite it. Citations are usually used to measure papers’ scientific impact. For each paper in set D, we can compute its citations \(c_{l}\), and then we calculate function \(f_{l}\)

$$\begin{aligned} {f_{l}} =\left\{ \begin{matrix} 0 , &{} if &{}c_{l} =0\\ c _{l}^{\theta }, &{} if &{} c_{l} \ne 0 \end{matrix}\right. \end{aligned}$$
(1)

where \(\theta \) is a tunable parameter, \(\theta \geqslant 0\). For example, for \(p_{1}\) in Fig. 3b, we have \(s_{1}\) = \(f_{1}\) because only one paper (\(d_{1}\)) cites \(p_{0}\) and \(p_{1}\) together, whereas \(s_{2}\) = \(f_{1}+f_{2}+f_{3}\) as three papers (\(d_{1}\) , \(d_{2}\) and \(d_{3}\)) cite \(p_{0}\) and \(p_{2}\) together. Note that the target paper \(p_{0}\) is also viewed as a co-cited paper of itself and the co-citation strength equals to the sum of the \(f_{i} \) of citing papers. By that way, we get the co-citation strength vectors (Fig. 3c). After that, we get a credit allocation matrix A by using the author list of the co-cited papers (Fig. 3d). The component \(A_{ij}\) refers to the credit share that author \(a_{i}\) gets from the co-cited paper \(p_{j}\). The value of \(A_{ij}\) is calculated by fractional credit allocation. For example, paper \(p_{2}\) assigns whole credit to the sole author \(a_{1}\), whereas \(p_{0}\) assigns equal (half) credit to authors \(a_{1}\) and \(a_{2}\) (Fig. 3d). Consequently, the total credit \(c_{i}\) of author \(a_{i}\) is the weighted sum of its local credit acquired from all co-cited papers.

$$\begin{aligned} c_{i}=\sum _{j}A_{ij}s_{j} \end{aligned}$$
(2)

or in the matrix form

$$\begin{aligned} C=AS \end{aligned}$$
(3)

The credit of all authors of target paper \(p_{0}\) is provided by the vector C. By normalizing C, the fractional credit share among coauthors can be obtained.

Fig. 3
figure 3

Illustrating the nonlinear credit allocation process. a The target paper \(p_{0}\) has two authors, \(a_{1}\) and \(a_{2}\), colored in blue and pink, respectively. We also show the citing papers \(d_{l}\) (\(1\leqslant l\leqslant 5\)), the citation of citing papers \(c_{l}\) (\(1\leqslant l\leqslant 5\)) and the co-cited papers \(p_{j}\) (\(0\leqslant j\leqslant 4\)) that were cited by these citing papers together with \(p_{0}\). b Calculating the function \(f_{l}\). b The co-citation network of \(p_{0}\) constructed from A, where the weights of links denote the co-citation strengths between the co-cited papers and the target paper \(p_{0}\). We get the matrix S. d The credit allocation matrix A obtained from the author lists. The matrix A provides for each co-cited paper the authors’ credit share. For example, because \(p_{1}\) has \(a_{1}\) as one of its two authors but it lacks the author \(a_{2}\), it votes 0.5 for author \(a_{1}\) and 0 for author \(a_{2}\). With the matrix A and the co-citation strength matrix S, the credit share of the two authors of \(p_{0}\) is computed according to Eq. 2 or  3 with a normalization. (Color figure online)

Computing the \(h_{c}\) index

According to Hirsch’s original definition, the h index can be obtained by

$$\begin{aligned} h=\max \limits _{r}(r\leqslant c(r)) \end{aligned}$$
(4)

r is the rank attributed to the paper when the publication list is in decreasing order based on the number of citations c(r). In Eq. (4) each paper is fully counted for the calculation of its rank,

$$\begin{aligned} r=\sum _{{r}'=1}^{r} 1 \end{aligned}$$
(5)

An effective rank can be yielded by counting a paper with the credit share c(r) from the CCA method or the NCCA method of the author,

$$\begin{aligned} r_{eff}(r)=\sum _{{r}'=1}^{r} c_{r} \end{aligned}$$
(6)

Which can be used to define the \(h_{c}\) index,

$$\begin{aligned} h_{c}=\max \limits _{r}(r_{eff}(r)\leqslant c(r)) \end{aligned}$$
(7)

Results

The basic statistics of the NCCA method

Fig. 4
figure 4

The basic statistics of the NCCA method. a The distributions of citing papers’ citations of Nobel-prize papers. It indicates that there are lots of low scientific impact citing papers. b The mean difference of authors’ credit share between the CCA method and the NCCA method with the change of the value of \(\theta \) in the NCCA method in Nobel-prize papers. The mean difference substantially increases before \(\theta \) < 5. c The distributions of citing papers’ citations of ordinary papers. The pink dots represent papers whose citation range is between 0 and 10. The green dots represent papers whose citation range is between 10 and 100. The blue dots represent papers whose citation range is between 100 and 1000. The phenomenon is that most citing papers’ scientific impact is low in ordinary papers too. d The mean difference of authors’ credit share between the CCA method and the NCCA method with the change of the value of \(\theta \) in the NCCA method in ordinary papers. The pink line represents papers whose citation range is between 0 and 10. The green line represents papers whose citation range is between 10 and 100. The blue line represents papers whose citation range is between 100 and 1000. The mean difference substantially increases before \(\theta \) < 5 too. (Color figure online)

The fundamental difference between the NCCA method and the CCA method is that the NCCA method gives different weights to the citing papers of a given paper when using them to allocate credit to authors. In this paper, citation is defined as the scientific impact. As an empirical support of the NCCA method, we compute the citation distributions of citing papers of Nobel-prize papers and three kinds of ordinary papers (Fig. 4a, c). We find that the citation distributions have fat tails, which indicates that most citing papers have low scientific impact. Due to the significant heterogeneous impact of the citing papers, they should be considered differently when relying on them to allocate credit to authors of a given paper.

Parameter \(\theta \) is important in the NCCA method. To investigate the influence of \(\theta \) on the results, we calculate the mean difference of authors’ credit share between the CCA method and the NCCA method with different \(\theta \) values. The results show that in both Nobel-prize papers and ordinary papers, the mean difference substantially increases before \(\theta \) < 5, and then stays roughly stable afterwards (Fig. 4b, d). In the following discussion, the value of \(\theta \) is set around 1. The \(\theta \) value can adjust the nonlinearity effect of the NCCA method. More weight is assigned to the citing papers with high impact when \(\theta \) is larger. When \(\theta \) is infinitely large, the credit allocation of a given paper is only determined by the highest impact citing paper and the result of this situation of Nobel-prize papers database is showed in Supplementary Material.

Fig. 5
figure 5

Identifying Nobel laureates from Nobel-prize papers by the NCCA method (\(\theta =1\)). For each paper, laureates are shown in red-filled circles. The author with the highest credit share is shown as a black-filled circle when he/she is not a laureate. Other coauthors are shown as empty circles. Hence the presence of black-filled circles indicates that the credit allocation offered by the NCCA method is inconsistent with the decision made by the Nobel committee. If all authors have the same share credit, this situation is considered as identifying the laureates correctly and two Nobel-prize papers have this situation in both the CCA method and the NCCA method. (Color figure online)

It is widely recognized that Nobel laureates are the authors who contribute more in Nobel-prize papers. They deserve more credit share than other coauthors. We apply the NCCA method to 20 multi-author Nobel-prize publications to explore whether Nobel laureates receive more credit than their coauthors. In Fig. 5, the result of the NCCA method with \(\theta \) = 1 at identifying laureates is showed. For each paper, a red-filled circle represents the laureate. A black-filled circle represents the author with highest credit when he/she is not a laureate. Empty circles represent other authors. Therefore, the existence of black-filled circles means that the credit allocation by the NCCA method contradicts to the real case. In 20 Nobel-prize papers, the accuracy of authors with the highest credit share corresponding to laureates by the NCCA method is 75%. However, the accuracy of the CCA method is only 70% (Supplementary Materials: Table 3).

The robustness of the NCCA method against random perturbations

Considering that the results of collective credit allocation methods may be influenced by the reference of some papers missing, we validate the robustness of the CCA and NCCA methods against random perturbations, removing a certain proportion of citing papers of a paper. With perturbations, we apply the CCA method and the NCCA method to Nobel-prize papers and ordinary papers whose citation range are between 10 to 100 and between 100 to 1000.

Firstly, we calculate the change of credit share of coauthors after the deletion of citing papers in each paper. We compute the mean difference of credit share in each kind of data. Then, we investigate whether the author with the highest credit share is the same one after removing a proportion of citing papers in each paper. We calculate the proportion of each kind of papers in which the author with highest credit share does not change after perturbations. From the results (Fig. 6), we find that not only in Nobel-prize papers but also in ordinary papers, the mean difference of credit share of coauthors is small and changes slightly with the proportion of citing papers removed increasing. Furthermore, the proportions of papers are above 95 percent satisfying the condition that authors with highest credit share are the same after perturbations. These results indicate that the CCA method and the NCCA method both are robust against random perturbations.

Fig. 6
figure 6

The robustness of the NCCA method against random perturbations. a The mean difference of authors’ credit share in Nobel-prize papers after removing a certain proportion of citing papers. b The proportion of Nobel-prize papers in which the highest credit share author does not change after removing a certain proportion of citing papers. c The mean difference of authors’ credit share in ordinary papers after removing a certain proportion of citing papers, whose citation range is between 10 and 100 (randomly choosing 100 papers). d The proportion of papers in which the highest credit share author does not change after removing a certain proportion of citing papers. The citation range of those papers is between 10 and 100 (randomly choosing 100 papers). e The mean difference of authors’ credit share in ordinary papers after removing a certain proportion of citing papers, whose citation range is between 100 and 1000 (randomly choosing 100 papers). f The proportion of papers in which the highest credit share author does not change after removing a certain proportion of citing papers. The citation range of papers is between 100 and 1000 (randomly choosing 100 papers)

The robustness of the NCCA method against the malicious manipulation

As we find the CCA method is sensitive to manipulation in Nobel-prize papers, we try to testify whether the NCCA method can suppress the manipulative behavior, adding low scientific impact citing papers. What we calculate is the number of added papers needed to make credit share of the second highest author exceed the original highest one. For the malicious manipulation, we suppose that each added paper randomly cites three papers written by the second highest credit share author but not by the highest one. In 20 Nobel-prize papers, we remove three papers and finally focus on 17 Nobel-prize papers. Since there are two Nobel-prize papers in which authors have the same credit share and there is another Nobel-prize paper in which the author with the second highest credit share does not have three papers satisfying the above hypothesis. For the parameters in the NCCA method, the value of \(\theta \) is 1 and we assume the citation of added papers is 1 or 5. This is because papers in APS database receive 1.38 and 4.17 on average in 1 and 3 years after they are published.

The results of different methods are listed in Table 1. Compared with the CCA method, the NCCA method need more added papers, which means that the NCCA method has the better performance in suppressing the effect of added papers, especially for those added papers with low scientific impact.

Table 1 The number of papers (N_papers) needed to make credit share of the original second highest credit share author exceed the original highest one in Nobel-prize papers by using different kinds of collective credit allocation methods.‘\(\theta =1\), cit=5’ means the NCCA method with parameter \(\theta =1\) and the citation of added spamming papers is 5; ‘\(\theta =1\), cit=1’ means the NCCA method with parameter \(\theta =1\) and the citation of added spamming papers is 1

In Nobel-prize papers dataset, the NCCA method performs well in suppressing the influence of added citing papers with low scientific impact. To verify the NCCA method is also effective to ordinary papers, we again compute the number of added papers needed to make credit share of the second highest author exceed the highest one. For the parameters in the NCCA method, the values of \(\theta \) are set to 0.5, 0.8 and 1, respectively. We assume citations of added papers are 1, 5 and 10.38. The reason why the citation is 10.38 is that the average degree of citation network is 10.38 in the APS dataset. To every kind of ordinary papers, we randomly select 50 papers from that kind of papers in APS database and do it 50 times, then we calculate the number of added papers needed on average. The result is showed in Fig. 7. We find that in all kinds of ordinary papers, the NCCA method always needs more added papers than the CCA method. Besides, more added papers are needed for ordinary papers with larger citations range. Overall, the NCCA method is more robust against the malicious manipulation compared with the CCA method.

Fig. 7
figure 7

The number of papers needed to make credit share of the second highest author exceed the highest one in ordinary papers with different parameters in the NCCA method. ‘CCA\(\_\)method’ means the CCA method; ‘aver\(\_\)cit = 1’ means the citation of added papers is 1 in the NCCA method; ‘aver\(\_\)cit = 5’ means the citation of added papers is 5 in the NCCA method; ‘aver\(\_\)cit = 10.38’ means the citation of added papers is 10.38 in the NCCA method. a Ordinary papers with the range of citation between 0 and 10. b Ordinary papers with the range of citation between 10 and 100. c Ordinary papers with the range of citation between 100 and 1000

The \(h_{c}\) index

Schreiber proposed \(h_{m}\) index which is modified by counting the papers fractionally based on the number of authors. The \(h_{m}\) index can result in a different rank compared with the original h-index. According to the way to compute \(h_{m}\) index, we put forward \(h_{c}\) index modified by the collective credit allocation method. In order to find out whether the \(h_{c}\) index can make any significance when ranking scientists, we select top ten authors according to the number of coauthors in APS dataset and compute their h index, \(h_{m}\) index, \(h_{c0}\) index(the CCA method), \(h_{c1}\) index( the NCCA method with \(\theta =0.5\)) and \(h_{c2}\) index( the NCCA method with \(\theta =1\)). The result is shown in Table 2.

Table 2 The h index, the \(h_{m}\) index, the \(h_{c0}\) index(the CCA method), the \(h_{c1}\) index( the NCCA method with \(\theta =0.5\)) and the \(h_{c2}\) index( the NCCA method with \(\theta =1\)) of top ten authors. “author 1” to “author 10” are ranked by the number of coauthors

We find that \(h_{c}\) index can lead to a different ranking of authors. Although some authors have the same rank in h index and \(h_{m}\) index, the rank of these authors in \(h_{c}\) index changes greatly, like author9 and author10. Compared with h index and \(h_{m}\) index, the value is smaller in \(h_{c}\) index and a few authors have same values in \(h_{c}\) index. In addition, the rank in \(h_{c0}\), \(h_{c1}\) and \(h_{c2}\) are almost same. The credit share of authors calculated by the CCA or NCCA method can have huge difference in different papers even though papers have the same number of authors. The collective process in the scientific community determines the credit allocation in the CCA or NCCA method. That’s why the \(h_{c}\) index is more reasonable than hm index, since the credit allocation is more fair in collective credit allocation method than other methods calculating with the inverse of the number of coauthors.

Discussion and conclusions

In this paper, we propose a nonlinear collective credit allocation method (NCCA method) to assign credits to authors in a given paper. Similar to the existing method (CCA method), the NCCA method is time dependent and is determined by papers citing the target paper. The author whose papers are more co-cited with the target paper will receive higher credit. Compared to the CCA method, the modification is that the NCCA method takes into account the impact of citing papers when evaluating the co-citation relations. This modification can significantly enhance the robustness of the NCCA method with respect to random perturbations and the malicious manipulation. The comparison of the CCA method and the NCCA method is conducted in both Nobel-prize papers as well as ordinary papers. The results indicate that the credit share assigned by the NCCA method is much less influenced by adding citing papers with low scientific impact. In addition, the CCA and NCCA method can be also applied to modify h index. \(h_{c}\) index can result in a different rank and measure the scientific impact of scientists more fairly. However, the procedure of the NCCA method is complicated and thus its wide applicability is questionable.

Additionally, several promising extensions can be made in the future. For example, with different definitions of papers’ scientific impact such as PageRank values, we may find one that can improve the results of the NCCA method. APS dataset only has the physics field data. We need to validate the results of the NCCA method using other fields data.