Introduction

The h-index has been a well honored concept since it was proposed by Hirsch (2005). A lot of variations have been proposed in the literature, see for instance the references within (Alonso et al. 2009). Many efforts enhanced the original h-index by taking into account age-related issues (Sidiropoulos et al. 2007), multi-authorship (Hirsch 2010), fractional citation counting (Katsaros et al. 2009), the highly cited articles (Egghe 2006). Other works explored its predictive capabilities (Hirsch 2007), its robustness to self-citations (Schreiber 2007), etc. Some of the proposals have been implemented in commercial and free software, such as MatlabFootnote 1 and the Publish or Perish software.Footnote 2

Even though there are several hundreds of articles developing variations to the original h-index , there is notably little research on making a better and deeper exploitation of the “primitive” information that is carried by the citation curve itself and by its intersection with the \(45^{\circ }\) line defining the h-index . The projection of the intersection point on the axes creates three areas that were termed in (Rousseau 2006; Ye and Rousseau 2010; Zhang 2009) as the h-core-square area,Footnote 3 the tail area and the excess area (see Fig. 1). The core area is a square of size \(h\) (depicted by grey color in the figure), includes \(h^{2}\) citations and it is also called Durfee square area (Anderson et al. 2008); the area that lies to the right of the core area is the tail or lower area, whereas the area above the core area is the excess or upper or \(e^2\) area (Zhang 2009). Both the absolute and the relative sizes of these areas carry significant information. The absolute size of the excess and core areas were directly used for the definition of e-index and h-index ; part of the absolute size of the tail area was used by García-Pérez (2012) to create a vector of h-indices; the relative size of the core to the tail area (without taking into account the tail length) was used by Ye and Rousseau (2010) for similar purposes, etc. (For a complete review of the relevant bibliography cf. section "Relevant work".)

The common characteristic of all these works is that they develop indices to “break ties”, i.e., to differentiate between scientists with equal h-indices.

Fig. 1
figure 1

Citation curve depicting the excess, core and tail areas. (Color figure online)

We believe that the latent information carried by these areas is not adequately explored, and most significantly, it can be used in a different way, not as a plain tie-breaker, but as a “first-class” citizen in the scientometric indices ecosystem.

Rosenberg (2011) took the first step towards this goal; he provided a qualitative characterization for the scientists with many citations in the upper area and a few citations in the tail area by referring to them as perfectionists. He referred to the authors with few citations in the upper area and many citations in the tail area as mass producers, since they have a lot of publications but mostly of low impact. Finally, he referred to the rest of the scientists as the prolific ones. The origin of this terminology is quite old; it was proposed by Cole and Cole (1967), and subsequently studied further by Feist (1997).

Motivated by Rosenberg’s classification scheme, we pose the following question: “Can we develop a quantitative methodology for identifying those scientists who are truly laconic and constantly influential compared to those who produce a mass of papers but relatively few of them contribute to their h-index ?”

In the present paper, we will present a methodology and an easy calculated criterion to categorize a scientist in one of two distinct categories: either an author is a “mass producer”, i.e., he has authored many papers with relatively few citations or he is “influential”, i.e., most of his papers have an impact because they have received a significant number of citations. This methodology will indirectly highlight the “attitude” towards publishing. Some scientists are acting in a laconic way, in the sense that they are not fond of having published “half-baked” articles that are soon superseded by mature and extended versions of their work. Others develop their work in a slow and incremental way, publishing their ideas in a step-by-step fashion producing a lot of moderate impact articles until they hit the big contribution. This attitude may be due to other reasons as well, e.g., the pressure to present published articles as deliverables to a project. In any case, it is not the purpose of the present article to discover and explain those reasons. The sole purpose of our work is to develop metrics that can be used complementary to the traditional ones such as the h-index , in order to separate the steadily influential authors from the mass producers. At this point we need to emphasize that the concept of “influential” scientists we develop here is not related to the notion of influential nodes in a social network of actors as considered by Basaras et al. (2013).

The area of scientometric performance indicators is very rich, and it is continuously flourishing. Vinkler (2011) provides a brief classification of the traditional and modern scientometric indicators explaining their virtues and shortcomings; it is shown there that Hirsch index is not the only indicator that combines impact and quantity, but \(\pi \)-index (Vinkler 2009) which introduced the concept of “elite set” is another competitor of it. Nevertheless, in this article we use Hirsch index as a basis to expose our ideas claiming that neither Hirsch index is the ’best’ one nor that our core methodology applies exclusively to it. We are strongly confident that our ideas can be applied also to the family of indices based on the Impact Factor by penalizing those journals that publish articles which accumulate far less citations than the Impact Factor of the journal they appear in.

The rest of the article is organized as follows. In the next section we will present the relevant literature and then define two new areas in the citation curve. Based on these two new areas, we will establish two new metrics for evaluating the performance of authors in terms of impact. In section "Penalty areas and the Perfectionism Indices" we will present our datasets, which were built by extracting data from the Microsoft Academic Search database, and analyze these data to view the dataset characteristics. Subsequently, we will present the distributions of our new metrics for the above datasets and compare them with other metrics proposed in the literature. Finally, in section "PI in action: Ranking scientists" we will present some of the resulting ranking tables based on the new metrics and h-index . Section "Conclusions" will conclude the article.

Motivation and contributions

During the latest years an abundance of scientometric indices have been published to evaluate the academic merit of a scientist. Despite the debate around the usefulness of any index in general, they remain an indispensable part of the evaluation process of a scientist’s academic merit. The ideas behind the h-index philosophy was so influential, that the vast majority of the proposed indices are about some variant or extension of the h-index itself. Despite the wealth and sophistication of the proposed indices, we argue that the relevant literature did not strive for an holistic consideration of the information carried by the citation curve and by the \(45^{\circ }\) line. In the next paragraph we will present the motivating idea with a simple example.

Let us consider author \(a\) who has published 13 articles, and author \(b\) who has published 24 articles with citation distributions \(\{29,24,20,17,15,14,13,12,11,10,9,3,0\}\) and \(\{29,24,20,17,15,14,13,12,11,10,2,1,1,1,1,1,1,1,1,1,1,0,0,0\}\) respectively. Both authors have the same “macroscopic” characteristics in terms of the number of citations, i.e., they both have the same total number of citations, identical core areas and h-indices equal to 10, identical excess areas with 65 citations there, and the same number of citations in the tail area, namely 12. However, author \(a\) has only 3 articles in his tail area, whereas author \(b\) has 14 articles.

At a first glance, we can simply use the number of articles in the tail as a tie-breaker to differentiate between the two authors, and characterize the first one as constantly “influential”, and the second one as a “mass producer”. But, how can we capture the fact that, in the short term, the first author’s h-index is more likely to increase. At the same time, we need a way to describe—actually, to penalize—the second author for this long and lightweight tail. Starting from these questions we will define the penalty areas and then develop the respective indices. In this article, we do not consider temporal issues, e.g., the time of publication of the articles in the tail area; such issues are part of our on-going work. Specifically, the article makes the following contributions:

  • It defines two areas to quantify the fact that some authors publish articles which eventually do not have analogous impact with those that contribute to their h-index .

  • It develops two new perfectionism indices taking into account the size of the penalty areas. There are the PI and XPI indices, which are statistically uncorrelated to the h-index , thus proving that they measure something that is different from what the h-index measures.

  • Using these indices, it proposes a filter to separate the authors into influential ones and mass producers. This filter partitions the authors irrespectively of their h-index , i.e., it can classify two authors as influentials, even if the values of their h-indices are significantly different. This two-way segmentation of the scientists is a significant departure from the earlier, rich classification schemes (Cole and Cole 1967; Feist 1997), since with a single integer number and its sign (plus or minus) it can provide rankings, contrary to intuitive mapping schemes such as that by Zhang (2013b).

  • It provides a thorough investigation of the indices against the h-index for three datasets retrieved by the Microsoft’s Academic Search.

At this point, we need to emphasize again that the proposed indices are neither a substitute for any of the already existing metrics nor a tool for identifying “bad” publishing behaviour. They are one more tool in the indices toolbox of someone who wishes to capture the multi-dimensional facet of a scientist’s performance.

Relevant work

The original article by Hirsch (2005) created a huge wave of proposals for indices attempting to capture the academic performance of a scientist. It is characteristic that at the time of writing the present article, the h-index ’s article had more than 3850 citations in Google Scholar. Since the focus of the present manuscript is not about the h-index in general, but about the exploitation of the information in the tail area, we will survey only the articles relevant to the usage and mining of that part of the citation.

Ye and Rousseau (2010) studied the evolution of tail-core ratio as a function of time, and later extended their study by Liu et al. (2013) including a few more ratios among the three areas. Similar in spirit is the work reported by Chen et al. (2013), which examines variations of the ratios across scientific disciplines. Baum (2012) introduced the ratio (the relative citedness) of the few, highly-cited articles in a journal’s h-core and the many, infrequently-cited articles in its h-tail as a way to improve journals’ Impact Factors.

Having as motive to consider each and every citation under the whole citation curve (and therefore under the tail area as well), Anderson et al. (2008) proposed a fractional citation counting scheme based on Ferrers graphs. Later, Franceschini and Maisano (2010) recognized the weaknesses of that scheme and proposed the Citation Triad method; both indices are striving to exploit the information under the whole citation curve in a way that creates a strictly monotonic (increasing) index for every new citation added to the curve, which is completely different to what we propose.

A kind of “quantization” scheme for the citation curve and the creation of multiple Durfee squares under that curve was proposed by García-Pérez (2012). The output of that method was a vector (i.e., multiple h-indexes) as a measure of the scientific performance. However, the method simply transformed the task of comparing different citation curves into the problem of comparing vectors, without setting clear rules. A study of the contribution of the excess, core and tail areas to the entire citation curve was performed by Bornmann et al. (2010) proving that this contribution varies across scientists. The study provided also a regression model for determining the most visible article of a scientist. The position of the centroids of the core and tail area was used by Kuan et al. (2011) as an index for comparing different scientists providing only straightforward characterizations for high-low impact and productivity. Along these lines of research, Zhang (2013b) proposed a triangle mapping technique to map the three percentages (of the excess, core and tail area) of these citations onto a point within a regular triangle; by viewing the distribution of the mapping points, different shapes of citation curves can be studied in a perceivable form. The work described by Dorta-González and Dorta-González (2011) sought selective and large producers considering only a part of the excess and a part of the tail area, thus again neglecting a part of the tail area which carries significant information.

The most closely relevant articles to our work are these from Rosenberg (2011) and Zhang (2013a). Rosenberg (2011) described a three-class scheme for scientists’ classification based on the length and thickness of the tail of the citation curve. Zhang (2013a) proposed the \(h^{\prime }\) index as a quantitative measure to discover which scientist belongs to which one of those three categories.

Collectively, the present work differentiates itself by the previous studies in a number of factors: (a) it exploits the full spectrum of information under the citation curve, (b) it is based on the definition of new areas (not below, but above the citation curve), (c) it penalizes those scientists with long and thin tails, (d) it proposes an index that can be used as a filter to separate the constantly influential scientists from the mass producers.

Penalty areas and the Perfectionism Indices

In this section, we will define the penalty areas which form the basis for the development of the respective scientometric indices. Before proceeding further, we summarize in Table 1 some basic symbols related to the productivity and impact of a scientist, along with their description and the relations among them.

Table 1 Basic symbols and their interpretation

The tail complement penalty area

We now get back to the motivating example presented in the previous section, and we illustrate graphically their citation distributions (see Fig. 2). We depict with red solid color the h-core area of each author.

Fig. 2
figure 2

Citation curves for two sample authors (a, b). (Color figure online)

It is intuitive that long tails and light-weight tails reduce an author’s articles’ collective influence. Therefore, we argue such kind of a tail area should be considered as a “negative” characteristic when assessing a scientists’s performance. The closer the citations of the tail’s articles get to the line \(y=h\), the more probable it is for the scientist to increase his h-index , and at the same time to be able to claim that practically each and every article he publishes does not get unnoticed by the community.

For this purpose, we define a new area, the tail complement penalty area, denoted as TC-area with size \(C_{TC}\). The size of the tail complement penalty area is computed as follows:

$$\begin{aligned} C_{TC} = \sum \limits _{\forall i \in P_T}(h-C_i) = h \times (p-h) - C_T. \end{aligned}$$
(1)

This area is depicted with the green crossing-lines pattern in Fig. 2, and fulfilling the motivation behind its definition, it is much bigger for author \(b\) than for author \(a\).

The ideal complement penalty area

If we push further the idea of the tail complement penalty area, we can think that “ideally” an author could publish \(p\) papers with \(p\) citations each and get an h-index equal to \(p\). Thus, a square \(p \times p\) could represent the minimum number of citations to achieve an h-index value equal to \(p\). Along the spirit of penalizing long and thin tails, we can define another area in the citation curve: the ideal complement penalty area (IC-area), which is the complement of the citation curve with respect to the square \(p \times p\). Figure 3 illustrates graphically the IC-area with the green crossing-lines pattern. The size of the IC-area (\(C_{IC}\)) can be computed as follows:

$$\begin{aligned} C_{IC} = \sum \limits _{\forall i \in P\ \wedge \ C_i<p}{(p-C_i)}. \end{aligned}$$
(2)
Fig. 3
figure 3

Graph illustrating all (existing and proposed) areas. (Color figure online)

Apparently, this area does not depend on the h-index value, as it holds for the case of the TC-area. Notice that the IC-area includes the TC-area defined in the previous paragraph. We realize of course that it is hard (if possible at all) to find scientists—with sufficiently large h-index , and—with zero-sized TC-area. Therefore, the index derived by this area is not expected to provide significant insights into scientists’ performance.

The new scientometric perfectionism indices: PI and XPI

The definition of the penalty areas in the previous subsection, allows us to design two new metrics which will act as the filter to separate influential from mass producers. Firstly, let us introduce the concept of Parameterized Count, \(PC\), as follows:

$$\begin{aligned} PC = \kappa *h^2 + \lambda *C_E + \mu *C_T \end{aligned}$$
(3)

where \(\kappa ,\lambda ,\mu \) are real numbers. Therefore, we define PC as the parameterized addition of the three areas we defined earlier. Apparently:

  • when \(\kappa =\lambda =\mu =1\), then it holds that \(PC=C\),

  • when \(\kappa =1\ \wedge \ \lambda =\mu =0\), then \(PC=h^2\),

  • when \(\lambda =1\ \wedge \ \kappa =\mu =0\), then \(PC=C_E=e^2=C_{h}-h^{2}\),

  • when \(\mu =1\ \wedge \ \kappa =\lambda =0\), then \(PC=C_T\).

By assigning positive values to \(\kappa \) and \(\lambda \), but negative values to \(\mu \), we can favor authors with short and thick tails in the citation curve. Even in this way, we cannot differentiate between the authors \(A\) and \(B\) of our example.

For this reason, instead of using the tail of the citation curve, we use the tail complement penalty area. Thus, similarly to Eq. 3, we define the concept of Perfectionism Index based on TC-area as follows:

$$\begin{aligned} {\rm PI} = \kappa *h^2 + \lambda *C_E - \nu *C_{TC} \end{aligned}$$
(4)

In the experiments that will be reported in the next sections, we will use the values of \(\kappa =\lambda =\nu =1\). We experimented with various combination of values for the parameters, but we used integer values equal to one, because these default values give a straightforward geometrical notion of the newly defined metric. Noticeably, it will appear that PI can get negative values. Thus:

  • if an author has PI\(< 0\), then we characterize him as a mass producer,

  • if an author has PI\( > 0\), then we characterize him as an influential.

In the same way as the PI’s definition, we define an extremely perfectionism metric, the Extreme Perfectionism Index, taking into account the ideal complement penalty area, as follows:

$$\begin{aligned} {\rm XPI} = \kappa *h^2 + \lambda *C_E + \mu *C_T - \nu *C_{IC}. \end{aligned}$$
(5)
Table 2 Traditional and proposed indices for authors \(a\) and \(b\)

As in the previous case, we will assume that \(\kappa =\lambda =\mu =\nu =1\). We will show in the experiments, that very few authors have positive values for this metric. Using the previously defined perfectionism indices, the resulting values for authors \(a\) and \(b\) are shown in Table 2. Author \(a\) has greater values than author \(b\) for both XPI and PI perfectionism indices. This is a desired result.

Before proceeding to the next section, which describes the detailed experiments that demonstrate the merits of the new indices, we provide an additional example of five authors with different publishing patterns, as an extension to our artificial motivating example which was presented in the beginning of the article. We use only initials but they refer to real persons and their data. In Table 3 we present the raw data (i.e., h-index , number of publications \(p\) and number of citations \(C\)) of 5 authorsFootnote 4 selected from Microsoft Academic Search.Footnote 5 The last column shows the calculated PI values, which can be positive as well as negative numbers. In Fig. 4 we present citation plots for these five authors.

Fig. 4
figure 4

Real examples. (Color figure online)

In Fig. 4a we compare three authors: AuthorA, AuthorB and AuthorC. They correspond to real names but we have preferred to present them anonymously. They all have an h-index equal to 10. Note that AuthorA has a comparatively large number of publications but the citation curve is cropped to focus on the lower values. As he has the bigger and longest tail (red line with crosses), he could be characterized as a “mass producer”. This is reflected in a PI value of −2,505 as shown in Table 3. AuthorB (green line with diagonal crosses) has shorter tail than AuthorA and higher excess area. From the same table we remark that his PI value is 11 (i.e. close to zero). Finally, the last author of the example, AuthorC (blue line with asterisks), has similar tail with AuthorB but a bigger excess area \((e^2)\). Definitely, he demonstrates the “best” citation curve out of the three authors of the example. In fact, his PI score is 717, higher than the respective figure of the other two authors.

Table 3 Computed h-index and PI values for 5 sample authors

In Fig. 4b, again we compare three authors: AuthorD, AuthorE and AuthorC. The first two have h-index value equal to 15. Comparing those, it seems that AuthorD (red line with crosses) has a better citation curve than AuthorE (green line with diagonal crosses) because he has a shorter tail and a bigger excess area. Indeed, the first one has PI = 717, whereas the second one has PI = −2,523. AuthorC (blue line with asterisks) has a smaller tail and a big excess area but since there is a difference in h-index we cannot say for sure if he must be ranked higher or not than the others.

In Fig. 4c we have scaled the citation plots so that all lines cut the line \(y=x\) at the same point. From this plot, it is shown that AuthorC has a better curve than AuthorE because he has a shorter tail and a bigger excess area. When comparing AuthorC to AuthorD, we see that the latter has a longer tail but also a bigger excess area. Both curves show almost the same symmetry around the line \(y=x\). That is why they both have similar PI values. This is a further positive outcome as authors with different quantitative characteristics (say, a senior and a junior one) may have similar qualitative characteristics, and thus be classified together.

In general, there is no upper (lower) limit for the size of the PI index. This depends on the productivity (number of articles) and impact (citation) distribution of each scientist. However, a positive PI means that the scientist is a perfectionist one, and a negative PI indicates a mass-producer (Table 4).

Table 4 Symbols used and their interpretation

Experiments

In this section, we will present the results of the evaluation of the proposed indexes. The primary goal of our experimentation is to study the merits of the PI index, since our results confirmed that the severe penalty that XPI imposes makes it a less useful scientometric tool. Firstly, we will explain the procedures for dataset acquisition, then we will present their characteristics, and finally we will give the results that concern the evaluation of PI.

Datasets acquisition and characterization

During the period December 2012 to April 2013, we compiled 3 datasets. The first one consists of randomly selected authors (named “Random” henceforth). The second one includes highly productive authors (named “Productive”). The last one consists of authors in the top h-index list (named “Top h”). The publication and the citation data were extracted from the Microsoft Academic Search (MAS) database using the MAS API.

The dataset “Random” was generated as follows: We fetched a list of about 100,000 authors belonging to the “Computer Science” domain as tagged by MAS. MAS assigns at least three sub-domains to every author. These three sub-domains may not all belong to the same domain (e.g., Computer Science). For example, an author may have two sub-domains from Computer Science and one from Medicine. We kept only the authors who have their first three sub-domains belonging to the domain of Computer Science. From this set, we randomly selected 500 authors with at least 10 publications and at least 1 citation.

The dataset “Productive” was generated as follows: from the set of 100,000 Computer Science authors we selected the top-500 most productive. The least productive author from this sample has 354 publications.

The third dataset named “Top h” was generated by querying the MAS Database for the top-500 authors in the “Computer Science” domain ordered by h-index .

Table 5 summarizes the information about our datasets with respect to the number of authors (line: # of authors), number of publications (line: # of publications), number of citations (line: # of Citations) and average/min/max numbers of citations and publications per author.

Table 5 Statistics of the datasets used in our study
Fig. 5
figure 5

Distributions of h-index , \(m\), \(p\), \(C\) indices (Left plots CDFs. Right plots PDFs) a \(h\)-index, b \(h\)-index, c \(C\,(*10,000)\), d \(C\,(*10,000)\), e \(m\), f \(m\), g \(p\,(*1,000)\) h \(p\,(*1,000)\). (Color figure online)

Figure 5 shows the distributions for the values of h-index , \(m\), \(C\) and \(p\). The \(m\) index was defined in Hirsch’s original article (Hirsch 2005) and is explained (quoting Hirsch’s text) in the next paragraph. Plots are illustrated in pairs. The ones on the left show cumulative distributions. For example, in Fig. 5a we see that 80 % of the authors in the sample “Random” (red solid line) have h-index less than 10. It is obvious that the sample “Top h” (blue dotted line) has higher values for the h-index . Figures 5c, d show the distributions for the total number of citations. As expected the sample “Top h” has the highest values.

Figures 5e, f show the distributions for the \(m\) value. Recall its definition from Hirsch (2005): A value of \(m\approx 1\) (i.e., an h-index of 20 after 20 years of scientific activity), characterizes a successful scientist. A value of \(m \approx 2\) (i.e., an h-index of 40 after 20 years of scientific activity), characterizes outstanding scientists, likely to be found only at the top universities or major research laboratories. A value of \(m \approx 3\) or higher (i.e., an h-index of 60 after 20 years, or 90 after 30 years), characterizes truly unique individuals. Indeed, only a few authors have \(m>3\).

Figure 5 g, h illustrate the distributions for the total number of publications. It is obvious that in the “Random” sample (red solid line) there are relatively low values for the total number of publications. Also, as expected, the distribution for the “Productive” sample has the highest values for the total number of publications.

For the completeness of the dataset description we present Fig. 6. In this plots is shown the distribution of the Career length of researchers in our samples. This information was necessary in order to compute the \(m-index\) presented in Fig. 5. The figures  6a, b depicting carreer length, show that since the distributions of carreers of top and productive are similar, this means that the PI index mainly characterizes the scientist’s behaviour towards publising and not the length of its scientific life

Fig. 6
figure 6

Distributions of career length (Left plot CDFs. Right plots PDF) a Career Length bCareer Length. (Color figure online)

We have conducted further experiments to study the behavior of other indices such as \(\alpha \) (Hirsch 2005) and \(e^2\) (Zhang 2009). However, the results did not carry any significant information, and therefore, the figures for these factors are not presented.

Does PI offer new insights about the impact and publication habits of scientists?

The first question that needs to be answered is whether a new index offers something new and different compared to the existing (hundreds of) indices. The answer is positive; our metric separates the rank tables into two parts independently from the rank positions.

Fig. 7
figure 7

Correlation of PI to standard bibliometric indices. (Q-Q plots: X- and Y-axis denote normalized rank positions (%).) a \(h\) versus \(C\) (unioned), b \(h\) versus PI (unioned), c \(C\) versus PI (unioned), d \(C/p\) versus PI (unioned), e \(C/p\) versus \(h\) (unioned). (Color figure online)

In Fig. 7a the x-axis denotes the rank position (normalized percentagewise) of an author by h-index , whereas the y-axis denotes the rank position by the total number of citations \((C)\). Each point denotes the position of an author ranked by the two metrics. Note that all three samples are merged but if the point is blue asterisk, then the author belongs to the “Top h” sample, if the point is green diagonal cross then he belongs to the “Productive” sample etc. If an author belongs to more than one sample, then only one color is visible since the bullet overwrites the previous one. From Fig. 7a the outcomes are:

  • “Top h” authors are ranked in the first 40 % of the rank table by h-index , as well as in the top 40 % by the total number of citations \((C)\).

  • “Productive” authors are mainly ranked by h-index between 30 and 70 %. The rank positions by \(C\) are between 20 and 70 %.

  • “Random” authors are mainly ranked below 60 % for both metrics with some outliers in the range 0–60 %, mostly by \(C\).

The aforementioned conclusions are as expected; it also occurs that h-index ranking does not differ significantly from the \(C\) ranking; i.e., they are correlated which is consistentFootnote 6 with earlier findings (Spruit 2012).

In Fig. 7b the h-index ranking is compared to PI ranking. It can be seen that there is no correlation between PI and h-index . Note that the horizontal line at about 32 % (also shown later in Table 6) shows the cut point of PI for the zero value. Authors that reside below this line have \({\rm PI}>0\) and authors above this line have \({\rm PI}<0\). This observation strengthens the motivation of the article; only one out of three authors is a perfectionist one. Even among those with high h-index (Table 6) only half of them are truly laconic.

  • “Top h” authors are split to two groups. The first group is ranked in the top 20 % of the PI rank table. The second group is ranked in the last 50 %. These two groups are also separated by the zero line of PI.

  • “Productive” authors are almost all ranked at lower positions by PI than by h-index . Almost all points reside above the PI zero line and also above the line \(y=x\) (with some exceptions at about 65–70 % of the rank list).

  • “Random” authors are also generally higher ranked by PI than by h-index . They are also split into two groups by the line \({\rm PI}=0\).

From the above, it seems that PI is not correlated to h-index , whereas the line \({\rm PI}=0\) plays the role of a symmetric axis. Thus, it emerges as the key value that separates the “influential” authors from the “mass producers”.

In Fig. 7c we compare PI ranking against \(C\) (total number of citations) ranking. It is expected that the plot would be similar to Fig. 7b based on the similarity of h-index with \(C\).

In Figs. 7d, e we compare h-index and PI with the average number of citations per publication (\(C/p\)) ranking. It is apparent that PI is not correlated to \(C/p\). h-index is also uncorrelated to \(C/p\), however the points of the qq-plot in Fig. 7e are closer to the line \(x=y\) than the points of Fig. 7d.

Conclusively, the PI ranking is not correlated to h-index , \(C\) and \(C/p\).

Fig. 8
figure 8

Correlation of Academic Trace to PI, XPI, C, hindex. (Q-Q plots: X- and Y-axis denote normalized rank positions (%).) a \(T\) versus PI, b \(T\) versus XPI c \(h\) versus \(T\) d \(C\) versus \(T\). (Color figure online)

We have also implemented several comparisons of PI and XPI with various type of ranking indices. We do not include all of them for brevity. All them show that our new metric is not tied correlated with none of them. One interesting metric is the “Academic Trace” (Ye and Leydesdorff 2013). This metric looks similar to PI in the way that it has a negative factor for the zero cited publications. For this reason we prove that neither PI nor XPI is correlated to “Academic Trace” metric.

Ye and Leydesdorff (2013) defined the metric named “Academic Trace” as:

$$\begin{aligned} T=tr(V) = \frac{h^2}{p}+\frac{C_T^2}{C}+\frac{C_E^2}{C}-\frac{p_z^2}{p} \end{aligned}$$

where \(p_z\) is the number of the zero cited publications. Figure 8 shows the comparison of Academic Trace \((T)\) with PI, XPI, \(C\) and \(hindex\). As it is shown in Fig. 8a, b, PI and XPI are not correlated to \(T\). That is because the negative factor of \(T\) (the \({p_z^2}/{p}\)) is usually too small so it does not affect significantly the resulting score. Academic Trace can only be used with the notion of a time window (like Impact Factor). When used with a time window the average number of zero cited papers increases and this negative factor affects the ranking. So, Academic Trace can mainly be used with a temporal notion. In addition, Fig. 8c, d show that it is correlated with \(hindex\) and especially with the total number of citations \((C)\), except for some outliers of the “Productive” sample which are ranked last with \(T\) but they are about at 70th (of 100) position by \(hindex\) and \(C\). These researchers probably have a very big set of zero cited publications. As a conclusion we can say that Academic Trace is strongly correlated with the total number of citations, and when used with a time window is an improvement to Impact Factor as it penalizes the big sets of zero cited publications (but not the once or twice cited onesFootnote 7).

Aggregate analysis of the datasets

Figure 9 shows the distributions for the areas defined in the previous section. In particular, Fig. 9a, b illustrate the distributions for the \(C_T\) (tail) area. It seems that the “Top h” cumulative distribution is very similar to the “Productive” one, however, the “Top h” distribution has slightly higher values.

Fig. 9
figure 9

Distributions of \(C_T\), \(C_{TC}\) (tail complement), \(C_{IC}\) (ideal complement). (Left plots: CDFs. Right plots: PDFs) a \(C_T\) (*1,000) b \(C_T\) (*1,000) c \(C_{TC}\) (*10,000) d \(C_{TC}\) (*10,000), e \(C_{IC}\) (*100,000), f \(C_{IC}\) (* 100,000). (Color figure online)

Figure 9c , d illustrate the distributions for the \(C_{TC}\) (tail complement) area. It seems that \(C_{TC}\) has the same distribution as \(C_T\) for all samples except for the sample “Productive”, for which \(C_{TC}\) has slightly higher values than \(C_T\) does. Note, also, that the “Productive” distribution has lower values for h-index than “Top h”. This means that the height of the \(C_{TC}\) areas is smaller for the “Productive” authors than for “Top H’ ones. The previous two remarks lead to the (rather expected) conclusion that the “Productive” authors have long and thin tails.

The \(C_{IC}\) distribution is shown in Fig. 9e, f. In these plots, it is clear that the “Productive” authors have clearly higher values than any other sample, since \(C_{IC}\) is strongly related with the total number of publications.

Fig. 10
figure 10

Distributions of PI, \({\rm PI}_{(\kappa =2)}\) and \({\rm PI}_{(\kappa =4)}\) a \({\rm PI} (* 10,000)\) b \({\rm PI} (* 10,000)\) c \({\rm PI}_{(\kappa =2)} (* 10,000)\) d \({\rm PI}_{(\kappa =2)} (* 10,000)\) e \({\rm PI}_{\kappa =4} (* 10,000)\) f \({\rm PI}_{\kappa =4} (* 10,000)\) g \({\rm PI} (limited\, \hbox {x} range)\) h \({\rm PI}_{\kappa =4} (limited\, \hbox {x}\, range)\). (Color figure online)

In Fig. 10 we see the distributions for the previously defined PI index. For all plots, the zero y-axis is the center of the figure. As seen in Fig. 10a, b most of the authors are located around zero. Note that in the right plots, a point at \(x=0, y=95\,\%\) with a previous value of \(x=-3{,}000\) means that the 95 % of the authors have values in the range \(-1{,}500, \ldots, 1{,}500\). The first two plots show that the “Top h” authors have the highest values for PI (about 10 % of them have values greater than 8,000). Interestingly, it seems that the value 0 is a key value. This practically means that the majority of the authors follows a conservative approach toward publishing; they publish significant articles hoping to attract also a significant number of citations. However, the fact that we can encounter authors far away (to the left and to the right) from zero, strengthens the value of the present research, because it shows two things: (a) there are scientists which publich very aggressively and they end up being “mass producers”; (b) there are a few scientists who only publish when they have produced ground-breaking results that really advance their field. We strongly believe that this is not a random concidence. The bell-shaped curves of “Top-h” and “Productive” authors confirm that there is a publishing pattern; otherwise, the curves would be flat ones. These outliers (scientists at the far left and at the far right) reveal the information that we are seeking, i.e., the mass-producers and the perfectionists. Another generic conclusion drawn from the figures is that it is very likely to find a perfectionist who is also a “Top-h” rather than a “Productive” one. This means that the “Top-h” scientists are more “selective” towards publishing, but this not the rule. Its exceptions also strenghten the value of the PI index.

Figure 10g is a zoomed-in version of Fig. 10a. It is clear that about 96 % of the “Productive” authors have \({\rm PI}<0\). This means that in this sample there are a lot of “mass producers” (people with high number of publications but relatively low h-index- or at least not in “Top h-indexers”). The other samples cut the zero y-axis at about 50 to 60 %, which means that 40 to 50 % are positive. It is also noticeable that about 70 % (15–85 %) of the “Random sample have values very close to zero within the range −200, …, 200.

In Fig. 10c–f) we present the distributions for \({\rm PI}_{\kappa =2}\) and \({\rm PI}_{\kappa =4}\). We remind that factor \(\kappa \) is the core area multiplier. In these plots, it is shown that these distributions behave like the basic PI distribution except that they are slightly shifted to the right. The “Productive” sample is affected less than the others. This outcome is understandable since they are the authors with small h-index core areas compared to their tail and excess areas.

Comparing subfigure 10h to g we can better see the differences. The number of authors in the negative side of samples “Random” and “Top h” has decreased from 57 and 58 to 24 and 23 % respectively, meaning that about 33–35 % of the sample members moved from the negative to the positive side. The number of “Productive” authors in the negative side has been decreased from 97 to 88 %, i.e. an additional 11 % of the sample members moved to the positive side.

In addition to the distribution plots, Table 6 presents the number of authors that have the mentioned metrics below or above zero for each sample. As mentioned before, 97 % of the “Productive” authors have \({\rm PI}<0\), whereas only 3 % reside in the positive side of the plot. This amount increases as we increase the core factor \(\kappa \). For \(\kappa =4\) the increment is 9 % (12 % from 3 %). In all other samples the increment is greater, i.e. for “Top h” the increment is 35 %, for “Random” is 33 %.

Fig. 11
figure 11

Distributions of XPI, a \({\rm XPI} (* 100,000)\), b \({\rm XPI} (* 100,000)\), c \({\rm XPI} (limited\,\hbox {x}\, range)\). (Color figure online)

In Fig. 11 the same kinds of plots are presented for the metric XPI. As expected, the difference is that most of the authors lie in the negative side of the graph. The cut points of y-axis are also presented in Table 6. About 2 % of the “Top h-index ” authors have \({\rm XPI}>0\) but none of the “Productive” authors do. The cut point for “Random” authors is at 6 %. Also, at this point we repeat the experiment of varying the \(\kappa \) value. The results do not match with those of the PI case. Incrementing \(\kappa \) does not increase the number of positive authors in the same way as the PI case. The increment is negligible for the “Productive” and “Top h” and very small for the sample “Random”. This leads to the conclusion that varying the \(\kappa \) factor does not affect XPI significantly. Probably different default values for the factors of Eq. 4 (especially for \(\kappa \) and/or \(\nu \)) may be needed for tuning the XPI metric. However, this task remains out of the scope of the present article.

Table 6 PI and XPI statistics

PI robustness to self-citations

Self-citations (a citation from an article to another article when there is at least one common author between the citing and the cited paper) is a common way for authors to increase the visibility of their works. It has been documented that self-citations can have significant impact upon h-index (Schreiber 2007). However, self-citations do not necessarily represent a bad practise, that the scientometric indices should punish. In many cases (Katsaros et al. 2009), “they can effectively describe the authoritativeness of an article.” Therefore, the aim is to design robust metrics (Katsaros et al. 2009) that will be unaffected by self-citations (Katsaros et al. 2009).

Fig. 12
figure 12

Robustness of h-index and PI to self-citations (Q-Q plots: X and Y axis denote the rank position normalized in percent) a \(h\) versus \(h\)(no-self) bPI versus PI(no-self). (Color figure online)

We performed an experiment to study the behavior of h-index and PI with respect to self-citations. In Fig. 12(a) a qq-plot is shown, which compares the ranking produced by h-index . The x-axis represents the rank produced by the computed h-index including self-citations, whereas the y-axis represents the rank of h-index after excluding self-citations. We have performed several experiments with different types of ranking and they all show similar behavior with respect to h-index . In Fig. 12(b) the same kind of qq-plot for the PI as a rank criterion is displayed. It is apparent that PI is much less affected by self-citations than the h-index . This is another advantage of the proposed metric; it is not affected by self-citations.

PI in action: Ranking scientists

In the previous two subsections, we performed an analysis of the datasets at a coarse level. In this section, we will provide an analysis at a finer level, that of individual scientists. We have emphasized from the beginning of the article that it is not this article’s purpose to explain the roots of the publishing behaviour of individual scientists. However, we will attempt to record those characteristics of the scientists (if there are such characteristics) that make them exhibit particular behaviours.

Table 7 Ranking by h-index (top-20 scientists)

Table 7 shows the rank table for the top-20 authors by h-index from all our samples. They are truly remarkable scientists with significant contributions to their field. The table also shows their corresponding PI values; it is remarkable that about half of them are characterized as “Mass Producers” (i.e., they have negative PI values). We will seek an explanation for that by contrasting these results in Table 8.

Table 8 Ranking by PI index (top-20 influential scientists)

Table 8 shows the rank list ordered by PI; all authors have high ranking positions by h-index as well. If we try to find what is common in all these persons, we could say that (most of) these scientists spend significant time of their careers in industrial environments making groundbreaking contributions, and being recognized as inventors whose ideas have been incorporated into many products that penetrated our lives. Examples include Tuecke, Rivest, Shamir, Agrawal, and Lamport. The personnel in these environments are highly trained, working on “real” problems whose solutions are part of business products. Thus, these groups are not publishing-prone, and (most of the time) whenever they publish their results, these are path-breaking and influential. Others, such as Vapnik, Zadeh, Kohonen, Aho and Schapire are pioneers, inventing brand new knowledge and developing it in a long series of articles. It might also be the case that these scientists work only with experienced researchers, thus being elitists (Cormode et al. 2013), because for instance their topics are very advanced. On the contrary, people coming solely from academic environments have the role of a mentor (Cormode et al. 2013) and are charged with the task of training young PhD students whose initial works (usually) do not have high impact. Moreover, sometimes they are involved in projects of exploratory nature, which eventually do not open new avenues. Finally, we should not forget the publish-or-perish pressure upon their students and themselves.

Table 9 Ranking by PI (bottom-20 by PI, i.e., top-20 mass producers)

Table 9 shows the top-20 “Mass Producers” from our samples. In this table we also present the average number of citations per paper (\(C/p\) column). It can be seen that there is a big range of average values from 4 to 45 citations per publication in the top “Mass Producers”. In this table we will recognize—consistent with what we said in the previous paragraph – some excellent academics who have trained many PhD students.

Table 10 Rank table by PI of sample “Networks”

In Tables 10, 11 and  12 we present the “toppers” with respect to the fields of “Networks” “Databases” and “Multimedia”, respectively. Starting from the “Networks’ table, we see Van Jacobson and Sally Floyd ranked first and second respectively; they are well-known inventors who contributed fundamental algorithms to the design and operation of the Internet. They both had careers in industry: in Cisco, Xerox, AT&T Center for Internet Research at ICSI, and worked extensively on developing standards (RFC) in the areas of TCP/IP congestion control. Similarly, looking at Table 11 for the database field, we will find in the top positions persons such as Rakesh Agrawal, Ronald Fagin, Jeffrey Ullman and Rajeen Motwani who also have spent their careers in companies such as Google, IBM and Microsoft, or have contributed fundamental algorithms in fields such as compilers, databases and algorithms. We can make similar observations from Table 12 where we find some entrepreneurs such as Ramesh Jain who founded or co-founded multiple startup companies including Imageware, Virage and Praja. The type of career is certainly a factor that helps categorize a scientist as an influential, since we can see that Van Jacobson (“Networks”) and Nick Koudas (from “Databases” who spend part of his career in AT&T) are the ones who gained the greatest rise in PI ranking compared to the h-ranking: 43 and 38 positions, respectively.

Table 11 Rank table by PI of sample “DataBases”

If we turn our attention to the bottom rows of these tables we will recognize some excellent mentors, but mass producers: Elisa Bertino and Jiawei Han from databases, Georgios Giannakis and Jack Dongarra from the networking community, Thomas S. Huang, Rama Chellappa and Ioannis Pitas from multimedia.

Table 12 Rank table by PI of sample “Multimedia”

But, is it really the case that only inventors and industry persons are influentials, whereas academia persons are mass producers? In that case, the PI index would be of little usefulness since the separation of influentials and mass producers would be quite straightforward. The answer to this question is definitely negative. From the beginning of our article we emphasized that this is a generic attitude of the scientists towards publishing, rather than an outcome of their type of careers. Thus, we can see in the “Networks” field some academia persons such as Hari Balakrishnan, David Johnsonand Ion Stoica, or Peter Buneman from “Databases” who are quite high in the PI ranking, even though they did not develop their careers in companies working with highly trained colleagues. On the other hand, P. S. Yu (“Databases”) who spend many years in IBM is found at the end of Table 11, remaining in the top-50 of “Databases”.

Conclusions

The development of indices to characterize the output of a scientist is a significant task not only for funding and promotion purposes, but also for discovering the scientist’s “publishing habits”. Motivated by the question of discovering the steadily influential scientists as opposed to mass producers, we have defined two new areas on an scientist’s citation curve:

  • The tail complement penalty area (TC-area), i.e., the complement of the tail with respect to the line \(y=h\).

  • the ideal complement penalty area (IC-area), i.e., the complement with respect to the square \(p \times p\).

Using the aforementioned areas we defined two new metrics:

  • The perfectionism index based on the TC-area, called the PI index.

  • The extreme perfectionism index based on the IC-area, called the XPI index.

We have performed an experimental evaluation of the behavior of the PI and XPI indices. For this purpose, we have generated three datasets (with random authors, prolific authors and authors with high h-index ) by extracting data from the Microsoft Academic Search database. Our contribution is threefold:

  • We have shown that the proposed indices are uncorrelated to previous ones, such as the h-index .

  • We have used these new indices, in particular PI, to rank authors in general and, in particular, to split the population of authors into two distinct groups: the “influential” ones with \({\rm PI}>0\) vs. the “mass producers” with \({\rm PI}<0\).

  • Also, we have shown that ranking authors with the PI index is more robust than h-index with respect to self-citations, and we applied it to rank individual scientists offering some explanations for the reasons behind their publishing habits.

We are already involved in the consideration of temporal issues into PI by integrating the concepts of contemporary h-index (Sidiropoulos et al. 2007) into the PI index.