Introduction

Ranking of researchers/authors in their respective domain can assist in multiple aspects. For instance, ranking of authors can help organizers of journal or conferences to locate a reviewer for a paper, it can help Ph.D. students to find a supervisor and also it can help any organization to recruit a candidate for a job. Researchers/Authors in any field are ranked on the basis of the impact of their research work in the scientific community. Moreover, ranking of authors helps to answer the questions pertaining to the issues like who should get the scholarship; who has more impactful research work; who should get tenure and who should be given promotions?

To date, various techniques have been introduced in the literature for ranking of authors. Each ranking technique has different criteria to rank authors. Some techniques rank quantitatively and some techniques rank authors by considering qualitative expertise. To measure the performance of an author, different author ranking techniques employ the parameters including publication count (Balog et al. 2006; Fang and Zhai 2007), citation count (Cameron et al. 2007; Borgers et al. 2008), co-authors (Deng et al. 2012; Sun et al. 2013) and hybrid approaches (Afzal and Maurer 2011; Hirsch 2005). Furthermore, many variants and extensions of h-index have also been proposed (Bornmannet al 2011) which include: a-index (Jin 2006), ar-index (Jin 2007), contemporary h-index (Sidiropoulos et al. 2007), raw h-rate (Burrell 2007), m-quotient (Hirsch 2005), f-index (Tol 2009), t-index (Tol 2009), Maxprod index (Kosmulski 2007), q2-index (Cabrerizo et al. 2010) and wu-index (Wu 2010) etc.

According to the nature of problem there is no benchmark available to evaluate these techniques and to discover the optimal parameter to rank the authors. In general, whenever a new technique is proposed in the literature, its evaluation is made on either imaginary case scenarios (Alonso et al. 2009; Kosmulski 2007) or on just small and different datasets (Cabrerizo et al. 2010; Bornmann et al. 2011). Furthermore, these techniques are evaluated on different datasets, making it difficult to grasp the contribution and importance of each technique over the others. To identify which technique performs better for ranking of authors, there should be a comprehensive evaluation of these techniques on the wide-ranging dataset from the same domain.

In this paper, we intend to investigate the role of indices based on citation intensity (h-index, a-index, f-index, t-index, Maxprod index, q2-index, Wu-index, tapered h-index) and publication age (ar-index, contemporary h-index, raw h-rate) on a comprehensive data set belonging to the domain of Civil Engineering. The purpose of this research is the evaluation of these indices and identification of the best performing indices for ranking of authors. For evaluation purposes, awards given by four prestigious societies in Civil Engineering field are considered as a benchmark, which include ASCE (American Society of Civil Engineering), CSCE (Canadian Society of Civil Engineering), ACI (American Concrete Institute) and ICE (Institute of Civil Engineering).

Literature review

The analysis of scientific production of authors plays pivotal role in different aspects of research, judgmental decisions, accepting research projects, nominating scientific awards, promoting an individual, allocating tenures to skilled persons and to grant the contracts to the experts. There exist various evaluation parameters upon which the entire research work of an individual author may be ranked. For this purpose, many techniques have been proposed in the literature. One of the earliest proposed parameter for scientist’s assessment is publication count. The number of publications is harnessed for evaluation of authors (Balog et al. 2006) but it does not cover the impact and quality analysis of one’s work (Cameron et al. 2007). To overcome this deficiency, the citation count was introduced (Bogers and Van 2008). But there are some limitations with citation count too e.g. newly published papers require a lot of time to acquire citations. Therefore, novice authors do not get a chance to be ranked in the top position because of having low number of citations even if they have large number of publication count. Another issue is that sometimes scientists cite the research papers of other authors to criticize them, the count of such citations can’t be deemed as a measure to analyze the quality of a publication.

To overcome the issues pertaining to these parameters, h-index was introduced by Hirsch (2005). H-index gained the popularity in a few years and nowadays the h-index has become the dominant parameter over all other measurements for evaluating research output of authors. H-index has some weaknesses as well, for example h-index is mainly based on long term observations, so the old authors have large h-index than novice authors and it is field specific. In order to overcome the deficiencies of h-index, various modifications of h-index are introduced which are known as variants and extensions (Bornmann et al. 2011).Many studies are presented in the literature in which different indices are applied on different data sets, and a few of those studies are presented in this section.

In 2006, Kosmulski tested the relationship between the h-index and h(2)-index. The data on which the relationship was examined includes 19 professors affiliated with the department of Chemistry in a Poland university (Kosmulski 2006). Results of Kosmulsiki’s study revealed the existence of a strong correlation between the two indices which indicates the results acquired from two indices are same.

In a recent study, Ayaz and Afzal (2016) have evaluated the h-index (Hirsch 2005), g-index (Egghe 2006) and complete-h (Dienes 2015) on a real data set in the field of Mathematics. Their proposed scheme has considered awarding winners from the Mathematics field as a benchmark. According to this study, complete-h, which is defined to complete the definition of h-index having added community factor in h-index definition, performs better than h-index and g-index in bringing award winners to the top rankings (Ayaz and Afzal 2016).

Another study is conducted by Van Raan (2006) to compare the h-index with other indicators. In this study, to examine the correlation between these indicators and h-index, the author has utilized the results of larger evaluation study of 147 research groups of chemistry in Netherland. The author focused on research groups rather than the individual performances and the citations are restricted to only three year window instead of whole life citation count (Van Raan 2006).

Yan et al. (2016) have considered 29 different indicators including 26 h-index variants, calculated correlation among the indices and with h-index and Wu-index, considering h-index and Wu-index as benchmarks. They found out that with some exceptions, mostly indices which are highly correlated with h-index are less correlated with Wu-index, similarly which are highly correlated with Wu-index have low correlation with h-index. It is also highlighted that those indices which have high correlation show little improvement over h-index or Wu-index and should be abandoned or integrated into these indices (Yan et al. 2016).

In 2007, Jin and colleagues proposed a scheme to utilize a combination of indices for the purpose of authors’ evaluation (Jin et al. 2007; Liu and Rousseau 2007). The scheme utilizes a combination of the h-index and r-index, or combination of the h-index and the ar-index. Both the pairs were used to find an indicator for scientist’s evaluation. In a pair, one index is used as quantitative measure while the other index is used as qualitative measure. According to their findings, pair (h-index, ar-index) is a good indicator for research evaluation.

In 2008, Schreiber performed an empirical study of the g-index in comparison with the h-index, a-index, and r-index for data sets of 26 physicists (Schreiber 2008). In this study, Schreiber analyzed the citation records of 26 physicists of the Institute of Physics at Chemnitz University of Technology. The data sets for analysis were compiled in January 2007 from Web of Science. According to Schreiber, the g-index is more suitable than the h-index to characterize the overall impact of the publications of a scientist.

In another interesting study, Wu and Zhang (2017) emphasized that validity of an index should be checked prior to its usage as an evaluation criteria for different purposes, like promotion, funding allocation or other such decisions. It is asserted that for an index to be valid, it should satisfy basic domination relation. Basic domination relation states, if scholar A’s papers have citations at least as much as citations of the scholar B’s corresponding papers, when sorted in descending order of citation count, then scholar A should not be considered inferior to scholar B. In other words, different indicators or indices values’ for scholar A should never be less than scholar B. It is identified that a-index, m-index, e-index and q2-index violate the basic domination relation so, these indices should not be used independently.

In the above paragraphs, we have discussed different studies in which the indices are used to evaluate/rank the authors. Following are the observations derived from the above discussions.

The indices which are discussed above in literature are mostly evaluated for hypothetical case scenarios to prove their accuracy. Secondly, these indices are applied on different and very small data sets. There is need to evaluate these indices on a comprehensive and large data set of a specific domain. Thirdly, there exist no standard benchmark to evaluate these indices. A benchmark data set is required to evaluate these indices.

Methodology

The research questions focused in this study are:

  1. Q1

    Is there any correlation between the ranked lists achieved from the previously known author ranking indices?

  2. Q2

    Whether the international prestigious awardees lie in the top ranking obtained from these indices?

  3. Q3

    Which index contributes the most to bring the international prestigious awardees in the list of top authors?

  4. Q4

    Which award honouring society (ASCE, CSCE, ICE, and ACI) depend upon these indices?

The methodology of proposed solution to answer these questions is discussed in this section and the results are discussed in the next section. The architecture diagram of our methodology is given in Fig. 1.

Fig. 1
figure 1

Architecture diagram of methodology

Domain selection

For the evaluation of indices, we have selected the Civil Engineering field as it is one of the most important fields of science consuming over 40% amount of world’s matter (mass and energy) and still not well explored for Scientometrics studies (Czarnecki et al. 2013).

Taxonomy building

Civil Engineering is a diverse and vast field of knowledge. To collect comprehensive data set for evaluation of indices, topic classification for Civil Engineering was acquired. The certified classification of Civil Engineering is CEDBFootnote 1 (Civil Engineering Data Base) which is the effort of ASCE (American Society of Civil Engineering). ACSE is a well-known society of Civil Engineering and publishes 37 journals in this field. The CEDB is used by all these journals of Civil Engineering, these journals request the authors to use the subject heading of CEDB. It helps in finding the items of potential interest to users. Every year the database of this classification is updated. We have started to collect data in September 2015, so we have used CEDB 2015 classification.

Search engine

To collect the data of authors, publications and citations against these categories of CEDB, we have used Google Scholar. Other resources like Web of Science and Scopus were also available, but the reason to choose Google Scholar is that it has large coverage of publications than other recourses (Falagas et al. 2008; Moed et al. 2016), Google Scholar allows open access to data, whereas other recourses are not openly available and Web of Science has limited coverage than Google Scholar. Another reason to choose Google Scholar is that it also does citation indexing and data relevant to every field of science is comprehensively available. A recent study reveals that the growth of Google Scholar is 13% more than that of Web of Science. The number of citations in Google Scholar increases by approximately 1.5% per month (Harzing 2014). In another study it has been demonstrated that the citation noise (non-Scholar citations, count citations double) in Google Scholar is less than that of Web of Science (Below 2005) and Scopus (Moed et al. 2016). According to Google scholar, they are adding new papers, several times in a week, so we can say that Google Scholar is also updated often. With respect to updating existing Google Scholar record and insertion of new records Harzing reports in a study that Google Scholar is updated every 2–3 months (Harzing 2014). Many researchers work on the comparison of Google Scholar with other sources like Web of Science, they finally conclude that the Google Scholar is the best source for publication, citation and other metadata (Moed et al. 2016; Meho and Yang 2007; De Winter et al. 2014; Noruzi 2005; Handerson 2005). Hence we can say that the Google Scholar is a preferable source to collect the data of publication, citation and authors of these publications. Therefore, to collect the data set of Civil Engineering the terms of classification categories are given to Google Scholar by using a crawler. It collects title of Publication, names of authors, Venue of publication, Citation Count, URL of Paper and year of publication.

Beel and Gipp (2009) have discussed how Google Scholar extracts the records against the given query. They concluded that the top results are more relevant to query term and this relevance is measured by the occurrence of the query in the paper’s title. The results shown by Google Scholar are first ranked according to relevance with query term and then the citations received by searched results. To collect the data we used Google Scholar because it gives top ranked highly cited papers and relevant to the query.

The terms from classification were given to the crawler to collect data of publications and authors. There were many generic terms in the classification which give irrelevant results when given to Google Scholar. The example of this problem is illustrated in the Fig. 2. The top result “construction grammar” is not relevant to construction in Civil Engineering. All such terms were further tuned by domain experts and only those terms were selected which only related to domain of Civil Engineering.

Fig. 2
figure 2

Irrelevant results against “Construction”

With the help of domain expert fine-tuned taxonomy is built that contains 152 terms. The new fine-tuned taxonomy is available as ``Appendix 3'' at http://cdsc-cust.org/Appendices/.

Crawling metadata of authors

For data collection, we used a dedicated crawler which retrieves the metadata of authors from Google Scholar. The metadata contains the title of paper, journal or conference in which it was published, author’s name, the URL of papers and citations of the paper. In order to collect data, the selected subcategories were given to crawler and crawler first identified the papers against given subcategory on Google Scholar and then extracted the records. The crawler extracted top 600 records from Google Scholar, the reason to select only top 600 records is that the Google Scholar gives irrelevant records beyond 600 records. The records are further saved in a database and the database was maintained in SQL server. It must be noted here that for some queries, the Google Scholar returns results below 600. In such cases, whatever number is returned by Google Scholar is saved into the database.

Table 1 shows total publications and authors collected through dedicated crawler.

Table 1 Data set before cleaning

Data cleaning

Many authors (Deng et al. 2012; Harzing 2014; Jacsó 2008; Ayaz and Afzal 2016) who collected data from Google Scholar pointed that the collected data should be cleaned as it has noise. Therefore, the data was further cleaned and verified, this step of cleaning was performed in two cases. In first case it is ensured whether the data is relevant to Civil Engineering field. In second case it is ensured that the authors are disambiguated (Rosenstreich and Wooliscroft 2009; Aguillo 2011).

To verify data in first case further three steps are followed:

  1. 1.

    Removal of crawled records which have invalid characters in the title ($, £, %, *, &, ?, /).

  2. 2.

    Verification of papers whether they belong to journals or conferences of Civil Engineering.

  3. 3.

    Manual verification of the remaining results.

After performing these steps, some publications were removed and final publications have been shown in Table 2.

Table 2 Publications data set after verification

In second step, the authors are disambiguated, for this purpose; we checked the duplication of author names. To perform this task first we find the second name shared with other authors. After this we have analysed the first names and checked whether the first names were distinct, if both first name and second name of more than one author were same, it was considered as duplication of authors. Further, we manually checked those authors by browsing their profile on home pages.

In second step, author disambiguation is performed. In data set the total number of authors were 36,921 and among those authors, 17,589 authors shared the same last name, those 17,589 authors further needed to be disambiguated. By verification, we found that 4130 distinct names were shared by 17,589 authors. This means that there were 4130 authors having a total of 17,589 variations. In 4130 names, some names were shared by 100 plus authors and some names were shared only by 2 authors.

After the data collection the numbers of authors in the dataset were 36,921. There were two cases in name variations, these two cases are highlighted in the Fig. 3. In case 1 the first and last names of authors are same, in case 2 the last names of authors are same but first names are different. We verified such type of record to ensure the author disambiguation and we handled both cases. We accomplished this task manually.

Fig. 3
figure 3

Show the cases of name variations

In case 1 the publication of each author among 4130 authors were checked manually to ensure that either the authors belong to different name or same name. Interestingly there was not any author having same first and last name. For case 2, we again manually visited the home profiles of authors and checked their publications to ensure either they are different authors or same. In this process 88 variations were detected as duplications of 45 authors, it means that 45 authors shared same last name with different first names and the duplications were removed from database. After all these verifications we obtained the dataset of 36,876 authors and 2,184,638 total citations against 152 subcategories or terms of CEDB. Table 3 shows number of authors before and after verification.

Table 3 Authors data after verification

Author ranking by different indices

After data collection and cleaning, we calculated 11 indices on data, ``Appendix 1'' shows calculation of these 11 indices. After calculation of indices we have separately ranked the authors according to each index. After this step, we get 11 distinct rankings of authors which are further evaluated based on four research questions postulated in this study.

Award benchmark

To evaluate the indices, we have considered international prestigious awardees given in the field of Civil Engineering as our benchmark. The data of awardees is collected from different societies which include ICE, ASCE, CSCE and ACI, these societies bestow awards to competent authors. List of awards is shown in Table 4 and data of awardees is available as ``Appendix 2'' at http://cdsc-cust.org/Appendices/.

Table 4 List of awards

Evaluation

After the ranked lists have been calculated, the evaluation of four research questions has been addressed in this section.

Correlation between the indices

The first research question is to find whether there is correlation between ranked lists obtained by these indices. The purpose of finding correlation is just to check as to what extent these indices are similar to each other. Spearman’s rank correlation coefficient is used keeping in view the rank nature of our results (Corder and Foreman 2009).

Awardees’ trend in ranked lists

In this step, we evaluated top ranked authors and focused on second question and third question. The second question is whether the international prestigious awardees lie in top ranked authors. Third question is which index contributes more to bring awardees in rank list. For both questions, we checked the percentage of occurrences of awardees in top 10% of ranked lists then 1–10%, 10–20%, 20–30% up to 90–100%. Moreover, the occurrences of awardees are checked in top 100, 500 and 1000 ranked authors.

To find out the percentage of occurrence of awardees in ranked lists, first we have calculated all the indices for all the authors. After that, we have ranked authors separately, according to each index. We have identified/marked the position of award winners in these ranked lists. Then it was calculated, that how many award winners lie in top 10% of each ranked list. Finally we have calculated that out of total award winners, which were found in our data set, what percentage of award winners lie in top 10% of each ranked list. Similarly for 1–10%, 20–30% and so on. Let us explain with an example, let us suppose there are 50 authors in data set, out of which 10 are award winners. We sort the authors according to h-index value, from high to low. We extract top 10% records from data set, it would comprise of 5 authors’ record which are on top, on the basis of value of h-index. Suppose in top 10% we have 2 occurrences of award winners. Hence in this case 20% of award winners lie in top 10% of authors ranked on the basis of their h-index value.

Dependency of award honoring societies on these parameters

This section focuses on fourth question which is to check dependency of indices on award giving societies (ASCE, CSCE, ICE and ACI). To accomplish this evaluation task we find the occurrence of awardees in top 10% of ranked lists obtained from these indices and according to our observation the index which brings more awardees is dependent on that society.

In previous research question we have identified/marked the position of award winners in these ranked lists and also calculated how many award winners lie in top 10% of each ranked list. We have separated the award winners found in our data set on the basis of their award giving societies, statistics are shown in table. Finally we have calculated that out of total award winners from one society how many lie in top 10% of ranked lists. We have done this separately for all the authors’ lists ranked on the basis of different indices. For example there are 299 award winners from ASCE which are also found in our data set. Let’s suppose out of these 299, 120 award winners are found in the top 10% of author’s ranked list based upon their h-index value. In this case 40% of award winners from ASCE are found in top 10% of the list ranked on the basis of h-index values.

Results and discussion

In this section the answers of our four research questions are discussed.

Correlation between the ranked lists obtained from indices

The purpose of this evaluation is to find the similarity between the ranked lists and to answer the first research question i.e. to compute the correlation among all the indices. To accomplish this task, the correlation of each ranked list with all other ranked lists is computed. The value of correlation greater than zero indicates the extent to which a specific index is positively correlated with other index, the value of correlation less than zero indicates that the indices are negatively correlated, and the value of zero indicates that there exists no correlation among indices. The obtained values of correlation of each index with every other index are presented in Table 5 and are graphically illustrated in Fig. 4. We followed guidelines regarding strength of correlation from (Evans 1996). The strength of correlation for absolute value of r is very weak (0.0–0.19), weak (0.20–0.39), moderate (0.40–0.59), strong (0.60–0.79) and very strong (0.80–1.0). In Fig. 4, only weak and strong correlation between indices is presented whereas detailed values are shown in the Table 5. The Table 5 shows the correlation between the ranked lists acquired from 11 indices and red color is used to highlight the values which show strong or very strong correlation. The correlation of index with itself is 1, and values in blue color represent weak or very weak correlation.

Table 5 Correlation of indices
Fig. 4
figure 4

Strong and weak correlation

The correlation values among different indices are graphically shown in Fig. 4. This figure presents only low and high correlation values. The relationship between those indices having values below 0.4 or having values equal to or above 0.6 are exhibited. The rectangle is used to represent indices, edges represent the strength of a correlation between two indices. The edges in red color stipulate the existence of a strong or very strong correlation between the indices. Similarly, the indices connected through blue edges have weak or very weak correlation. The rankings acquired from 11 indices are not similar because most of the edges are of blue color, which depicts the existence of vast cases of a weak correlation.

Table 6 shows the frequency of low, high and negative correlation among indices. For instance, h-index has strong correlation with 5 indices and weak/low correlation with 5 indices. It is not negatively correlated with any of the indices. Overall the number of indices having low correlation is greater than those indices, which have strong correlation.

Table 6 Strong and weak correlation

From these correlation results, it can be observed that overall weak correlation cases prevail the strong correlation cases. Moreover, raw h-rate and a-index have negative correlation with each other. Hence, it can be stated that in most of the cases the indices are not strongly correlated. The findings from first question further motivates to scrutinize next three research questions.

Awardees trend in author ranked lists

This section presents the evaluation of next two research questions, which are to assess the contribution of indices in bringing award winners in top ranks and to find the occurrence of award winners in top ranks.

To evaluate these two questions, first we examined the occurrence of awardees in top 10% of ranked list for each parameter. Secondly, we examined the occurrence of awardees in 1–10%, 10–20%, 20–30% and up to 90–100% of ranked list calculated by each parameter. The total number of considered awardees are 1060, from which 135 awardees are duplicated, after removing duplication, and remaining 925 are unique awardees. Ideally, all the award winners of this field should be present in the data set, but it only contains 501 awardees. Therefore, we checked occurrences of these 501 awardees in the ranked lists.

Usually, it is assumed that award winners hold a strong research background, their number of citations, publications must be high. Based on this assumption, it was expected that all the award winners should lie in top 10% of authors in lists ranked according to these indices. Whereas, our results portrayed a different picture as shown in Fig. 5 which illustrates the occurrence of these awardees in top 10% of ranked lists acquired from indices.

Fig. 5
figure 5

Percentage of occurrence of awardees in top 10% of ranked lists of each index

From Fig. 5, it can be observed that maximum occurrences of awardees (around 47%) lies in the top 10% of the authors in the list ranked by their Wu-index value, followed by f-index and t-index. Comparatively the better performance of Wu-index could be due to the fact that it pays more attention to the highly cited papers (Egghe 2011; Schreiber 2010; Kosmulski 2013). The indices that focus on highly cited papers are also found better in terms of evaluating economists (Ellison 2013). Wu-index is also known for more accurately reflecting the influence of a scientist’s top papers (Panaretos and Malesios 2009) as well as in assessing the integrated impact of a researcher’s work, especially the most excellent papers (Wildgaard et al. 2014). Furthermore, overall performance of tapered h-index and contemporary h-index is also around 45% and h-index brings 42% awardees. In rest of ranked lists, the occurrences of awardees in top 10% are lower, which are astonishing results as raw h-rate performs 30%.

None of the indices succeeded in bringing even 50% of award winners in top 10% of the ranked lists. We decided to analyze the trend in 100% results, although it is expected that award winners would merely be present below the top 50% of ranked lists, it is quite surprising that the award winners are dispersed from top 10% to the least 10% throughout the experiments.

The Fig. 6 depicts the overall percentage of occurrence of awardees in 1–10%, 10–20%, 20–30% up to 90–100% from the ranked lists. Wu-index, f-index, t-index, hc-index and h T -index perform better in top 10% of the ranked lists, but a-index and raw h-rate have low performance even in top 10%. Other indices also perform better in 1–10% of the ranked list and bring average number of awardees. The occurrence of awardees gradually decreased as we moved in downward of the ranked list. In 20–30%, 30–40% and other lower portions up to 90–100%, the performance of all indices remained same except a-index and raw h-rate.

Fig. 6
figure 6

Occurrence of awardees in partial 10% parts of ranked lists

Dependency of awarding society on indices’ results

This section presents the evaluation of fourth research question i.e. to find out which awarding society depends upon these indices. To answer this question, we examined the occurrences of awardees of each society in top 10% of the ranked list. As we have mentioned that 501 award winners were present in our data set, out of these, 299 were from ASCE, 76 from CSCE, 96 from ACI and 30 award winners from ICE. Figure 7 shows the dependency of awarding societies on these indices. Following observations are formed from the obtained results.

Fig. 7
figure 7

Trend of indices on awarding societies

  1. 1.

    ASCE

    1. a.

      Overall the Wu-index, t-index and f-index outperformed all other indices and around 46% of awardees from ASCE exist in the ranked lists of Wu-index, t-index and f-index.

    2. b.

      H-index, h T -index and hc-index also performed sufficiently by securing 43% approximately.

    3. c.

      The performance of a-index, ar-index and raw h-rate remained lower which is around 30%.

  2. 2.

    CSCE

    1. a.

      The Wu-index brought 41% of the awardees from CSCE. The h T -index, f-index, t-index and hc-index almost performed equally, with around 38, 38, 37 and 36% respectively.

    2. b.

      The a-index, Maxprod index, ar-index, q2-index and raw h-rate didn’t perform well.

  3. 3.

    ACI

    1. a.

      Hc-index brought highest number of awardees from ACI awarding society and performs 59%.

    2. b.

      H T -index, Wu-index, f-index and t-index performed equally, brought around 58% awardees from ACI society.

    3. c.

      The raw h-rate didn’t perform well to bring awardees from ACI society.

    4. d.

      The performance of other indices, a-index, ar-index and q2-index is not very optimal, but better than raw h-rate.

  4. 4.

    ICE

    1. a.

      Hc-index and t-index performed better than other indices with respect to ICE awarding society, as around 43% of the awardees are brought in by these indices.

    2. b.

      AR-index and f-index equally performed well and brought 40% awardees from ICE.

    3. c.

      A-index and q2-index performed inferior to all other indices.

The Fig. 7 shows the trend of awardees of each society.

Evaluation of awardees in top ranked authors

We examined the occurrences of awardees in top 100, 500 and 1000 ranked authors. Figure 8 shows the percentage of awardees’ existence in top ranked authors.

Fig. 8
figure 8

Occurrence of awardees in top ranked authors

Figure 8 shows the occurrences of awardees in top 100, 500 and 1000 ranked authors. In the top 100, the maximum number of awardees, which could be brought are just close to 6% which is an astonishing fact. In top 100 ranked authors, the Wu-index and hc-index equally performed well (6%). The performance of h-index and h T -index remained equal, which is 5%. F-index and t-index also performed well around 4.6%. The performance of a-index and raw h-rate remained lower, the ar-index, Maxprod index and q2-index performed equally but brought only a few of awardees to the top 100. In top 500 awardees, around 17% of the awardees lie in the ranked lists obtained from f-index and t-index. These indices have brought the highest number of awardees to the top 500. A-index, q2-index and raw h-rate brought a small number of awardees, which are around 9%. In top 1000 ranked authors, h T -index, hc-index and Wu-index remained on the top in bringing awardees around 26%, f-index, t-index and h-index performed equally by bringing 25% of the awardees to top 1000.

Conclusion

In this study, the h-index, its variants and extensions based on citation intensity and publication age are evaluated. We have investigated the role of h-index, raw h-rate, a-index, ar-index, contemporary h-index, tapered h-index, q2-index, t-index, f-index, Maxprod index and Wu-index on comprehensive data set collected from the domain of Civil Engineering. To compare the results obtained from these indices, we have used international prestigious awardees in Civil Engineering as benchmark.

The Spearman Rank correlation is calculated among indices to address the first research question. The results revealed that there exists a weak correlation among most of the indices except h-index and f-index as these two indices have obtained the strong value of correlation. The negative correlation is found between raw h-rate and a-index. The positive correlation indicates the extent to which rank lists are similar and negative correlation indicates the extent to which rank list are opposite to each other. Overall, weak correlation among most of the indices implies that 11 rankings are different from each other, these findings motivated us to explore other research questions.

To answer the second and third question i.e. the evaluation of occurrence of awardees in these ranked lists, we compared the awardees with the top 10% ranked authors. Unexpectedly, none of the indices succeeded in bringing even 50% award winners in the top 10% authors. Wu-index, t-index and f-index brought 47% awardees whereas tapered h-index or h T -index and hc-index equally brought around 45% awardees. 31 and 30% awardees are found in top 10%, when ranked according to a-index and raw h-rate respectively. Wu-index, f-index and t-index brought highest number of awardees to the top 10% of the ranked list, it is evident that these indices contribute more to bring awardees at the top.

To answer the fourth question related to the dependency of award giving societies on these indices, we compared the awardees individually from each society with ranked lists. The results depicted that almost 60% of the awardees from ACI awarding society lie in all ranked lists. The existence of more awardees from ACI awarding society shows its dependency on these indices.

Apart from these indices, various other indices have been proposed by researchers’ ranking community that should be evaluated on a comprehensive dataset to signify the potential of each index. In future, we intend to evaluate these remaining indices on a comprehensive data set from different domains.