Introduction

A clear picture of the dynamics of national disciplinary profiles helps identify the strengths and weaknesses in research fields of a nation and is thus key to science policy-making, especially with regards to allocation of funding. Recent years have seen an emerging strand of research in scientometrics focusing on quantitative investigations of disciplinary distributions of national research capacity measured by bibliometric indicators. The approaches of this research strand have mainly been either a comparative study of national research profiles, a time series analysis of the evolutionary patterns of national disciplinary specializations, or a combination of both.

The comparative studies have provided insightful information. King (2004) compares the disciplinary strength and weakness of the G8 nations using disaggregated citation share as the indicator and identifies some marked asymmetries in national disciplinary profiles, i.e., the relative strength in the physical sciences and engineering for Russia and Japan, and in the medical, life, and environmental sciences for the United States and United Kingdom. Hu and Rousseau (2009) assess research performance in selected Western and Asian countries and find that Western countries outperform their counterparts in Asia in most biomedical fields. Yang et al. (2012) evaluate the similarity and balance of disciplinary structures of the G7 countries as opposed to the BRIC nations and assert the G7 countries have more balanced disciplinary structures than the BRIC nations but the latter are catching up. They also find that the G7 countries are specialized in life sciences and the BRIC nations have relative strength in physics, chemistry, mathematics, and engineering.

Evolutionary patterns are investigated at both national and regional levels, with mixed research findings. Glänzel and Schlemmer (2007) analyze research profiles of the EU nations and confirm that while a converging process is in place, the process is slow and EU member countries have maintained their individual peculiarities and preferences. Horlings and van den Besselaar (2013) verify that there has been a process of convergence in both the level and structure of scientific outputs. It is found that convergence is more evident within each of the nine convergence clubs of nations grouped according to their socio-economic and political nature. Radosevic and Yoruk (2014) investigate whether the catching-up/transition regions have converged with the world frontier regions in their research profiles and find an overall pattern of persistence in regional disciplinary structures.

In addition, there have been noteworthy scholarly efforts devoted to the development of methodologies for the study of national disciplinary profiles. Zhou et al. (2012) propose a general framework for describing the degree of inequality within systems and similarity between systems. They suggest that the Gini coefficient and the Herfindahl–Hirschman index be used for measuring diversity within systems and the Salton’s Cosine measure for measuring structural similarity. Bongioanni et al. (2014a, b) propose a different measure of structural similarity—the spin glass model of the physics of complex systems—to analyze national research systems’ disciplinary orientations based on the percentage of publications in various research fields. This model is applied to European countries as well as a sample of 40 nations. Their studies confirm the existence of a converging process in the disciplinary profiles across nations.

While existing literature has revealed important findings, more in-depth investigations are still in demand to reveal insights of structural changes in national research profiles in at least two aspects. First, it is desired that research of this topic keep focusing on the evolutionary patterns of national disciplinary profiles. National research systems are ever changing; their disciplinary profiles can only be better assessed if the changes over time are taken into account. Second, given that the approaches utilized in studies of research profiles are often descriptive in nature, appropriate inferential statistical methods need to be applied to testing the stability and mobility of national disciplinary structures.

This paper contributes to the literature in the two aforementioned aspects by providing an empirical study of the evolutionary patterns of national research systems in terms of their disciplinary profiles. The nations examined cover the 45 top producers of publications around the world during a period spanning two decades from 1996 to 2015. For every nation, the level of specialization for each discipline is measured and the level of dispersion among disciplinary specializations is calculated. Significance of the structural changes is tested through simple regression models. Three related issues are examined: (1) the distribution of disciplinary specializations across nations; (2) the patterns of structural changes in the world’s disciplinary profiles across nations, i.e., whether there has generally been a convergent process in national disciplinary structures towards a unique profile defined as the global disciplinary distribution of publications; and (3) the evolutionary patterns of research profiles in individual nations.

Data and methodological issues

The number of publications from each nation extracted from Scopus is used as the measure of national research outputs. Changes of structural changes in national research profiles are to be analyzed through descriptive tools and inferential statistics such as regression models. However, the research design should be applied with caution, as using publication counts extracted from bibliometric databases as a measure of scientific outputs is subject to biases that may lead to questionable research findings. In this section, relevant reliability and validity issues are discussed and steps taken to reduce the problems are described, including the approach to better meet the assumptions of regression analysis. Methods for interpreting regression results are also elaborated.

The reliability and validity of bibliometric analysis are of central importance in the study of national research system (Nederhof 1988; Tijssen 1992) and relevant issues need to be identified and tackled. First, coverage of the database constrains the quality of data in representing the world scientific outputs, causing reliability issues to the studies that involve comparing across nations and/or disciplines and between different time periods. On one hand, every bibliometric database, even the largest one, has its own selection policies and there is no one single database that covers every valuable piece of the research literature in the world. Issues linked with the balance of subject areas and with languages bias have been identified: (1) some subject areas in social sciences and humanities are less represented; and (2) major bibliometric databases have strong presence of publications from the Anglo Saxon world (van Leeuwen 2004; van Raan 2004, 2014). On the other hand, time-series studies encounter problems with the inconsistence of database coverage. Both the selection policy and the strategic expansion of databases may lead to significant changes in coverage. The level of influence of inclusion and/or exclusion of journals in bibliometric databases differs more or less across subject areas. Second, using publication counts as the unit for quantitative analysis of national research profiles has critical validity concerns. For comparative purposes, an essential assumption is often made—all publications included in the database are of equal values in measuring scientific outputs. In the real world, however, values of papers differ greatly. Furthermore, many quantitative studies employ statistical techniques that require normality of the distributions of variables but the distribution of publication counts is very likely highly skewed (van Raan 2006).

Problems with the reliability and validity issues described above may lead to invalid or partially valid findings and these problems need to be carefully addressed. A common practice in reducing these problems is to use macro-level bibliometric data. Nederhof (1988) argues that comparative studies become more reliable when the unit of study becomes more aggregated. This is because in general a more aggregated unit of study means a larger sample size and a larger sample size reduces uncertainty caused by random errors. This paper follows this common practice of addressing issues related to database coverage in four aspects. First, data for analysis are extracted from a bibliometric database of large scale. The present study uses data from SCImago Journal & Country Rank (SCImago 2007), a free access portal that provides country level scientific indicators developed from the Scopus database. Launched in 2004, with new titles added in each year, Scopus now contains over 60 million records of bibliometric data, including 14,200 journals. Although Scopus does not backfill content for newly added journals, the database does add cited references going back to 1996. Through the Cited Reference Expansion project, Scopus was 99% complete for titles originally included in the database back to 1996 by the end of 2015 (Scopus 2016). Therefore, the post-1995 publication counts from Scopus is more reliable than the pre-1996 counts.

Second, this study sets a relatively higher threshold for sample selection. Changes in coverage from inclusion or exclusion of journals tend to have higher impact on small countries with less scholarly outputs. Characteristics of disciplinary distribution are only best measured if a nation reaches a certain level in its total publications. Inclusion of too many nations may result in distortion of analyses, as relative indicators may become statistically unreliable when a country only has a small number of publications (Schubert et al. 1989). This study sets 100,000 total publications as the threshold for sample selection. Above the threshold are 45 nations representing 96% of the world’s research outputs (see Table 1).

Table 1 Sample nations and number of publications 1996–2015.

Third, in measuring individual nation’s research profile, instead of relying on publication counts for each year, this paper divides the time span of 1996–2015 into three 6-year periods and uses measures for each period as the unit of analysis. The more aggregated data reduce the impact of sudden fluctuations that may happen to individual nations. The three 6-year periods are defined as Period I—1996 to 2001, Period II—2003 to 2008, and Period III—2010 to 2015. There is a 7-year gap between the consecutive periods and a 14-year gap between the starting and the ending periods. In the traditions in public policy studies (Johns and Inwood 2014) as well as in financial management (Brigham and Houston 2016), a long term is defined as a period of more than 10 years and a medium term as 5–10 years. Therefore, comparison between the three periods allow exploration of both medium-term and long-term changes in national disciplinary structures.

Fourth, this paper makes an attempt to decide on the appropriate level of classification system. In the Scopus database, publications are grouped into 4 board areas, 27 subject areas, and 313 specific subject categories. A more detailed classification may provide more insights, but using disaggregated data may cause the concerns of reliability. Meanwhile, if the more aggregated classification is chosen, then it is at the price of sacrificing insightful information in particular disciplines. Therefore, the 27 Scopus subject categories are chosen for analysis of disciplines. Table 2 shows that there is a high level of distributive disparity of publications across the 27 disciplines, ranging from 0.54% (Veterinary) to 16.70% (Medicine). Note that due to double counting, there is a large difference between the number of world total publications in Tables 1 and 2.

Table 2 Distribution of world publications across disciplines: 1996–2015.

It is believed that the aforementioned four steps taken to reduce the problems associated with the reliability issues caused by the database coverage should be effective. Nonetheless, as this study uses Scopus as the sole source of data, the analysis is still subject to the biases inherent to this particular database, although to a much lesser extent.

The validity problems in the study are addressed through the selection of a more appropriate indicator for analyzing national patterns in disciplinary specializations. The process is elaborated in the following discussions on the statistical methods employed in this paper.

To reveal insights of the evolutionary patterns of individual nation’s research profiles, both descriptive tools and inferential statistical techniques are utilized in the present study. As for statistical analysis, research methods are borrowed heavily from the tradition in the studies of sectoral specialization in two related fields: international competitiveness and technological development (see Cantwell 1989; Dalum et al. 1998; Amendola et al. 1998). This borrowing of research methods encounters little obstacle because the quantitative studies of disciplinary profiles are analogous to these two research fields in that they all deal with compositional data. In fact, the three fields share the same measures of specialization with different terminologies. Activity Index (AI), the commonly used measure of disciplinary specialization, is calculated in the same way as the Revealed Comparative Advantages (RCA), the most popular measure of specialization in trade and technology policy. Specifically, the formula for AI in subject field i of country j is:

$$ {\text{AI}}_{{ij}} = \frac{{P_{{ij}} /\sum\nolimits_{i} {P_{{ij}} } }}{{\sum\nolimits_{j} {P_{{ij}} } /\sum\nolimits_{i} {\sum\nolimits_{j} {P_{{ij}} } } }} $$

where P ij represents publications in field i from country j. The numerator is the percentage share of a given field in a nation’s total publications, and the denominator is the percentage share of a given field in the world total publications. Therefore, the AI values compare national disciplinary profiles against the world research profile. For a given discipline of a nation, if AI > 1, the nation is said to be specialized in that discipline and vice versa if AI < 1.

Statistical tests of the stability of the national disciplinary profiles are to be conducted through simple regression models, following the methods utilized by Cantwell (1989) and Dalum et al. (1998). Theoretically, a simply regression analysis makes key assumptions of linear relationship, multivariate normality, no autocorrelation, and homoscedasticity (constant variance) of the errors. The major concerns in this study are multivariate normality and auto-correlation issues.

Using simple regression models with AIs as the variable invokes a risk of violating the assumptions of normality. The value of AIs falls between 0 and infinity with an average of 1 (or very close to 1) and it is likely that the distribution of AIs is skewed. One possible fix of the problem is to use an alternative indicator—the Relative Specialization Index (RSI), a widely adopted indicator in the study of national disciplinary profiles (Aksnes et al. 2014). An RSI is calculated as (AI − 1)/(AI + 1), and takes a value between −1 and +1. For a discipline, if AI > 1, then RSI > 0; if AI < 1, then RSI < 0. For a nation in a given subject discipline, a positive RSI indicates that the nation has a higher-than-average activity, or is specialized in that discipline; a negative RSI means the nation has a lower-than-average activity, and a zero RSI reflects the nation has equal-to-average activity or is completely balanced. As RSI values fall between −1 and +1, distribution of RSIs tends to be symmetric, with better normality than AIs. This notion is supported by studies of trade specialization. Dalum et al. (1998) and Laursen (2015) propose using RSCA (Revealed Symmetric Comparative Advantage) rather than RCA in their regression models and an RSCA is calculated in the same way as an RSI.

As to the autocorrelation issue, the fact that scientific disciplines are interrelated raises concerns about independence between consecutive observations. The effects of interrelatedness between disciplines are complicated and hard to estimate. On one hand, research project in one subject field may trigger related projects in other closely related fields, i.e., in life sciences. On the other hand, competing for research funds between two subject fields may result in an increase of publications in one field and a decrease in another.

To diagnose the possible violations of the assumptions for regression analysis, this paper employs the Shapiro–Wilk test for normality and the Durbin–Watson test for autocorrelation. The purpose of the test is to verify whether RSI is a more appropriate indicator for regression analysis then AI. Hence, the following regression models are to be tested and compared:

$$ {\text{AI}}_{ij}^{{t_{2} }} = \alpha_{i} + \beta_{i} {\text{AI}}_{ij}^{{t_{1} }} + \varepsilon_{ij} $$
(Model I)

and

$$ {\text{RSI}}_{ij}^{{t_{2} }} = \alpha_{i} + \beta_{i} {\text{RSI}}_{ij}^{{t_{1} }} + \varepsilon_{ij} $$
(Model II)

where AI ij stands for Activity Index for discipline i of nation j, RSI ij stands for Relative Specialization Index for discipline i of nation j, and t 1 and t 2 refer to the initial time period and the final time period, respectively. α and β are linear regression parameters and ε is the residual term. For the Shapiro–Wilk test, test values are calculated for each nation and for each period for both AI and RSI. For the Durbin–Watson test, simple regressions are performed for both AI and RSI for the changes over three time-spans: Periods I–II, Periods II–III, and Periods I–III. The results are reported in Table 3 and both tests confirm that RSI is more appropriate than AI. Overall, of problems associated with the no-autocorrelation and normality assumptions, the latter deserves more serious concerns. It shows the normality of the error terms can be rejected for 54 out of 135 cases (45 nations for 3 periods) when using AI, and for 27 out of 135 cases when using RSI. That is, using RSI instead of AI reduces the violation of normality from 40 to 20% of the cases. Using RSI rather than AI in the regression model also reduces the problems of autocorrelation. Results of the Durbin–Watson test indicate that 23 out of 135 regressions when using AI and 18 out of 135 regressions when using RSI violate the no-autocorrelation assumption.

Table 3 The Shapiro–Wilk test for normality and the Durbin–Watson test for autocorrelation

Thus, Model II, with RSI at different periods as the dependent and independent variables, will be used for analysis. Note that Model II is for the test of the stability of research profiles in individual nations. When the subscript j is removed, it becomes:

$$ {\text{RSI}}_{i}^{{t_{2} }} = \alpha_{i} + \beta_{i} {\text{RSI}}_{i}^{{t_{1} }} + \varepsilon_{i} $$
(Model III)

Here, RSI i stands for Relative Specialization Index for discipline i of a nation for all nations in the sample. Since this study covers 45 nations that collectively represent 96% of the world’s total research output, it is proper to use Model III to test the stability of the global disciplinary profiles across nations.

The degree of dispersion of a nation’s disciplinary specialization measure by RSIs corresponds to the extent to which the nation’s research capacity is more concentrated or diffused across disciplines. A high degree of dispersion means strong specialization or a highly polarized disciplinary structure. A lower degree of dispersion indicates a relatively more balanced disciplinary profile. Furthermore, measures of dispersion presented for different time periods help determine whether a nation has experienced a converging, diverging, or a mixed process in reference to the world’s overall disciplinary profile. An increase in variability means a nation has become more specialized or polarized; a decrease is an indication that a nation has become more balanced or converged with the world research profiles.

This study uses Standard Deviation (SD) as the measure of dispersion across RSIs in different disciplines. Although SD is an absolute measure for dispersion, given the fact that RSI is a bounded relative indicator without unit, SD is suitable for comparison. SD has the property of being very sensitive to extreme values or outliers, allowing the analysis to sensitively reflect changes of nations in disciplines with high level of specialization or in disciplines with relatively very low level of outputs.

Also noteworthy is that using SD as the measure links the present analysis to the literature on growth in which the convergence/divergence issues have been thoroughly discussed. The growth literature specifies two kinds of convergence: β-convergence and σ-convergence. According to Sala-i-Martin (1996), β-convergence is the phenomenon that poor economies grow faster than rich ones and σ-convergence results from the decrease of the dispersion of nations’ real per capital GDP levels over time; β-convergence shows the mobility of individual economies and σ-convergence captures the features of distribution of the group in question as a whole. Sala-i-Martin (1996) also demonstrates that a σ-convergence must result in β-convergence, but β-convergence does not necessarily lead to σ-convergence. Apparently the analysis of disciplinary structures in this paper has much in common with σ-convergence. Let σ be the dispersion of the numbers of publications across disciplines, if \( \sigma_{{t_{2} }} < \sigma_{{t_{1} }} \), where t 1 and t 2 refer to the initial and the final time periods, then there is σ-convergence. In this study, σ is the SD.

For statistical testing of structural stability, the β value should be tested against zero for each regression function. If β = 0 cannot be rejected at a certain significance level, then no relationship is found between the initial and ultimate values of RSIs across disciplines. In the cases that there is a linear relationship between the initial and ultimate RSIs, if β is significantly below zero, the ranking of publication activities across disciplines has been deeply changed or even reversed (the case of β = −1), meaning disciplines with lower-than-average activities became higher-than-average and vice versa. If β is significantly above zero, then the null hypothesis of β = 1 should be tested. For interpretation of the regression coefficient, β = 1 corresponds to perfect stability in the profile of a nation’s research specialization; β > 1 means the nation has become more specialized in disciplines where specializations were in place or became even lower in disciplines with previously lower-than-average activities; 0 < β < 1 indicates the nation has decreased in the degree of specialization in disciplines with higher-than-average activities and became higher in previously lower-than-average disciplines in terms of publication activity. Therefore, if the alternative hypothesis of β ≠ 1 is accepted, then the signs and values of \( \hat{\beta } \), the estimated value of β, signal the degree and direction of disciplinary structural changes for a nation revealed by regression. Thus, the magnitude of (\( 1 - \hat{\beta } \)) can be treated as the estimated measure of the “regression effect”, an indicator of the degree of structural change explained by the regression model.

The process of convergence or divergence can be confirmed by comparing the β value against the Pearson correlation coefficient value (R). With reference to Cantwell (1989, p. 30):

$$ (\sigma_{i}^{{t_{2} }} )^{2}{/}(\sigma_{i}^{{t_{1} }} )^{2} = \beta_{i}^{2} /R_{i}^{2} .\quad {\text{Thus}}\;\sigma_{i}^{{t_{2} }} /\sigma_{i}^{{t_{1} }} = \left| {\beta_{i} } \right|/\left| {R_{i} } \right| $$

Since σ measures the level of dispersion, using the estimated values, when \( | {\hat{\beta }}| = | R | \), the dispersion of a given distribution is unchanged; \(| {\hat{\beta }} | > | R | \) indicates an increase in the dispersion, equivalent to a process of divergence or specialization; \( | {\hat{\beta }}| < | R | \) implies a decrease in the dispersion, equivalent to a process of convergence or de-specialization. Furthermore, the value of coefficient of determination (R 2) represents the percentage of the total variation that is explained by the regression model, and the value of (1 − R 2) is the percentage of total variation caused by random errors. Thus, the magnitude of (1 − |R|) can be used as a measure of random variation. Cantwell (1989) terms the value of (1 − |R|) as the measure of the “mobility effect”. If |R| is close to 1, then the relative position of disciplines has been fairly stable. If the |R| value is low, then the linear relationship between RSIs in t 1 and t 2 are relatively weak and the mobility effect is more significant. This is caused by the changes in the ranking of disciplines—some disciplines are moving closer together and others are moving further apart.

Hence, the dispersion in a nation’s disciplinary profiles is decomposed into a “regression effect” (1 − |β|) and a “mobility effect” (1 − |R|). Structural changes in research profiles of different nations result from a variety of combinations of the two types of effects. Since the condition of (1 − |β|) > (1 − |R|) is equivalent to that of |β| < |R|, a convergence process in a nation can be explained as the case when the regression effect outweighs the mobility effect.

Findings and discussions

This section is devoted to the exploration of the characteristics of specialization patterns in national research profiles. The exploration starts with observations on important patterns. Then, the statistically significance of these observations is verified through regression models described in the previous section. Structural changes in national specialization patterns are interpreted through the decomposition of dispersion in the regression effect and the mobility effect. This section proceeds with an in-depth comparative study of the he G7 and the BRIC countries.

Observations on important patterns

Figure 1 shows the SDs of RSIs for each nation during 1996–2015 as an overall picture exhibiting the degrees of balance in national disciplinary profiles. Clearly, the level of balance differs greatly from nation to nation. SDs of RSIs across disciplines range from 0.08 to 0.44. The 45 nations are divided into three groups based on their SD values—Group One: SD < 0.2; Group Two: 0.2 < SD < 0.32; and Group Three: 0.32 < SD < 0.44. 21 nations belong to Group One and all are members of OECD. The top 17 nations in Group One are all developed countries located either in Western Europe or in North America. The top 5 nations are Germany, Austria, Spain, Switzerland, and Italy, all located in continental Europe. 21 nations belong to Group Two, representing the moderate level of balance. Most Asia Pacific economies (such as Japan, Korea, Taiwan, Singapore, Hong Kong, Malaysia, Thailand, India, and New Zealand), and countries of South America (Brazil, Chile and Argentina), Middle East (Iran, Egypt, and Saudi Arabia), and Eastern Europe (Hungary, Poland, Czech Republic, and Romania) belong to this group. 3 countries—Ukraine, Russia, and China—stand out as Group Three, the one with the highest level of disciplinary polarization. History of these counties significantly matters, notably the great influence of former Soviet Union. Observations from Fig. 1 suggest that factors such as a nation’s level of development, geographical location, and history might have significant impacts on its research profiles. However, determinants of whether a nation is more balanced or polarized in its disciplinary profiles are complex and a better understanding of how these factors play their roles requires more systematical studies.

Fig. 1
figure 1

Standard deviations of RSIs for selected nations: average of 1996–2016

The dynamic patterns of the structural changes in national disciplinary profiles are illustrated by Fig. 2 in which the heights of column represent SD values of the three 6-year periods. Comparison of the heights of the three columns for each nation offers very suggestive observations. First, for the vast majority of nations, the heights of the white columns representing Period I are the greatest and the heights of the dark columns representing Period III are the shortest among the three columns. This indicates a prevailing trend of convergence in national disciplinary profiles. Second, differences in height among the three columns for a nation may illustrate the extent to which structural changes have been made. However, while there are observable variations among individual nations in their dynamics of dispersion across disciplines, Fig. 2 along does not provide enough information for explaining these variations. This paper relies heavily on the statistical analysis approach for a more in-depth exploration.

Fig. 2
figure 2

Standard deviations of RSIs for selected nations: comparison of three periods

Statistical analyses

The statistically significance of the observations made from Figs. 1 and 2 is verified using regression models as for whether there has been an overall pattern of convergence in national disciplinary profiles and whether all individual nations have experienced significant structural changes converging towards more balanced research profiles. Findings are discussed through the analysis of the regression effect as opposed to the mobility effect.

The overall pattern of convergence is tested using regression Model III with the dataset that covers RSI values for all sample nations for each of the three periods. Regression analysis is conducted for three pairs of values: RSIs of Period II on Period I, of Period III on Period II, and of Period III on Period I. Regression results are provided in Table 4. For all three regressions, the \( \hat{\beta } \)-values are significantly different from 0 and from 1, meaning there have been statistically significant changes in the global disciplinary structure across nations for both medium and long terms. The \( \hat{\beta } \)-values are all between 0 and 1, indicating that, in general, nations have decreased in the degree of specialization in disciplines with higher-than-average activities and became higher in previously lower-than-average disciplines in terms of publication activity. Further, values of \( \left| {\hat{\beta }} \right|/\left| R \right| \) are all below 1, confirming a process of convergence in national disciplinary profiles. Note that since the RSI values are calculated in reference to the world’s disciplinary distribution of publications, the convergent process implies the changes of national research profiles toward the global structure of disciplines. Also noteworthy here is the cumulative effect. It seems that the long-term change between Period I and Period III results from a sort of additive of the changes during Period I–Period II and Period II–Period III. That the values of \( \hat{\beta } \) are among the lowest for regression of Period III on Period I among the three regressions suggests a continuous convergence process.

Table 4 Regression results for disciplinary profiles in the world across nations (N = 1215)

The detailed regression results for individual nations are displayed in Table 5. For all but two cases (Spain and Malaysia, regression of Period III on Period I), there is a linear relationship between RSIs across subject areas of the initial and ending periods for each nation and for both medium and long terms. The two outliers should have experienced some turbulence in their research profiles—some disciplines lost while others gained specializations, resulting in some tremendous changes in their disciplines’ ranking as well as RSI values.

Table 5 Regression results for national disciplinary profiles (N = 27)

As to whether there have been significant structural changes with individual nations, the findings are mixed. The 45 nations are grouped into various categories based on two dimensions: whether it is about the medium or long term structural change and whether a nation has experienced significant changes (see Table 5). Of the 45 nations, 76% had significant structural changes during the First Medium Term (Periods I–II), 67% during the Second Medium Term (Periods II–III), and 87% during the Long Term (Periods I–III). 5 nations (Australia, Canada, Israel, New Zealand, and the United States) have experienced no significant changes in research profiles both for the medium term and the long term. 4 of the 5 nations have English as their official languages, and among them 2 are neighboring countries in North America and 2 are neighbors in Oceania. In fact, the UK, as another English speaking country, also has fairly stable research profiles. Neither of the two medium terms has seen significant structural changes for the UK. The structural change for UK is only significant for the Long Term, showing the nation has become slightly more polarized with a \( \hat{\beta } \) value of 1.115. The Netherlands only shows significant change for the Second Medium Term when it also became slightly more polarized (\( \hat{\beta } \) = 1.115). Like the UK, Belgium also only shows significant changes for the Long Term. However, changes in Belgium are more dramatic than in the UK as indicated by the \( \hat{\beta } \) value (0.642). Note that the UK, the Netherlands, and Belgium are also geographical neighbors.

While nations with no structural changes or with relatively less changes exhibit some relationships, such as having English as the commonly spoken language and the geographical closeness between nations, those experienced continuous changes in their research profiles do not show a clear pattern. 26 nations had significant structural changes during both the First and the Second Medium Terms (see Table 6). Nations in this category includedeveloped countries with both large size (i.e., Japan and Italy) and relatively small size (i.e., Denmark, Finland, Greece, and Portugal) of economies, historically Newly Industrialized Ecomies (NIEs—Korea, Hong Kong, Singapore and Taiwan), Eastern European countries (such as Czech Republic, Romania and Hungary), current Newly Industrialized Countries (NICs—China, India, Thailand, and Brazil), and Middle East countries (such as Iran, Saudi Arabia, and Turkey).

Table 6 Medium-term and long-term changes in national research profiles

Table 5 also provides information for individual nations regarding whether their structural changes lead to convergence towards the global disciplinary distribution of publications. Using the \(| {\hat{\beta }} |/| R | \) values as indicators, among the 103 cases of which nations experienced significant structural changes (34 for the First Medium Term, 30 for the Second Medium Term, and 39 for the Long Term), only 4 have their \(| {\hat{\beta }} |/| R | \) values above 1 and thus involve a divergent process: Malaysia during the First Medium Term, the Netherlands during the Second Medium Term, and the UK and Belgium during the Long Term. Of the 45 nations, 33 and 29 had their disciplinary structures converged towards to the global distribution during the Frist and the Second Medium Term, respectively, and among them 26 had continuous converging processes through both medium terms. 37 nations experienced the conrging process through the Long Term.

The decomposition of a nation’s disciplinary structural changes into a “regression effect” (1 − β) and a “mobility effect” (1 − R) enables a more in-depth exploration. Based on the regression results, values of the regression effect and the mobility effect for all nations are plotted in Fig. 3 to show medium term changes and in Fig. 4 for long term changes. Correlation analysis is conducted for each term and a significant linear relationship between the two effects is found for the First Medium Term (r = 0.71, t = 6.62, p < 0.001, n = 45), the Second Medium Term (r = 0.72, t = 6.84, p < 0.001, n = 45), and for the Long Term (r = 0.78, t = 8.23, p < 0.001, n = 45). Thus, in general, the larger the extent to which a nation’s disciplinary structural changes are caused by the “mobility effect”, the more those changes are influenced by the “regression effect” as well.

Fig. 3
figure 3

The regression effect versus the mobility effect: medium term changes

Fig. 4
figure 4

The regression effect versus the mobility effect: long term changes

For each chart in Figs. 3 and 4, an estimated regression line is added. In the scatter plots, nations that have experienced higher level of the mobility effect and lower level of the regression effect than the expected levels are located above the estimated regression lines. Figure 3 indicates that Malaysia stood out as an outlier from Periods I–II but returned normal from Periods II–III, and Spain was normal from Periods I–II but stood out as an outlier from Periods II–III. In Fig. 4, both countries are located relatively further away from the regression line and both are in the area of high regression effect and high mobility effect. Note that Figs. 3 and 4 also indicate that changes in the two nations result more from the mobility effect than the expected levels.

In addition, for each chart, a dashed diagonal line connecting the origin and the upper right corner of the square is added. Structural changes in nations below the lines where (1 − |β|) > (1 − |R|) result more from the regression effect than the mobility effect. Since (1 − |β|) > (1 − |R|) is equivalent to |β| < |R| and, thus, |β|/|R| < 1, nations below the diagonal lines have experienced a convergence process. Both Figs. 3 and 4 show that the vast majority of nations are located below the diagonal lines, serving as strong evidence to support the statement of an overall phenomenon of converging process. For the two outlier nations, Fig. 3 indicates that the drastic structural changes in Malays during the First Medium Term led the country to the direction of specialization, while in Spain the harsh changes happened during the Second Medium Term, leading the nation to be more balanced.

Take the long term change case as an example for a close look at the patterns of structural changes. In Fig. 4 where the levels of the regression effect and the mobility effect are labelled with low, moderate, or strong, making the chart into a matrix, and in the High Regression Effect-High Mobility Effect category are the aforementioned two outliers, Malaysia and Spain. Nations close to the origin are categorized as Low Regression Effect and Low Mobility Effect. Among them, 6 nations have been fairly stable (β-values not significantly different from 1) in their disciplinary profiles: United States, Canada, Australia, Netherlands, Israel, and New Zealand. They rank 1st, 7th, 11th, 14th, 23rd, and 36th respectively in terms of number of publications in the world, and together they represent 31.8% of the world’s total publications (see Table 2). That is, close to one third of the total publications are produced by nations that did not experience significant changes in their research profiles.

Figure 4 also reveals that, for nations that have experienced significant structural changes in research profiles, the changes in Hong Kong, Japan, Taiwan, China, Russia, Poland, Czech Republic, and Mexico are mainly attributed to the so-called “regression effect.” The mobility effect in these nations is limited, implying a high level of stability in the ranking of disciplines. On the other extreme, the nations for which the mobility effect contributes more significantly than the expected level to their changes in national research profiles are Brazil, Chile, Austria, Finland, and Saudi Arabia. No clear geographical or socio-economic pattern among these nations is identified.

The value of (1 − |β|) implies the extent to which structural changes have happened in a nation’s research systems. Figure 4 shows that, among those that have undergone higher degree of changes are Portugal, Romania, Italy, Finland, Germany, France, Chile, Brazil, Austria, Thailand, Iran, Korea, Sweden, Switzerland, Turkey, and Denmark. This group of nations includes quite a few advanced Western European nations such as Portugal, Italy, Finland, Germany, France, Austria, Sweden, Switzerland, and Denmark, all of which are among nations with homogeneous research profiles.

The convergence patterns revealed above are similar to findings of previous studies. Glänzel and Schlemmer (2007) confirm a convergence process across EU-15 nations but point out that the process is slow. The 45 nations in this paper cover 14 members of the EU-15. For the 14 members, Fig. 3 shows that, from Periods I to II, which corresponds to the time frame of Glänzel and Schlemmer (2007), the UK and the Netherlands had no significant structural changes and all other nations have moved towards convergence, resulting from a combination of low level of mobility effect and low to moderate level of regression effect. The study by Horlings and van den Besselaar (2013) is a more current investigation of the convergence process. The study uses regression and factor analysis for data from Web of Science for 200 nations in 150 subject areas. It confirms a global process of both β-convergence and σ-convergence. Although differing in data source and research methodology, conclusions reached by the Horlings and van den Besselaar (2013) study are in consistent with the findings of the present paper. Furthermore, studies presented in Bongioanni et al. (2014a, b) aggregate data of the 27 subject categories extracted from the Scopus database into 4 broad areas and use multiple indicators for analysis. Both studies confirm the converging process towards a unique disciplinary profile—either towards the European profile among 27 European countries (Bongioanni et al. 2014b), or towards the global profile among 40 nations around the world (Bongioanni et al. 2014a).

A close look at the G7 and the BRIC countries

While due to limited space a careful scrutiny of each of the 45 nations is not feasible, it makes much sense if the present study pays more attention to the G7 and the BRIC countries. The G7 group consists of the world’s 7 major advanced economies: Canada, France, Germany, Italy, Japan, the United Kingdom, and the United States. They also represent a high level of academic research. The BRIC nations are Brazil, Russia, India, and China. They were at a similar stage of economic development and have enjoyed impressive economic growth during the past decades. The BRIC countries also made great achievements in scientific research. As a side note, measured by the total number of publications from 1996 to 2015, countries belonging to the two groups are all among the top 15. The G7 represents 51.7% of the world’s total publications, and the BRIC nations represent 16.5% (see Table 2). What follows is a comparative investigation between the world’s biggest developed nations and the biggest catching-up nations.

As illustrated in Figs. 1 and 2, the G7 nations have much lower level of dispersion among RSIs across disciplines than the BRIC countries, with Japan and Brazil as an exception in each group. Statistical analysis shows that during the long term from Period I to Period III there has been no significant structural changes for the United States and Canada, and the UK has become slightly more specialized. Measured by the regression effect (1 − |β|), structural changes in the three continental Western European nations are harsh than in three BRIC nations—China, Russia, and India. Structural changes in Japan are much less drastic than those of the three Western European nations of the G7, and changes in Brazil are more significant than the changes in the other three members of the BRICs.

To help analyze evolutionary patterns of structural changes in individual countries, scatter plots are constructed for each of the 11 countries with RSI of Period I as the X-axis and RSIs of Period III as the Y-axis (Fig. 5). Overall, nations with data points more close to the origin have higher level of balance among disciplines than those with data points showing a higher level of variability. For data points located in the four quadrants in each chart, quadrants I and III represents the regression effect and quadrants II and IV represents the mobility effect. Specifically, data points in quadrants II represents disciplines in which a nation did not specialize during 1996–2001 and become specialized during 2010–2015, and data pints in quadrant IV represents disciplines in which a nation specialized during 1996–2001 and lost specialty during 2010–2015. If a nation has more data pints in quadrants II and/or IV, then it has more changes in ranking of disciplines.

Fig. 5
figure 5

Long term structural changes in national research profile: G7 versus BRICs

Charts for individual nations in Fig. 5 reveal some patterns of differences and similarities among the G7 and BRIC countries. France, Germany, and Italy form the group with the highest level of balance across disciplines. This group is also characterized with the highest level of changes of ranking of disciplines. This leads to the observation that the homogeneity of disciplinary structure makes it easier for disciplines to switch their relative positions in the national research profile. This perception is exemplified in the group of Japan, Russia, India, and China, all of which are with high level of variability of RSIs. Structural changes in these 4 nations to a large extent result from the regression effect and the role of mobility effect has been very limited. Canada, the UK, and the USA are similar in that they all possess very stable and balanced disciplinary structures. Brazil does not belong to any of the three groups as it has a level of variability higher than G7 nations (except Japan) and lower than other BRIC nations and Japan. However, like the group of France, Germany, and Italy, structural changes in Brazil result as much from the regression effect as from the mobility effect.

There is also worth in analyzing how RSIs in each nation have evolved over time. Table 7 lists disciplines that remain in the top 10 most specialized disciplines for all the three periods for the G7 and BRIC nations. Some disciplinary structural similarities among certain nation groups are revealed. Namely, the United States, United Kingdom, and Canada share similar specializations in Psychology, Nursing, Arts and Humanities, Social Sciences, and Health Professions; China, Japan, and Russia in Engineering, Energy, Materials Science, Chemical Engineering, Chemistry, and Physics and Astronomy; Germany, France, Italy, and Russia in Mathematics and Physics and Astronomy. Note that patterns revealed in Table 7 should be interpreted with cautions as levels of data coverage in the Scopus database are not evenly distributed across research areas. Research in arts and humanities and social sciences is often more nationally, regionally, and locally engaged and published in languages other than English. These fields are underrepresented in large bibliometric databases, namely Scopus and Web of Science, due to the strong bias towards English-language journals in their coverage (Mongeon and Paul-Hus 2016).

Table 7 Stability of specialized disciplines in the G7 and the BRIC countries

In terms of disciplinary specialization, Yang et al. (2012), using data from Web of Science, find that the G7 is life science dominated and the BRIC members specialize in mathematics–physics–chemistry–engineering. However, this present study reveals two patterns within the G7 group (except Japan)—the English speaking nations (USA, UK, and Canada) versus the continental European members (Germany, France, and Italy) plus Russia. It is also found that China, Japan, and Russia share a high level of similarity in their disciplinary structures.

The present study is also closely related to the framework of paradigmatic patterns proposed by Glänzel (2000). This framework identifies four types of models: (1) the Western model: developed Western countries, dominated by clinical medicine and biomedical research; (2) the former socialist model: economies in transition and China, specialized in physics and chemistry; (3) the bioenvironmental model: developing and natural-resource-oriented countries, focusing mainly on biology and earth and space sciences; and (4) the Japanese model: Japan and NIEs, predominant in engineering and chemistry. Findings of the present paper do not fit into this framework very well due to the different classification systems employed. The study by Glänzel (2000) excludes social sciences and arts & humanities, while these two subject areas are included in the present paper. It is also argued that the Western countries differ greatly in their specialized areas, as shown above for the G7 nations.

Concluding remarks and policy implications

This paper provides an empirical study examining the evolutionary patterns of national research systems using a dataset that covers publications of 45 nations across 27 disciplines. Possible reliability and validity concerns associated with the data source and with the research model are addressed and steps have been taken to improve the validity and reliability of the study.

Overall, a converging process is confirmed in the disciplinary structure of worldwide scientific outputs. The time span between 1996 and 2015 has seen a continuous process of convergence towards a unique global profile. It also shows that all of the nations with low level of dispersion or with homogenous disciplinary structures are developed countries. Meanwhile, regression analysis for individual nations reveals that nations differ greatly in their evolutionary patterns of disciplinary structural changes over time. There are nations that have experienced no significant structural changes and nations that have undergone high level changes, while the majority of nations have gone thorough moderate or low degrees of changes. However, there is no clear pattern showing either the level of development of nations or their geographical locations playing a decisive role in the changes of structural specializations.

Causes of changes in disciplinary profiles result from a combination of the regression effect and the mobility effect. In some nations such as Spain and Malaysia, salient alterations in their rankings of disciplines have been the main sources of structural changes in their research profiles, while others such as Japan and China have been very stable in their rankings of disciplines and the evolutions of their national research profiles have shown a continuous process of de-specialization.

Careful scrutiny of the G7 and the BRIC countries reveals some important features. First, France, Germany, and Italy form the group of homogeneity characterized by the highest level of changes of ranking of disciplines. They all specialize in Mathematics and Physics and Astronomy. Second, the group that includes Japan, Russia, and China is of heterogeneity in the variability of RSIs. National structural changes result from the regression effect and the role of mobility effect has been very limited. The three nations specialize in Engineering, Energy, Materials Science, Chemical Engineering, Chemistry, and Physics and Astronomy. Third, Canada, the UK, and the USA have very stable and balanced disciplinary structures. They specialize in Psychology, Nursing, Arts and Humanities, Social Sciences, and Health Professions.

The above findings have important implications in science policy. First, as disciplinary profiles differ greatly across nations, there should be no single pattern that fits all nations, and a ‘one size fits all’ research policy that ignores national differences may not lead to the desired results. Policy-makers must be sensitive to how their respective nations differ from others; particularly, they must possess a good understanding of patterns of structural changes as well as the specialization in their national research profiles. The analytical approaches described in this paper, namely the regression models and the decomposition of the variability of structural changes in national research profiles into the regression effect and the mobility effect, may help individual nations better understand their evolutionary patterns of disciplinary profiles.

Second, for any nation, a notable policy question that needs to be answered is whether the nation will better off if it strives to be more homogenous in its disciplinary structures or if it specializes in just a few disciplines. Unfortunately, the present study does not address this issue. In fact, despite the existence of a general process of convergence across national research profiles, nations diverge in their evolutionary patterns. Evidence from the comparison between the G7 and the BRIC nations shows that all advanced economies but Japan maintain their research profiles with high level of balance, and the BRIC nations, with the exception of Brazil, retain relative stability in the ranking of their disciplines and move towards a more balanced structure in an incremental way. However, it could be dangerous if this pattern is generalized to other nations for their strategies in scientific research as each nation is unique to some extent and the strategic planning should be made on the case by case basis.

Third, no matter whether a nation sets its development strategy as having a more balanced or more specialized in its discipline profile, the nation may still consider having science policies that take advantage of the overall convergence in national disciplinary profiles. The convergence phenomenon at least results from two contributing factors: one is the indigenous improvement in disciplines with previously lower-than-average publication activities, and the other relates to the international transfer of knowledge through collaborations and mobility of talents. The latter deserves attention. Scientific knowledge, unlike technology know-how, is often regarded as a type of public goods that knows no national borders. This feature eases the process of international transfer of scientific knowledge. Moreover, the movement of globalization speeds up the international exchanges of ideas, goods, and people. Therefore, implementing policies that promote international collaboration and mobility of ideas and talented people not only can facilitate catching-up in disciplines with lower-than-average publication activities but also may strengthen those already specialized fields. Even if a nations aims to become more specialized, it could still benefit from the converging process.

The foregoing interrelated points lead to an important research question: what are the major determinants of a nation’s research profile? That is, what leads a nation to be more balanced or to be more specialized and what contributes to the structural changes in the nation’s research profile? National differences in research profiles may be caused by direction of persistent investment (i.e., investment in certain fields of R&D), by factor endowments (i.e., big country versus small country), by level of development (developed country versus developing country), by geographical location (continental location, whether adjacent to a large economy, etc.), and by cumulative mastery of knowledge through national culture and traditions, among other factors. It is not clear how the abovementioned factors play their respective roles, nor the relative importance of each factor. This is a promising direction for future research.

It should be emphasized that, although steps have been taken to alleviate problems of reliability and validity, interpretations of the findings of the present study still need to be made with caution. The analysis still suffers to some extent from limitations inherited from the use of the bibliometric databases. In addition, although normality issues have been greatly reduced, the assumption is still not met with 20% of the cases.