Introduction

What characteristics of research systems influence scientific output? That question has a long tradition in the economics of science. Numerous works throughout the last decades have shown that bibliometric data can be used to answer this type of question and to compare the performance and effectiveness of research systems across countries (King, 2004; May, 1997). Garfield (1979), one of the founders of bibliometrics, held the opinion that if the scientific literature reflects research activities, then a comprehensive multidisciplinary reference index provides an appropriate overview of them.

Previous research using cross-national and historical comparison has shown that investments in research and development (R&D) do not solely determine the differential growth of scientific productivity and visibility (Powell & Dusdal, 2017). Many other factors affect bibliometric research impact besides the level of funding such as size (Allik et al., 2020), hyperauthorship (mass authorship) (Hirv, 2019; Thelwall, 2020), collaboration (Adams, 2012; Wagner et al., 2018) and the way of financing (Auranen & Nieminen, 2010; Dialogic/Empirica, 2014; Jonkers & Zacharewicz, 2015). These factors have not been studied simultaneously, but rather incoherently as separate issues. This article takes an initial step toward integration by using Lepori’s (2011) theoretical structure of public research funding to bring relevant characteristics together systematically. In this scheme, public funding is viewed as a multi-level, multi-actor system in which stable patterns are mostly formed through collective interaction among actors and where coordination among actors is critical to system performance. The scheme gives an overview of public research funding, distinguishing between four organizational layers—the policy layer, funding agencies, performing organizations and research groups/individual researchers—as well as two main allocation methods, core funding to research organizations and project funding to research groups.

Despite the complexities of research funding systems and the vast differences between them, they fall under the theoretical structure of public research funding proposed by Lepori (2011). In order to contribute to research on scientific productivity, we investigate simultaneously how returns to scale, level of funding, way of financing and forms of collaboration (e.g., international and industry–university collaboration) affect research impact in terms of the number of research articles published and citations received. Based on our adaptation of Lepori’s (2011) work on the research system’s theoretical structure, we have three research hypotheses on what could affect countries’ scientific impact. These hypotheses could bring a new perspective to debates on the science production function’s returns to scale and effects of competitive funding as well as explain why some small countries perform significantly better than expected based on their R&D spending.

Firstly, we propose that the size of the research system affects countries’ citation rates positively. In larger research systems, there is more room for specialization and diversification, which are associated with higher research impact (Herron et al., 2016; Parker et al., 2013). We argue that digital innovations have significantly reduced transaction costs of collaboration (Goldfarb & Tucker, 2019) and increased the role of specialization and diversification. Since co-authorship has increased from 1900 onward (Lariviere et al., 2015), it is also reasonable to assume that increasing returns to scale, described by total number of researchers in the system, is associated with higher scientific impact.

Secondly, we propose that transnationally coordinated research has a positive effect on the country’s citation rates per article. This question has increased in relevance as the relationship between public funding and scientific output has become more complicated than previously believed. Scholars have recently found that research done in consortia has significantly higher citation impact than the world average (Thelwall, 2020), and small countries participate proportionally more in consortia than larger ones (Must, 2014). This could explain why several small nations perform above their weight in impact relative to spending (ibid).

Thirdly, we propose that the financing model for research organizations and individuals (e.g., core versus project-based funding) has a significant impact on the number of research articles published and citations received. Many have noted that countries differ considerably in terms of the efficiency of turning financial inputs into bibliometrically measurable outputs (Guan & Zuo, 2014; Leydesdorff & Wagner, 2009; Rousseau & Rousseau, 1997). We expect the share of project-based funding to increase the number of citations per article but also to decrease the number of articles published as a consequence of higher transaction costs.

In this study, we first give an overview of the literature on factors affecting national scientific impact. Secondly, we construct a regression model to test the proposed hypotheses. Relevant metrics are retrieved from Clarivate Analytics, the European Commission and Eurostat. Based on the results, we discuss the strengths and weaknesses of the national research systems of European countries.

General scheme of the research system

In the early 1960s, Nelson (1959) and Arrow (1962) laid the foundation for the economics of science. These two works accentuate that scientific non-rivalry and excludability—that its use by someone does not reduce its availability to others—make science a public good and prevent firms from obtaining exclusive benefits through investment. As a result, most of this public good is produced in publicly funded basic research, which typically leads to peer-reviewed publications as an output (Sugimoto & Larivière, 2018). Garfield (1979) developed the publishing and referencing concept as a method to evaluate research, thereby creating the science of bibliometrics. The computer age has made the use of bibliometrics easy; however, simplified use of it has raised legitimate concerns among its critics that it mechanizes science evaluation and overlooks intrinsic value being evaluated (Bornmann & Haunschild, 2017).

Determinants of the scientific productivity of countries described in this study are based on our adaptation of Lepori’s (2011) work on the theoretical structure of research systems. The theoretical framework of how different aspects interact with each other and can influence countries’ scientific impact is outlined in Fig. 1. Our adaptation gives an overview of the public research system by distinguishing among five layers—scale layer, policy layer, funding agencies, performing organizations, and research groups/individual researchers—and between two allocation methods (core and project-based funding) for how resources are brought down to research groups and individuals. We made three substantial changes in comparison to Lepori’s (2011) initial scheme. Firstly, we put research impact to the top of the scheme which was previously left implicit. Secondly, we added the scale layer to reflect possible size advantages of larger research systems due to increasing returns to scale. Thirdly, we added the possibility to participate in transnationally coordinated research to the policy layer.

Fig. 1
figure 1

Source: author’s adaptation based on Lepori’s (2011) model

An overall view of public research systems.

Based on the scheme elaborated in Fig. 1, we investigate how financial resources are brought down from the policy layer to the layer of research groups and individuals (core vs project-based funding) and the overall effect on research systems at the country level. We also take into consideration the collaboration activities of research groups and individuals such as university–industry collaborations (UIC) and collaborations with scientists from abroad.

Determinants of the scientific productivity of countries

In this article, we focus only on factors brought out in our adaptation of Lepori’s (2011) scheme of research systems and leave factors such as language barriers (Van Leeuwen et al., 2001), political stances (Gantman, 2012), and history (Allik et al., 2020) aside. We assume these factors have no effect because of the current homogeneity of Europe. Previous empirical findings on how factors mentioned in our adaptation can affect a country’s scientific impact are outlined in the following.

Returns to scale

An awareness of the production function’s returns to scale is crucially important in order to allocate resources efficiently. Schumpeter (1942) argues that only large firms could induce technological change because small firms are incapable of allocating the optimal level of resources for R&D. It is stated that in some cases, interactions among researchers, fixed setup costs (ibid; Romer, 2012) or even choice of research topics could lead to a situation where doubling capital and labor more than doubles the output. Although Schumpeter received heavy criticism (Fisher & Temin, 1973), his theory has stayed relevant since.

Recent findings on returns to scale in science production are primarily based on university-level data, where results are mixed. All the possible returns to scale are presented in those studies, including (1) increasing (Koshal & Koshal, 1999; Laband & Lentz, 2003); (2) decreasing (Worthington & Higgs, 2011); and (3) constant returns to scale (Abramo et al., 2012; Bonaccorsi & Daraio, 2005; Clermont et al., 2015).

Results from the macro-level (country) are also mixed. Some state that returns to scale is country-dependent and varies from decreasing to increasing (Guan & Zuo, 2014; Sharma & Thomas, 2008), while others find that smaller countries seem to have an advantage over larger ones in producing highly cited articles (Allik et al., 2020). On the other hand, others reject constant and decreasing returns to scale for publications and citations and point toward only increasing returns to scale (Crespi & Geuna, 2008). A possible explanation for why some studies have shown increasing returns in small countries and decreasing in large ones could come from the methodology of how research publications were counted. The full counting scheme is found to boost the position of more internationally tied countries with a relatively weak science base in relation to scientifically stronger but less internationally connected countries (Gauffriau et al., 2007).

Including international collaboration could bring out the significance of increasing returns to scale on the country level. Recent research has found digital innovations that have significantly reduced the transaction costs of collaboration (Goldfarb & Tucker, 2019) and increased the role of specialization and diversification. In larger economies, there is more room for both, which are associated with higher research impact (Cimini et al., 2014; Herron et al., 2016; Parker et al., 2013).

Expenditure on R&D

Expenditure on R&D is the most significant factor determining the number of articles and their citations. According to Griliches (1979), the scientific output \(Y\) is a function \(F\left( {X, \;K,\; u} \right)\) of conventional inputs of production X, such as labor and capital, the current level of technological knowledge K, and unmeasured determinants of output and productivity. Expenditure on R&D covers both capital and labor used in the production of science. Causality is argued from the observation that funding is required to pay for researchers’ time, other resources needed to conduct research and to publish the results (Shelton, 2020).

The link between financial resources available and scientific impact has been repeatedly demonstrated in recent literature (Docampo & Bessoule, 2019; Shelton, 2020; Vinkler, 2018). It is a general understanding that national GDP has a linear or exponential relationship with scientific publications (Allik, 2013; Docampo & Bessoule, 2019; King, 2004; May, 1997; Mueller, 2016). Although national GDP and expenditure on R&D are highly correlated, they are not the same. It is assumed that R&D expenditures do not depend on real needs because economically well-off countries can afford to spend more money, while developing countries less on scientific research (Vinkler, 2008)

Governmental R&D expenditures (GERD) are found to be directly proportional to the number of scientific articles published (Cimini et al., 2014; Shelton, 2020) and are the main source of the heterogeneity in the scientific output of the EU countries (Pastor & Serrano, 2016). Also, it is noted that more government spending does not necessarily result in more citations because additional government funding is not always absorbed by authors and institutions who produce highly cited research (Leydesdorff et al., 2018).

Large transnationally coordinated research projects

The relationship between public funding and scientific output has become more complicated than previously believed due to the prevalence of large publicly funded, transnationally coordinated R&D projects. For example, in 2010, approximately 3.8 percent of EU Member States’ R&D budgets was directed to these projects (Eurostat, 2010). Transnationally coordinated R&D projects are seen as policy instruments through which states pursue their policy objectives (Reale et al., 2013). Smaller states compensate for their lack of size and specialization by funding and participating in this type of collaboration (ibid). Furthermore, many modern scientific challenges cannot be solved without international networks involving a large number of countries and sometimes even thousands of collaborators (Adams, 2012; Castelvecchi, 2015).

Transnationally coordinated research projects are linked with hyperauthorship (mass-authorship), which is seen as the postmodern pervasion of science (Cronin, 2001), because it is doubtful that all collaborators could have possibly written, edited and approved the final work. The exact threshold which marks the start of hyperauthorship is open to debate. Based on King’s (2012) and Pacchioni’s (2018) suggestions, we interpret articles with more than 50 authors as hyper-authored.

Like hyperauthorship, there is no agreed way to measure the output of transnationally coordinated R&D projects. One method is through a proxy. As we can see in Fig. 2, the total number of hyper-authored articles and investments made to transnationally coordinated research have a very strong correlation (r = 0.88). Therefore, hyper-authored articles are a reasonably good proxy for measuring the scientific output of transnationally coordinated research.

Fig. 2
figure 2

Source: author’s calculations based on Eurostat and Clarivate Analytics’ data (2008–2018)

Hyper-authored articles and investments made in transnationally coordinated research.

Large publishing consortia produce highly cited research, but, at the same time, it is harder to evaluate them bibliometrically. Determining the allocation of credit when publishing in consortia is especially a problem for smaller countries that participate more actively in consortia relative to their size than medium-sized or large countries (Must, 2014). For example, in the case of Estonia, articles with up to 50 authors account for 66 percent of total citations (Hirv, 2019). In addition, an increase in consortia publishing during the decade 2005–2015 led to a significant increase in overall impact (ibid), even though Estonian expenditures on R&D were decreasing during the observed period (Lauk & Allik, 2018).

Additionally, we have to keep in mind that although countries’ participation rates in transnationally coordinated research could be similar, the projects in which countries participate could be different. We expect the effect of transnational coordination to vary between countries.

Core versus project-based funding

There are two main ways to fund research: project-based funding and core funding. According to Eurostat, project funding is part of government budget allocations for R&D (GBARD) provided to a group or individual to carry out an R&D activity with a limited scope, budget and time frame, typically based on the submission of a project proposal specifying the research activities to be carried out (Eurostat, 2020). On the other hand, core (institutional) funding is defined as the part of GBARD granted to institutions with no direct selection of R&D projects or programs to be carried out. Under this type of funding, the receiving institution, not the funding organization, has authority over the R&D projects that are to be carried out (ibid). In the case of core funding, research agencies are skipped, and a state allocates money from the state budget to research organizations, such as universities, for their normal functioning. It is important to note that the divide between project-based and core funding has narrowed since some countries have introduced performance-based funding formulas (Klein, 2015; Salomon-Fernandez, 2014). When project-based and core funding are compared, researchers usually do not focus on absolute amounts of funding but rather on the balance between these two methods.

The share of project-based funding varies notably in European countries (Jongbloed & Lepori, 2015; Reale, 2017), demonstrating significant differences between countries on how they are pursuing research excellence, economic and social impact (Lepori et al., 2007; Van Steen, 2012). As noted by many authors (Guan & Zuo, 2014; Leydesdorff & Wagner, 2009; Rousseau & Rousseau, 1997), countries differ considerably in terms of the efficiency of turning (financial) input into bibliometrically measurable output. Some of these differences could be explained by the variation of funding scheme used.

Competitive funding mechanisms for project-based funding have a number of common features: a call for proposals, a panel assessment, peer review, scoring, award of funding for a limited time and follow-up (OECD, 2018). According to García and Sanz-Menéndez (2005), research allocation, especially in the case of project-based funding, goes through the scientist’s credibility cycle. The scientist’s credibility cycle is a relationship between production, communication and collective evaluation of the results that expands through the process of competing for funding to carry out research. In the setting of project-based funding, government agencies select the agent who best meets their goals. In contrast, research agents decide where to apply for funding and strategically develop their competencies, depending on the available sources of funds (Lepori et al., 2009). The competition takes place concerning quality, while the price is fixed. Research groups need to decide if it is reasonably feasible to achieve the level of quality required and if its cost is worth the price, depending on the other options available and the perceived position of potential competitors. In the evaluation of research proposals, government agencies take into account agent credibility (reputation), which is based on their past scientific achievements (usually number of articles and citations received) (García & Sanz-Menéndez, 2005).

The evidence that excellence in science is linked to competition between researchers is mixed (Auranen & Nieminen, 2010; Dialogic/Empirica, 2014; Jonkers & Zacharewicz, 2015). Although the broader conclusion is that although increased competition in research funding allocation is associated with higher performance, the underlying situation is more complicated and subject to contextual factors (Georghiou, 2013), leaving the question of which is the most effective way to improve research performance open to discussion. The main advantages of project-based funding from the literature contain the following: (1) tendency to increase quality and relevance of research proposals; (2) way to ensure research proposals meet a minimum standard of quality; (3) opportunity to test their ideas among peers; and (4) fair and impartial way of distributing funds (OECD, 2018). On the other hand, project-based funding has got some potential side effects too. For example, with regard to the quality of science, it has been found that competition-based funding leads to the emergence of mainstream, thematically stable but also lower-quality science (Laudel, 2006; Sandstrom & Van den Besselaar, 2018) and also puts considerable costs on the system (Cocos & Lepori, 2020; Gross & Bergstrom, 2019).

We suggest that there are transaction costs in the case of project-based funding in the form of a decreased number of research articles. Competitive project-based funding via funding agencies does guarantee research quality, but it comes with a cost. In addition to higher monetary costs, it also has non-monetary transaction costs, such as time spent on writing research grants and follow-up monitoring. Obtaining funds from any source requires the actors to apply for funds, negotiate and conclude contracts, monitor the project’s implementation and submit reports to the funder (Raudla et al., 2015). The time that goes into these administrative duties is the time that does not go into doing research.

International collaboration

Alongside funding and way of financing, collaboration has been found to play a crucial role in the innovation process (De Solla Price & Beaver, 1966). From 1900 onward, co-authorship and international collaboration have been increasing (Lariviere et al., 2015) and are seen as crucial to success in science (Adams, 2013; Figg et al., 2006; Glanzel, 2001; Glänzel & Schubert, 2005; Moed, 2005). For example, a fraction of papers that are entirely ‘homegrown’ is falling in countries that are at the forefront of science (Wagner & Jonkers, 2017).

International collaboration allows scientists to combine conventional production inputs, such as labor and capital, and exchange technological knowledge internationally. In many instances, international collaboration is necessary because specific technological knowledge or/and capital needed for scientific breakthroughs are difficult or impossible to find within country borders. In the current age, internationally collaborated projects account for at least 20 percent of nationally funded research (Wagner & Jonkers, 2017).

International co-authorship has been very thoroughly researched. On average, publications with international collaboration have higher citation rates than domestic papers (Adams, 2013; Wagner et al., 2018). It is found that openness among advanced science systems is strongly correlated with impact—the more internationally engaged a nation is in terms of co-authorship and researcher mobility, the higher the impact of scientific work (Wagner et al., 2018). Also, it is noted that when the effect of the number of authors is taken into account, the difference in citation impacts between international and domestic publications becomes significantly smaller (Puuska et al., 2013).

University–industry collaboration

In the present era, we cannot overlook university–business collaboration (UIC). National governments support knowledge transfer from universities to the industry as part of their innovation and competitiveness agenda. According to Leydesdorff and Henry (1996), economic, scientific and social development takes place in academia, industry and government. Existing literature on UIC has focused predominantly on how firms can benefit from collaborations with academic institutions, and the effects on academia have been covered to a lesser extent.

University–industry articles are more cited on average than single-university research, indicating that university researchers can often increase their research impact by collaborating with industry researchers (Abramo et al., 2009; Lebeau et al., 2008; Morillo, 2016). We have to keep in mind that this is not the case every time, even though there tend to be positive expectations from collaboration (Bloch et al., 2019). Moreover, papers that mention having received industry funding receive more citations if their results are favorable to industry (Farshad et al., 2013; Kulkarni et al., 2007).

Data

Metrics used in this article are retrieved from Eurostat, European Commission reports and the online InCites database, a research analytics solution provided by Clarivate Analytics (2018). InCites has the advantage of extensive coverage of recognized and widely read scientific journals. The present study uses metrics from 2008 to 2018 and covers 30 countries (28 EU Member States plus Iceland, Norway and Switzerland). For the basis of this analysis, the Essential Science Indicators (ESI) scheme from the InCites database is used. Selected documents include research and review articles from the Science Citation Index Expanded and Social Science Citation Index. The articles’ country identification is based on the article’s address section. If there are two country addresses, then it is read as a product of international collaboration, and an article is designated to all collaborated parties. Unfortunately, due to InCites limitations (Todeschini & Baccini, 2016), fractional counting, as recommended by numerous experts in the field (Gauffriau & Larsen, 2005; Leydesdorff et al., 2018), could not be applied.

In addition to the number of articles, we measure articles’ scientific impact via their citations. Using the number of articles and their citations as an output of production is common practice in the field of scientometrics (Waltman, 2016). Variables used in the study are normalized in order to avoid problems with the difference in scale between countries. For example, GBARD for each country is normalized by per capita. Detailed descriptions of dependent and independent variables used in the econometric models, their sources and normalization are shown in Table 1.

Table 1 Detailed description of variables used

Descriptive statistics of the main variables are shown in Table 2. Missing values in GBARD (3) and RESEARCHERS (13) are dealt with by using linearly interpolated and extrapolated values. In the case of COMPETITION, where a majority of observations (166) are missing, we use country averages for the entire period because core funding is relatively stable (Zacharewicz et al., 2019) and countries are not required to regularly report it to Eurostat.

Table 2 Descriptive statistics

The sample average for CNCI is 132.5 and has a minimum of 56 and a maximum of 232. On average, countries included in the analysis published 1542 articles per million inhabitants. Research system size measured by full-time researchers varies greatly. The sample average is approximately 58,000, but a country with the smallest research system has only around 500 full-time researchers. The proportion of hyper-authored articles is low, but in the case of smaller countries, it can peak as high as 18 percent. In general, countries are more tilted toward institutional funding, as seen in a moderately low share of project-based funding. The data also confirms that international collaboration has become the norm. The mean share of internationally co-authored articles exceeds 50, varying between 32 and 86 percent. As expected, the number of publications with industry collaboration is low, averaging slightly above four.

The correlation matrix of main variables is presented in Table 3. Based on correlations, the best predictor of research impact per article is an international collaboration (r = 0.75). The number of articles has the strongest correlation with GBARD (r = 0.87).

Table 3 Pairwise correlations of variables used

Model

In order to deal with the invariant variable and to get the most information out of the given data, we use a within‑between random effects (REWB) model proposed by Allison (2009) that includes random effects (RE), i.e., between-group and within-group estimates.

The econometric models constructed, where the dependent variable is CNCI or number of articles published, are described as follows:

$$\begin{gathered} Y_{{{\text{it}}}} = a + + \gamma_{1} \,\overline{{\ln \;{\text{RESEARCHERS}}_{{\text{i}}} }} + \beta_{{\text{i}}} \left( {\ln \,{\text{GBARD}}_{{{\text{it}} - 4}} - \overline{{\ln \,{\text{GBARD}}_{{\text{i}}} }} } \right) + \gamma_{2} \,\overline{{\ln \,{\text{GBARD}}_{{\text{i}}} }} + \gamma_{3} \,\overline{{{\text{COMP}}_{{\text{i}}} }} \hfill \\ \,\,\,\,\,\,\,\,\,\,\,\; + \gamma_{4} \,\overline{{{\text{COMP}}_{{\text{i}}}^{2} }} + \beta_{2} \,\left( {{\text{HYPER}}_{{{\text{it}}}} - \overline{{{\text{HYPER}}_{{\text{i}}} }} } \right) + \overline{{{\text{HYPER}}_{{\text{i}}} }} + \beta_{3} \left( {{\text{COLLAB}}_{{{\text{it}}}} - \overline{{{\text{COLLAB}}_{{\text{i}}} }} } \right) \hfill \\ \,\,\,\,\,\,\,\,\,\,\,\; + \gamma_{6} \,\overline{{{\text{COLLAB}}_{{\text{i}}} }} + \beta_{4} \left( {{\text{INDUSRY}}_{{{\text{it}}}} - \overline{{{\text{INDUSRY}}_{{\text{i}}} }} } \right) + \gamma_{7} \,\left( {\overline{{{\text{INDUSRY}}_{{\text{i}}} }} + \delta_{{\text{t}}} + u_{{\text{i}}} + \varepsilon_{{{\text{it}}}} } \right) \hfill \\ \end{gathered}$$

whereby

$$u_{{\text{i}}} \sim N\left( {0,\sigma_{{\text{u}}}^{2} } \right)$$
$$\varepsilon_{{{\text{it}}}} = \left( {0,\sigma_{{\text{u}}}^{2} } \right)$$

\(~\overline{{\ln {\text{RESEARCHERS}}_{{\text{i}}} }} ,\;~\overline{{\ln {\text{GBARD}}_{{\text{i}}} }} ,~\;\overline{{{\text{HYPER}}_{{\text{i}}} }} ,~\;\overline{{~{\text{COLLAB}}_{{{\text{i}}~}} ,{\text{INDUSRY}}_{{\text{i}}} ~}} ,~\;\overline{{{\text{COMP}}_{{\text{i}}} }}\) and \(\overline{{~{\text{COMP}}_{{\text{i}}}^{2} }}\) represent clustered means of a country i. When subtracting the country-specific mean from each of their respective values at time t, those differences become fully independent explanatory variables just like in the fixed effect (FE) model. Therefore, the coefficients of \(\beta_{1}\)\(\beta_{5}\) will have the same interpretation as in the FE model, since they capture within effects. More importantly, the coefficients of \(\gamma_{1}\)\(\gamma_{7}\) represent between effects, as they capture a change in any of the country-specific means, which are time-invariant. Time dummies \(\delta_{{\text{t}}}\) are also included in the models to take into account macroeconomic trends.

All other dependent variables besides the number of full-time researchers \(\left( {{\text{RESEARCHERS}}_{{\text{i}}} } \right)\) and project-based funding (\({\text{COMP}}_{{\text{i}}} )\) are time-varying. The number of researchers did not vary sufficiently to create a within-cluster variable. We interpret the country’s random effect (ui) as a country-specific ‘unknown ability’. Project-based funding square is added to have more flexibility in the case when the effect is nonlinear. Variables used in the models are either measured in percentage points or taken into a logarithmic form to assure the comparability of variables. A logarithmic unit is chosen for its familiarity: in small quantities, a log point is equivalent to a percentage point.

It is important to recognize that there are lags between changes in research funding and outputs (publications), and between outputs and their impact (King, 2004). Therefore, we use a lagged funding variable and assume that a change in expenditures shows an effect in output with a 4-year lag (t–4). Therefore, expenditures made on research in 2008 would affect research output in 2012. The length of the lag was decided based on a panel vector autoregression model (Abrigo & Love, 2016).

Results

The main results are shown in Table 4. We find that project-based funding does not vary within-groups and therefore cannot be used to create a within-group regressor. For interpreting the analysis, we are interested mostly in between-group estimates (BE) because those are more relevant for comparing countries.

Table 4 Within-between regression models

Researchers

Surprisingly, results do not confirm the proposition that the research system size has an effect on research impact.

GBARD

When GBARD increases within-group by one percent, the number of research articles per million inhabitants increases by 0.84 percent. The same number for the between-group estimate is 0.18 percent. A significant difference amongst within and between estimates hints that there is some contextual factor(s) underneath. For example, whole counting could benefit more internationally linked countries with a relatively weaker research base in comparison to scientifically advanced but less internationally connected countries.

A one percent between-group increase in GBARD increases citation impact per article by 0.16 points, confirming that GBARD is the main source of heterogeneity between EU countries’ scientific output (Pastor & Serrano, 2016). For example, when holding other variables constant, the country with the lowest level of funding (Bulgaria) would only have a citation impact of 112. At the same time, the country with the highest level of funding (Luxembourg) would have a citation impact of 151.

Hyperauthorship

A one percentage point within-group increase in the proportion of hyper-authored articles increases the number of articles published per million inhabitants by 1.7 percent. In addition, transnationally coordinated research has a significant positive effect on citation rates. A one percentage point increase in the proportion of transnationally coordinated research articles increases CNCI by 3.19 points, explaining why several small nations perform above their weight in citations relative to spending. For example, holding other variables constant, the country with the highest proportional participation rate in transnationally coordinated research (Cyprus) gains 29 CNCI points over the country with the lowest proportional participation rate (United Kingdom). It is also important to highlight that the effect of transnationally coordinated research on citations is homogenous for countries (i.e., within and between-group estimates do not differ).

Competition

Contrary to our expectations, the share of project-based funding does not affect the number of research articles produced. On the other hand, project-based funding has a nonlinear U-shaped relationship with citation impact, showing that there are two alternative financing strategies to maximize research impact per article, namely having a strong preference toward core or project-based funding. A U-shaped relationship is, to some extent, consistent with previously found mixed results. Firstly, a U-shaped relationship is tricky, and when a model’s specification is not flexible and does not allow nonlinearity, then the actual U-shaped relationship would come off as a non-significant linear relationship. Secondly, in a sample where there are not enough observations with a high share of project-based funding, it can seem that project-based funding is only associated with mainstream, thematically stable, and lower quality science (Laudel, 2006). Thirdly, when there are not enough observations with a low share of project-based funding, then only a positive tendency to increase research quality would emerge.

Collaboration

A one percentage point increase in collaboration decreases the number of articles produced within-group by 1.7 percent, but the change between-groups is insignificant. It shows there are some contextual factors underneath.

A one percentage point increase in international collaboration increases between-group citation impact by 1.31 points, which aligns with the understanding that collaboration is crucial to success in science. When holding other variables constant, the country with the highest level of international collaboration (Luxembourg) would have a citation impact of 165, while the country with the lowest level (Poland) would have a citation impact of 104. We have to keep in mind that larger countries probably have more domestic collaboration that is to some extent if not totally interchangeable with international.

Industry

Also, UIC is an important predictor of the between-group difference in research impact. A one percentage point increase in UIC increases between-group citation impact by 4.07 points. When holding other variables constant, the country with the lowest level of UIC (Romania) would have a citation impact of 129. At the same time, the country with the highest level of UIC (Switzerland) would have a citations impact of 155.

Country level results

The significant factors forming the basis of our discussion of countries’ strengths and weaknesses are presented in Table 5, where countries are ranked based on their CNCI. All linearly interpretable variables are presented as a heatmap with green indicating strengths and red weaknesses. The column describing the importance of the funding scheme is left colorless because of the nonlinearity and is visualized separately. The country’s random effect (ui) is interpreted as a country-specific ‘unknown ability.’

Table 5 Countries’ strengths and weaknesses

We see that countries from the former Eastern Bloc are strongly disadvantaged by low governmental expenditures on R&D. The problem of underfinancing is the largest in Romania, followed by Bulgaria and Latvia. Countries with smaller research systems, such as Cyprus, Estonia, Bulgaria and Iceland, participate far more proportionally in transnationally coordinated research than others. This is the reason why these countries are surprisingly successful in terms of citations when whole counts are used.

Share of project-based funding has a U-shaped relationship with research impact measured in citations. Figure 3 shows the predicted citation impacts of countries when other variables are held constant at their respective means. When policymakers are only interested in citations, then a strong preference for core or project-based funding is preferred. As seen in the figure, only Malta, Italy, Latvia and Estonia are strongly tilted to either extreme.

Fig. 3
figure 3

Predicted effect of project-based funding on CNCI. Note: estimates based on predictive margins from the REWB model

Most microstates collaborate more internationally than large ones. Luxembourg collaborates internationally more than other countries and is the most active in that regard. Luxembourg is followed by Iceland, Cyprus and Malta. Larger countries do not collaborate so much internationally, probably because they can find suitable partners and necessary know-how within their own country.

UIC is especially frequent in rich countries where expenditures on R&D are also high. This is understandable because more prosperous economies are more knowledge-based and more dependent on new knowledge and information (OECD, 2006). This is most seen in Switzerland, Belgium and Scandinavian countries (Denmark, Sweden, Norway, Iceland). The former Eastern Bloc countries are notably lagging in UIC.

As for unknown ability (ui), Greece, Iceland and Romania have a noticeable advantage, while Slovakia has a significant disadvantage. Unfortunately, what is behind these numbers is only a speculation, hence its name. Iceland has a tiny research system, but they have ten out 3000 in Thomson Reuters’ list of the most cited researchers—a share out of all proportion to the country’s population (Iceland Monitor, 2015). A large value for Greece could possibly be explained by strong connections in scientific networks that remain despite the turmoil caused by economic decline (Trachana, 2013). Romania’s large positive value could be caused by scientists who are more influential than expected based on the country’s shortcomings. A large negative value for Slovakia could be caused by brain-drain, which has been noted to be a problem by the chairman of the Slovak Academy of Sciences (Zubașcu, 2020).

Conclusion

This article analyzed the interplay of size, funding and collaboration in determining bibliometric outputs on the country level. Previous studies have addressed these factors incoherently as separate issues. This article has taken an initial step toward integration by using an adaptation of Lepori’s (2011) theoretical structure of public research funding to bring relevant characteristics together systematically. The hypotheses raised add a new perspective on how research system structure affects scientific output and explain why some microstates perform significantly better than expected based on their R&D spending.

The first research proposition—that the science production function has increasing returns to scale—was unsupported by evidence. No association was seen in the case of the number of articles published or citations received per article.

Secondly, as proposed, transnationally coordinated research projects have a positive association with citations. Larger countries invest more in transnationally coordinated research and have more hyper-authored articles. However, they have a much lower percentage of hyper-authored articles as a proportion of total articles. This creates a problem when comparing countries, since smaller ones are more affected by hyperauthorship, causing misleading information on the country’s contribution to the scientific world (Hirv, 2019; Zanotto et al., 2016). As a result, several small nations perform above their weight in impact relative to spending, although their intent for such collaborations remains unclear—is it better science, learning or something else?

Although transnationally coordinated research boosts countries’ impact significantly, policymakers need to keep in mind that most of it is artificial because of problems with credit allocations that occur with mass-authorship. Fractionalization of articles would significantly diminish the overall output of microstates with a higher prevalence of international collaboration, especially with hyperauthorship.

Thirdly, results show that not only the level of funding but also the way of financing is an important factor determining research impact. Surprisingly, project-based funding does not affect the number of research articles published, but, at the same time, has a U-shaped association with citations per article. This highlights that there are two alternative financing strategies to maximize research impact per article—strong preference toward core or project-based funding.

The reason why this relationship is U-shaped in regards to citations per article could be a result of the Matthew effect in financing (Bol et al., 2018), where early successes increase future success chances. High impact research is concentrated in a few premier institutions in countries where the overall research impact is higher than average (Prathap, 2017). In the case of core funding, institutions get a sufficient amount directly from a government, but in the case of a high share of project-based funding, research groups may get the same amount in the end by getting a higher number of research grants. Also, because of path dependence, there is pressure to publish well-cited articles. Scientists may combine funds from different grants to achieve a sufficient level of funding in order to cope with a high level of competitiveness (Laudel, 2006). The pressure to publish is probably less with a balanced strategy because core funding is more stable than project-based.

In addition, science evolution seems to undergo taxonomic change, where the impact is no longer measured by citations but in terms of impact on all sectors of society (Bornmann & Haunschild, 2017). A balanced funding strategy gives a lower impact in terms of citations per article, but its effects could be positive on culture and economics, that we did not cover here.

As the last step of the analysis, we created an overview of the factors in which countries are performing and lagging. As mentioned, the former Eastern Bloc countries are strongly disadvantaged by shortcomings in R&D expenditures. However, as results show, this is not the only reason why they lag. In most cases, they have significantly less UIC. UIC is especially frequent in countries with high living standards where expenditures on R&D are also high, such as Switzerland, Belgium and Scandinavian countries. Although UIC is associated with higher research impact, it seems to go hand in with the economic structure and cannot be forced upon by policy in the short run. Less developed countries will probably catch up in UIC when the economic structure is advanced enough for a more knowledge-based economy. Microstates compensate for their shortcomings by participating more in international collaboration and transnationally coordinated research than larger countries do. When it comes to financing, only Malta, Italy, Latvia and Estonia gain in impact thanks to having a strong preference toward institutional or project-based funding.

A limitation of the bibliometric method used in this study is that it measures innovativeness through the number of articles and citations that show relevance and standing at a current timepoint compared to other countries and overlooks comparisons between time periods. One way to compare innovativeness in different periods is to look at the number of macro-innovations made. Macro-innovations, as defined by Surry (2008), are innovations that impact a broad spectrum of people or processes within an organization, often requiring significant modifications to the organization’s structures and policies. These types of innovations have been declining since the end of the Industrial Revolution (Huebner, 2005; Woodley of Menie, Figueredo & Sarraf, 2019). Therefore, standings based only on the current independent variables may be inadequate in explaining long-term science production.