Introduction

Many studies have attempted to explore the impacts of different variables on research outputs (Rousseau, 2000; Van Raan, 1998). Some researches focus on the relationship between scientific collaboration and research impact (cf., Coccia & Bozeman, 2016; Ronda-Pupo & Katz, 2016a, 2016b). Other papers explore the scaling relationship between the size of the innovation system and citation-based performance (Katz, 2016; Ronda-Pupo, 2017). In this context, funding plays a critical role in the quality and performance of scientific studies, such as on the impact factor of journals (Wang & Shapira, 2015; Lyall et al. 2013). Although some studies do not find any relationship between funding and first citation, there can be a strong link between funding and top percentile citations across different funding sources (Gök et al. 2016). Huang et al. (2006) argue that a significant relationship exists between patent citations and NSF-funded researchers. Wang and Shapira (2015) identify a relationship between research impact and research sponsorship in nanotechnology papers. In particular, they suggest that sponsored research publications have a higher impact than researches without grant sponsors. Quinlan et al. (2008) confirm funded publications receive more citations compared to unfunded research. In computer science, Shen et al. (2016) show a connection between funding and the citations of research. Yan et al., (2018) also show that funded publications in science, technology, engineering, and mathematics (STEM), multi-author, and multi-institution papers tend to receive more citations than single-authored and single-institution researches. Other studies do not show significant differences in terms of scientific impact between funded and unfunded research (Jacob and Lefgren, 2011; Rigby, 2013). Cronin and Shaw, (1999) maintain that the citation of articles is associated with journals and the author's nationality in four information science journals. Hicks and Katz, (2011) argue that in science, research funding should be based on the merit-based evaluation of scholars, but there is unequal distribution of performance in science, and the decision-makers in power face a lot of difficulties in supplying public funds to overcome the level of inequality in research performance. Zhao et al. (2018) argue that research funding is one of the most important resources in the reward system of science. They suggest that funding has a main impact on usage and citations, and funded papers can attract more usage in different fields of research (Morillo, 2020; Pao, 1991). In general, the literature suggests that funded studies are likely to enjoy a higher research impact than other publications.

However, in scientometrics the investigation of the scaling relationship between funding of research and impact of paper in terms of higher citations between different disciplines is hardly known. Stimulated by these fundamental problems, the main goal of the present study is to verify the presence of power-law distribution between the citation-based performance of articles and their funding status in three main disciplines (computer science, medicine and economics) in order to suggest empirical properties that can explain and generalize whenever possible factors associated with the impact of science in society. These topics of ‘the science of science’ can offer a deeper understanding of successful driving factors of science to support best practices of research policy (Coccia, 2018, 2020; Coccia & Wang, 2016; Fortunato et al. 2018).

Theoretical framework

In order to position our analysis in scientific literature, we begin by reviewing existing approaches and studies. A power-law approach implies an inverse exponential relationship between two variables (cf. Barabási & Albert, 1999). Scaling behavior in bibliometrics is based on the Lotka, (1926) rule, which describes the ‘20/80’ distribution of scientific productivity through authors and articles. Cumulative advantage (de Solla-Price, 1976) and success-breeds-success (Merton, 1968; Merton, 1988) are two main concepts associated with many power-laws and scaling behavior. Lotka, (1926) claimed that only a few authors account for around 80 percent of scientific findings. Scholars have attempted to provide evidence that the distributions of other aspects in science (e.g. citations, number of articles in journals, word counts, number of authors for each paper, etc.) can have a power-law distribution (Ye & Rousseau, 2008). For instance, de Solla-Price (1976) argues that the probability of receiving citations for a paper depends on previous citations, such that scientific papers with the highest number of citations tend to be cited more quickly than papers with a small number of citations. In particular, the number of citations as a function of the number of publications tends to produce a heavy-tailed distribution illustrated by power-law. Katz (1999, 2000, 2005) studied scaling relationships between the number of publications and citations across research institutes, various countries, and disciplines (cf. Coccia, 2005a, b, 2019). Katz (2000, 2005) provides strong evidence that the scientific system is characterized by a size-dependent cumulative advantage: a non-linear increase of impact is led by a growth in size. It is assumed that most publications receive the least citations while an only a few papers enjoy a significant impact, becoming the main node of the citation network (cf. van Raan, 2006, 2008). Ronda-Pupo and Katz, (2018) have evaluated the power-law correlation between multi-authorship and citation impact in articles of information science & library science journals, revealing the existence of the Matthew effect among some of the scientific resource elements. Wang and Shapira, (2011) and Gao, (2019) have analyzed the distribution of citations of over-funded publications. Scholars also suggest two models that indicate the scale-invariance property (Katz, 2005; Ronda-Pupo, 2017). The first one is a power-law probability distribution, and the second is the power-law correlations. A scaling relationship exists between two entities (x and y) if they are correlated by a power-law function (Ronda-Pupo & Katz, 2018).

The present paper empirically analyses the power-law distribution of citations of total (funded + unfunded)/funded/unfunded studies and the power-law correlation between the number of total (funded + unfunded)/funded/unfunded papers and their citations impact in different disciplines. The scaling exponent approach will be used in the present study to explore whether any scale-invariant measures and models can aid the decision-making of policymakers to invest in different disciplines in order to predict and increase the diffusion of scientific results and their impact on society.

Materials and methods

Research questions and motivation of the study

The main research questions of this study are:

  • How does research funding affect citations of papers in different disciplines?

  • Do the number of citations in funded, unfunded and total (funded + unfunded) papers follow a power-law distribution?

  • Is the growth of the citations in funded research different from unfunded research?

We confront these questions here by developing an inductive study, which endeavors to explain, whenever possible, common patterns of the distribution of citations of funded and unfunded articles in different disciplines, and how research funding affects the citations and as a consequence the diffusion of scientific research. The development of the study is based on a comparative analysis of three disciplines, given by computer science, medicine, and economics, to analyze similarities and/or differences in the relationship between research funding and citations and to clarify article-level success and failure. We are interested in providing best practices of research funding for universities and at a national level to foster the diffusion of good research (Coccia, 2005a, 2019).

Data, sources and retrieval strategy

  • Disciplines under study here are computer science, medicine and economics. These disciplines fields are representative of critical research fields in different sciences (cf. Coccia & Wang, 2016).

  • The data of the study are all publications of the year 2015 in the fields of computer science, medicine, and economics from the Web of Science database, including inclusive and citation counts for six fixed citation windows (Web of Science, 2021).

We used the publication types of articles and reviews that were published in journals. The string search for retrieving computer science data was WC= "Computer Science", refined by document types: (Articles OR Review). We changed the WC value to "Medicine" and "Economics" to retrieve data related to the fields of medicine and economics. Although a significant number of publications in the field of computer science were conference proceedings, the study did not consider these documents as less than ten percent of them were funded, creating distortion in the comparative analysis done here. For the field of medicine, the number of funded and unfunded papers were roughly equal, providing a representative power. Funding research in social sciences were scarce compared to the two other fields under study here (Coccia et al. 2015). Accordingly, we chose the field of economics in which the portion of funded papers was significant enough to be analyzed in this study.

Measures

The number of papers in computer science, medicine and economics was the explanatory variable taken from the Web of Science database (2021).

Citation-based performance was measured with the number of citations in three different sets: funded articles, unfunded articles and the total (funded and unfunded) articles between disciplines under study (Web of Science, 2021).

This study focused on the papers published in 2015 to consider a time lag of more than five years for assessing a comprehensive amount of citations to support theoretical and empirical analysis.

The theoretical and empirical strategy of the model and the data analysis procedure

The analysis of research questions is based on three main steps.

Firstly, we analyzed the power-law distribution of the citations of total (funded + unfunded)/funded/unfunded articles published from computer science, medicine, and economics.

Secondly, we compared the power-law distribution fitness with other plausible distributions.

Thirdly, we analyzed the power-law correlation between citations and numbers of total (funded + unfunded)/funded/unfunded papers.

Verification of the power-law distribution

In order to test the existence of the power-law distribution of the citations of total (funded + unfunded)/funded/unfunded articles in all documents, we followed the procedure developed by Clauset et al. (2009, p.3, box. 1). At first, we estimated the Xmin value and scaling parameter α of the power-law. According to Clauset et al. (2009), Xmin is the value when the power law begins. Given that the data are discrete, to calculate the Xmin value, we used the formula devised by Clauset et al. (2009):

$$\hat{\alpha } \simeq 1 + n\left[ {\mathop \sum \limits_{{i = 1}}^{n} \ln \frac{{x_{{\text{i}}} }}{{x_{{{\text{min}}}} - \frac{1}{2}}}} \right]^{{ - 1}}$$

To calculate the parameter α, we used the maximum likelihood method. After that, we calculated the goodness of fit between the data and power-law. The aim here is to verify whether the power-law distribution can be a suitable distribution of data for the disciplines under study here. In this step, we used the bootstrap function and ran 2500 Monte-Carlo simulations. We fitted each sample individually to its own power-law model and calculated the Kolmogorov–Smirnov (KS) statistics for each item in relation to its model. The Kolmogorov–Smirnov (KS) statistic quantifies a distance between the sample's empirical distribution function and the reference's cumulative distribution function.

The aim of this step was to determine the p-value. According to Clauset et al. (2009), if the p-value is greater than 0.1 in power-law distribution, the power-law is a suitable distribution for the scatter of data. In addition, we compared the power-law with alternative distributions. The primary purpose was to verify that the power-law distribution was the best fitting model compared to alternative distributions. In this step, we compared the power-law distribution to lognormal, exponential, Poisson, stretched exponential, and power-law with exponential cut-off distributions as alternative options. We also used the Kolgomorov-Smirnov (KS) test to measure the distance between the distributions. Then, we calculated the p-value for each of the distributions and compared them with the p-value of the power-law distribution. According to Clauset et al. (2009), the likelihood ratio (LR) is a determinant factor to choose between two alternative distributions. The distribution with the most negative LR and a significant p-value is a better fit than the power-law distribution. If all the alternative distributions have a positive LR, the power-law distribution can be considered as a proper fit for the model. All of these calculations were performed by Power-Law packages in the R software that was proposed by Gillespie, (2015). We used the Python package as proposed by Alstott et al. (2014) to analyze the stretched exponential and power-law with cut-off distributions. As Clauset et al. (2009) offered, we used key terms to assess the statistical analyses for the power-law distribution of each data set:

  • ‘None’ indicates that the data is probably not power-law distributed. It means that the p-value of the power-law distribution is not significant.

  • ‘Moderate’ indicates that the power-law is a good fit but that there are other plausible alternatives as well. It means that the power-law and other distributions have significant p-value, but the LR of the alternatives is positive.

  • ‘Good’ indicates that the power-law is a good fit and that none of the alternatives considered are plausible. It means that just the power-law p-value is significant among all of the other distributions.

  • ‘Cut-off’ indicates that the power-law with exponential cut-off is favored over the pure power-law. It means that the p-value of the power-law with exponential cut off is significant and has the most negative LR.

Scaling correlation

Finally, we analyzed the power-law correlation between the number of citations and the number of total (funded + unfunded)/funded/unfunded papers in all articles published in computer science, medicine and economics. Citation-based performance (CBP) refers to all citations received by the articles within computer science, medicine and Economics. CBP is a dependent variable, whereas the number of papers published in journals is an explanatory variable. We used funding variables to separate articles and their citations into two categories:

  1. o

    Funded articles refer to articles that are published in journals with funding.

  2. o

    Unfunded articles are defined as the published articles in journals that had not received any funding.

In order to test the power-law correlation between citations and total (funded + unfunded)/funded/unfunded articles, we used the model developed by Ronda-Pupo and Katz, (2016a, b). This model analyzes the correlation between Citation-based Performance (CBP) and the number of articles published in journals of computer science, medicine and economics,

$${\text{CBP}} = kc^{\alpha }$$
(1)

CBP = for citation-based performance.

c = the number of articles (total: funded + unfunded/funded/unfunded).

k = constant.

α = the scaling factor.

Parameter α is a measure of the magnitude of the Matthew effect (Ronda-Pupo, 2016; Katz, 1999) and can be used to explain the exponent of the scaling correlation between articles that are published in journals and their citations. This parameter (α) and the parameter of the correlation k were calculated with the Ordinary Least Squares (OLS) method (Leguendre & Leguendre, 2012). According to Ronda-Pupo and Katz, (2017), since the Web of Science data is relatively noise-free, and as there is minor error in measuring the number of citations of articles, OLS was used as a line fitting method. The scaling exponent α can be used to predict the behavior of correlation.

In particular,

  • When α > 1, then CBP increases non-linearly, and citations of articles grow faster than the number of articles published in computer science, medicine, and economics journals. Thus, correlations are non-linear, and we have a positive Matthew effect or positive cumulative advantage.

  • If α < 1, the correlation is sub-linear, and its magnitude is a measure of the inverse Matthew effect or cumulative disadvantage. It means that the citations of articles grow slower than the number of published papers in journals.

  • Finally, when α = 1, the correlation is linear, and citations and articles published in journals grow at the same rate (Ronda-Pupo and Katz, 2017).

Student's T-statistics were used to confirm if the scaling exponents of the power-law correlation suggested is significant.

Results and discussion

Table 1 shows documents retrieved from the Web of Science database, organized per funding status.

Table 1 The number of documents and citations of computer science, medicine, and economics according to funding status

We analyzed 55,071 documents, including articles and reviews, with 870,807 citations published in 712 journals of computer science. 34,080 (61.88%) of documents had received funding, and 70.06% of all citations are associated with this category of documents. Furthermore, 38.12% of the papers published in computer science were unfunded, and 29.94% of the citations are associated with this category. We did not consider conference proceedings as they are rarely funded. In brief, the main patterns in this discipline suggest that computer science journals published more funded papers than unfunded ones.

We also analyzed 113,310 documents with 1,967,600 citations in 973 medicine journals. In this field, 71.13% of the citations account for 50.14% of all documents, which are funded research. Medicine journals published equally funded and unfunded documents.

Results also show 23,398 documents in 537 economics journals in which 51% of citations accounted for 38.98% of all documents, which were funded papers. Economics journals published more unfunded papers than funded.

In general, funded documents received more citations than unfunded papers in all three fields. These findings suggest that computer science, medicine, and economics have a high correlation between funding for research and studies being cited at least once. These findings confirm results by Wang and Shapira's study (2015) about the impact of research sponsorship on the field of nanotechnology science research.

The power-law behavior of the distributions of citations and the number of papers of journals was analyzed using the procedure developed by Clauset et al. (2009). We summarize the procedure in two steps:

  1. 1.

    estimating parameters Xmin and alpha;

  2. 2.

    comparing the power-law with alternative distributions. We also analyzed the power-law correlation between the number of citations and the number of papers of journals.

Verification of the power-law distribution for the papers in computer science, medicine, and economics

Table 2 shows results of fitting data to a power-law distribution by 2500 iterations using the Monte-Carlo bootstrapping analysis to distribute citations to the following three data sets. Each of the datasets includes total (funded + unfunded), funded, and unfunded papers. Statistical evidence seems, in general, to support the empirical results of the power-law distribution for all three categories of papers in computer science, medicine, and economics. Interestingly, in Total categories, medicine has the lowest Xmin, showing that with the lowest number of primary citations, the Matthew effect begins, and then citations start scaling exponentially. Economics has the highest Xmin: journals need more initial citations before their exponent scaling starts.

Table 2 Results of fitting the power-law to the datasets

Comparing the power-law with alternative distributions

According to Clauset et al. (2009), we cannot conclude that if the p-value of the power-law distribution is significant (p > 0.1), then it would be the best fit for the data. We compared the power-law model to lognormal, exponential, Poisson, stretched exponential, and power-law with exponential cut-off distributions as possible alternatives. According to Clauset et al. (2009), by using log-likelihood ratios (LR), we can identify which distribution can be considered as the best option. A distribution with the most negative LR is assumed to be the best distribution to fit the data. If the LR values of alternative distributions are positive, we can conclude that the power-law model is a plausible fit for the data. Table 3 shows the comparison between different distributions, including the magnitude of the scaling exponent α, its p-value, and the log-likelihood ratio test (LR) for alternative distributions. The final column of Table 3 shows the statistical support indicating if the power-law fits for distributions. We used ‘Moderate’ to indicate that the power-law is a good fit, but there are other plausible alternatives as well. In some cases, the notion of ‘with cut-off’ indicates that the power-law with exponential cut-off is clearly preferable over the pure power-law (cf. Clauset et al. 2009, p. 687).

Table 3 Results of comparing the power-law with alternative distributions

Computer Science: The p-value of the power-law distribution for computer science in the total (funded + unfunded), funded, and unfunded categories was significant (0.36, 0.7, 0.79), and we can conclude that the power-law distribution can be a plausible fit for our data. In the total (funded + unfunded)category, the p-values for exponential and Poisson distributions were statistically significant (p < 0.1). Therefore, the exponential and Poisson distributions cannot be completely ruled out for our data. Still, the power-law distribution was likely a good fit as the LR of the alternatives were positive. In the funded sub-dataset, the p-values of exponential, Poisson, and stretched exponential distributions were significant (0.001, 0.001, 0.02), so they could not be completely dismissed merely because their LR was 3.11, 3.42, and 2.25.Moreover, in the unfunded sub-dataset, the p-value of the exponential distribution was also statistically significant, but the LR was 3.81. These results suggest that we cannot rule out the exponential distribution for this data. The power-law seems to have a good fit for the distribution of citations.

Medicine: All alternative distributions had significant p-values for total (funded + unfunded) and funded data. The power-law distribution with cut-off had the most negative value of LR (− 1.48 for the total (funded + unfunded) category,  − 1.83 for the funded category), so this distribution was preferable compared to the pure power-law distribution. In unfunded data, the p-values of the exponential and Poisson are also statistically significant with p < 0.1. Accordingly, the exponential and Poisson distributions could not be completely ruled out in this category. However, the power-law distribution was marked as ‘Moderate’ as it was favorable because of the positive LR of alternatives. In particular, the power-law distribution with cut-off was generated from the pure one, and Katz, (2016) asserts, "models show that the probability distribution tends to evolve from a power-law with an exponential cut-off to a pure power-law given enough time". In life sciences, Su and Hogenesch, (2007) show power-law-like distributions in gene annotation (measuring links to the biomedical literature) and research funding, measured by gene references in funded grants (cf. Lewison & Dawson, 1998).

Economics: In the journals of this field, for total (funded + unfunded) and funded categories, the p-values for exponential and Poisson distributions were significant (p < 0.1). Therefore, the exponential and Poisson distributions could not be completely ruled out for our data in these two classes. Still, the power-law distribution was a plausible fit as the LR values of other distributions were positive. In the unfunded sub-dataset, the statistical p-values of exponential, Poisson, and power-law with cut-off distributions were significant (0.001, 0.001, 0.03). This distribution was preferable compared to the pure power-law distribution in the presence of the most negative LR value of power-law with cut-off. The results seem to support that the power-law can be a good fit in all categories of our dataset.

Power-law correlation between citation-based performance and number of papers in three categories of total (funded + unfunded), funded, and unfunded documents

Table 4 shows that all correlations are statistically significant, p-value < 0.001.

Table 4 Values of the exponents for the power-law correlation

In the field of computer science, Table 4 reveals that the power-law correlation between the number of papers and citations is statistically significant t(1,707), 10.6, p < 0.001. The power-law correlation between funded papers and citation-based performance is statistically significant: t(1, 660), 9.3, p < 0.001. The power-law correlation between unfunded documents and the number of citations is statistically significant: t(1, 705), 11.7, p < 0.001 (cf. Figure 1). The results suggest that the Matthew effect (Merton, 1968, 1988) is greater for funded articles compared with unfunded documents. The estimation of α for funded papers is 1.30 ± 0.03, which means that the citation performance of research will increase by 21.30 or 2.46 times when they double the number of papers. The exponent > 1 shows that citations grow faster than the rate of publication for funded papers. The exponent for the relationship between citations and unfunded papers is 1.17 ± 0.05. To put it differently, for a doubling of numbers of unfunded publications, the number of citations increases 21.17 or 2.25 times. Although there is a positive exponential relationship, the alpha is low compared to funded category. With a higher magnitude, doubling the number of funded papers increases the citation performance of research. The higher exponent for the funded papers suggests that an approach of increasing funding is required in this field to have a greater impact on the diffusion of science in society.

Fig.1
figure 1

The power-law relationship between citation performance and papers in total (funded + unfunded), funded, and unfunded of computer science, medicine, and economics

In the field of medicine, the power-law correlation between papers in funded + unfunded and citations is statistically significant: t(1,972), 9.02, p < 0.001. The power-law correlation between funded papers and citation-based performance is also statistically significant: t(1, 866), 7.18, p < 0.001. Finally, the power-law correlation between unfunded documents and the number of citations is statistically significant: t(1, 963), 14.9, p < 0.001. The α value of the total (funded + unfunded) category is estimated at 1.39 ± 0.02, suggesting a positive Matthew effect of 21.39 or 2.62 between the citation impact and the number of papers published in journals. Figure 1 shows that the power-law correlation between citations and total (funded + unfunded), funded, and unfunded research in medicine has a greater exponential scaling relationship with an α value of 1.36 in funded research compared to the unfunded category having an α value of 1.09. With an increase in the number of funded papers, the scaling of the citation performance for funded research increases by about 21.36 or 2.57 times. The expected number of citations increases by 21.09 or 2.13 times when the number of unfunded papers published in medicine journals doubles. The Matthew effect is stronger for funded papers than unfunded papers. In fact, by increasing the number of funded papers in this field, we can expect a higher magnitude of citation impact compared to unfunded category.

In the field of economics, the power-law correlation between papers in funded + unfunded and citations is statistically significant t(1,537), 10.1, p < 0.001. The power-law correlation between funded papers and citation-based performance is statistically significant t(1, 470), 19.63, p < 0.001. The power-law correlation between unfunded documents and the number of citations is statistically significant t(1, 534), 10.8, p < 0.001. The discipline of economics also shows a scaling relationship. The value of α for this field in the total (funded + unfunded) set is estimated at 1.51 ± 0.04, which means the citation performance of research will increase by 21.51 or 2.85 times when they double the number of published papers. Figure 1 shows the power-law correlation results between citations and total (funded + unfunded), funded, and unfunded research. The value of α in funded papers is 1.46 ± 0.03, indicating that by doubling the number of funded documents in this field, the number of citations multiplies by more than two times: about 21.46 or 2.75. The Matthew effect is greater than the unfunded dataset, in which the value of α is estimated to be 1.37 ± 0.047.

Finally, Table 4 also shows that coefficients of R2 of the funded datasets in computer science, medicine, and economics are higher than R2 of the power-law model for unfunded papers. As a result, the model has more explanatory power for the funded dataset.

Overall, then, statistical evidence of this inductive study based on three vital disciplines can be generalized in the following basic properties of dynamics of science.

Property 1. Research citations of funded and unfunded papers in different disciplines evolve with a power-law distribution.

Property 2. The number of articles and citation-based performance in different disciplines have a scaling correlation.

Property 3. The Matthew effect is greater for funded articles compared to unfunded documents in different disciplines.

Conclusions

The present study explored the scaling relationship between citation-based performance and articles published in vital disciplines given by: computer science, medicine and economics. We analyzed the distribution of citations for papers in total (funded + unfunded), funded and unfunded researches separately. The results suggested that the power-law with exponential cut-off distribution analysis is the best fit for distributing citations to total (funded + unfunded) and funded papers of the medicine and unfunded papers of economics journals. All three categories of computer science, unfunded papers of the field of medicine, total (funded + unfunded) and funded papers of the field of economics are supported with the power-law distribution. Table 3 shows that the power-law is a plausible distribution for these datasets.

In computer science, results suggest a scaling correlation between the number of articles published in journals and citation-based performance (CBP) as: CBP ≈ C1.31 for the total (funded + unfunded) dataset, CBP ≈ C1.30 for unfunded papers. Results reveal that the power-law correlation between the number of funded papers and citations in the field of computer science is statistically significant. Findings here show that the Matthew effect in funded papers is more powerful in comparison to unfunded studies.

With regards to the field of medicine, results also show a significant scaling correlation between the number of articles and citation-based performance (CBP), as: CBP ≈ C1.39 for the total (funded + unfunded) dataset, CBP ≈ C1.36 for funded papers, and CBP ≈ C1.09 for unfunded papers. Findings also show that the Matthew effect in funded papers is greater than that of the unfunded dataset.

Finally, with regards to the field of economics, our findings show a scaling relationship between the number of papers and citation-based-performance (CBP) as: CBP ≈ C1.51 for the total (funded + unfunded) dataset, CBP ≈ C1.46 for funded papers, and CBP ≈ C1.37 for unfunded papers. The Matthew effect in funded papers is also more considerable in comparison to unfunded research.

These results have main implications of research policy because a strategy of increasing funding for research can generate a higher diffusion of scientific results and a higher impact of science in society (Coccia et al. 2015; Coccia, 2005a, b).

These conclusions are certainly tentative. There is a need for much more detailed research into the relations under study here. As a future development, additional empirical research can further clarify an optimal scheme of research funding between different disciplines in order to support a vast diffusion of vital discoveries and new technologies for progress and well-being in human society.