Abstract
Science is a driving force of positive social evolution. And in the course of this evolution, research systems change as a consequence of their complex dynamics. Research systems must be managed very carefully, for they are dissipative, and their evolution takes place on the basis of a series of instabilities that may be constructive (i.e., can lead to states with an increasing level of organization) but may be also destructive (i.e., can lead to states with a decreasing level of organization and even to the destruction of corresponding systems). For a better understanding of relations between science and society, two selected topics are briefly discussed: the Triple Helix model of a knowledge-based economy and scientific competition among nations from the point of view of the academic diamond. The chapter continues with a part presenting the minimum of knowledge necessary for understanding the assessment of research activity and research organizations. This part begins with several remarks on the assessment of research and the role of research publications for that assessment. Next, quality and performance as well as measurement of quality and latent variables by sets of indicators are discussed. Research activity is a kind of social process, and because of this, some differences between statistical characteristics of processes in nature and in society are mentioned further in the text. The importance of the non-Gaussianity of many statistical characteristics of social processes is stressed, because non-Gaussianity is connected to important requirements for study of these processes such as the need for multifactor analysis or probabilistic modeling. There exist entire branches of science, scientometrics, bibliometrics, informetrics, and webometrics, which are devoted to the quantitative perspective of studies on science. The sets of quantities that are used in scientometrics are mentioned, and in addition, we stress the importance of understanding the inequality of scientific achievements and the usefulness of knowledge landscapes for understanding and evaluating research performance. Next, research production and its assessment are discussed in greater detail. Several examples for methods and systems for such assessment are presented. The chapter ends with a description of an example for a combination of qualitative and quantitative tools in the assessment of research: the English–Czerwon method for quantification of scientific performance.
Dedicated to Derek John de Solla Price and to all Price award winners whose contributions established scientometrics, bibliometrics and informertics as important and fast developing branches of the modern science.
Access provided by Autonomous University of Puebla. Download chapter PDF
Keywords
- Research Organization
- Research Publication
- Scientific Achievement
- Expert Evaluation
- Scientific Organization
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
1 Introductory Remarks
The word science originates from the Latin word scientia, which means knowledge. Science is a systematic enterprise that builds and organizes knowledge in the form of testable explanations and predictions about the Universe. Modern science is a discovery as well as an invention. It is a discovery that Nature generally acts regularly enough to be described by laws and even by mathematics; and it required invention to devise the techniques, abstractions, apparatus, and organization for exhibiting the regularities and securing their law-like descriptions [1, 2]. The institutional goal of science is to expand certified knowledge [3]. This happens by the important ability of science to produce and communicate scientific knowledge . We stress especially the communication of new knowledge, since communication is an essential social feature of scientific systems [4]. This social function of science has long been recognized [5–9].
Research is creative work undertaken on a systematic basis in order to increase the stock of knowledge, including knowledge of humans, culture, and society, and the use of this stock of knowledge to devise new applications [10]. Scientific research is one of the forms of research. Usually, modern science is connected to research organizations. In most cases, the dynamics of these organizations is nonlinear. This means that small influences may lead to large changes. Because of this, the evolution of such organizations must be managed very carefully and on the basis of sufficient knowledge on the laws that govern corresponding structures and processes. This sufficient knowledge may be obtained by study of research structures and processes. Two important goals of such studies are (i) adequate modeling of dynamics of corresponding structures and (ii) design of appropriate tools for evaluation of production of researchers.
This chapter contains the minimum amount of knowledge needed for a better understanding of indicators, indexes, and mathematical models discussed in the following chapters. We consider science as an open system and stress the dissipative nature of research systems. Dissipativity of research systems means that they need continuous support in the form of inflows of money, equipment, personnel, etc. The evolution of research systems is similar to that of other open and dissipative systems: it happens through a sequence of instabilities that lead to transitions to more (or less) organized states of corresponding systems.
Science may play an important role in a national economic system. This is shown on the basis of the Triple Helix model of a knowledge-based economy. Competition is an important feature of modern economics and society. Competition has many faces, and one of them is scientific competition among nations. This kind of competition is connected to the academic diamond: in order to be successful in globalization, a nation has to possess an academic diamond and use it effectively.
In order to proceed to the methods for quantitative assessment of research and research organizations and to mathematical models of science dynamics, one needs some basic information about assessment of research. A minimum of such basic information is presented in the second part of the chapter. The discussion begins with remarks about quality and measurement of processes by process indicators. Measurement can be qualitative and quantitative, and four kinds of measurement scales are described. The discussion continues with remarks on the non-Gaussianity that occurs frequently as a feature of social processes. Research also has characteristics of a social process, and many components and processes connected to research possess non-Gaussian statistical characteristics.
If one wants to measure research, one needs quantitative tools for measurement. Scientometrics , bibliometrics, and informetrics provide such tools, and a brief discussion of quantities that may be measured and analyzed is presented further in the text. In addition, another useful tool for analysis of research and research structures, the knowledge landscape, is briefly discussed. Next, research production is discussed in more detail. Special attention is devoted to publications and citations, since they contain important information that is useful for assessment of research production. The discussion continues with remarks on methods and systems for assessment of research and research organizations. Tools for assessment of basic research as well as the method of expert evaluation and several systems for assessment of research organizations applied in countries from continental Europe are briefly mentioned. The discussion ends with a description of the English–Czerwon method for quantification of performance of research units, which makes it possible to combine qualitative and quantitative information in order to compare results of research of research groups or research organizations.
2 Science, Technology, and Society
Knowledge is our most powerful engine of production
Alfred Marshall
Science, innovation, and technology have led some countries to a state of developed societies and economies [11–16]. Thus science is a driving force of positive social evolution, and the neglect of this driving force may turn a state into a laggard [17]. Basic research is an important part of the driving force of science. This kind of research may have large economic consequences, since it produces scientific information that has certain characteristic features of goods [18] such as use value and value. The use value of scientific information is large if the obtained scientific information can be applied immediately in practice or for generation of new information. One indicator for the measure of this value is the number of references of the corresponding scientific publication. The value of scientific information is large when it is original, general, coherent, valid, etc. The value of scientific information is evaluated usually in the “marketplace” such as scientific journals or scientific conferences.
The lag between basic research and its economic consequences may be long, but the economic impact of science is indisputable [19, 20]. This is an important reason to investigate the structures, laws, processes, and systems connected to research [21–26]. The goals of such studies are [27]: better management of the scientific substructure of society [28–30 ] , increase of effectiveness of scientific research [31–34 ] , efficient use of science for rapid and positive social evolution. The last goal is connected to the fact that science is the main factor in the increase of productivity. In addition, science is a sociocultural factor, for it directly influences the social structures and systems connected to education, culture, professional structure of society, social structure of society, distribution of free time, etc. The societal impact of science as well as many aspects of scientific research may be measured [35–43].
Science is an information-producing system [44, 45]. That information is contained in scientific products. The most important of these products are scientific publications, and the evaluation of results of scientific research is usually based on scientific publications and on their citations. Scientific information is very important for technology [46–48] and leads to the acceleration of technological progress [49–59]. Science produces knowledge about how the world works. Technology contains knowledge of some production techniques. There are knowledge flows directed from the area of science to the area of technology [60, 61]. In addition, technological advance leads to new scientific knowledge [62], and in the process of technological development, many new scientific problems may arise. New technologies lead also to better scientific equipment. This allows research in new scientific fields, e.g., the world of biological microstructures. Advances in science may reduce the cost of technology [63–66]. In addition, advances in science lead to new cutting-edge technologies, e.g., laser technologies, nanoelectronics, gene therapy, quantum computing, some energy technologies [67–74]. But the cutting-edge technologies do not remain cutting-edge for long. Usually, there are several countries that are the most advanced technologically (technology leaders), and the cutting-edge technologies are concentrated in those countries. And those countries generally possess the most advanced research systems.
In summary, what we observe today is a scientifically driven technological advance [75–81]. And in the long run, technological progress is the major source of economic growth.
The ability of science to speed up achievement of national economic and social objectives makes the understanding of the dynamics of science and the dynamics of research organizations an absolute necessity for decision-makers. Such an understanding can be based on appropriate systems of science and technology indicators and on tools for measurement of research performance [82–87]. Because of this, science and technology indicators are increasingly used (and misused) in public debates on science policy at all levels of government [88–96].
3 Remarks on Dissipativity and the Structure of Science Systems
The following point of view exists about the evolution of open systems in thermodynamics [97, 98]:
The evolution of an open thermodynamic system is a sequence of transitions between states with decreasing entropy (increasing level of organization) with an initial state sufficiently far from equilibrium. If the parameters of such systems change and the changes are large enough, the system becomes unstable, and there exists the possibility that some fluctuation of the parameters may push the system to a new state with smaller entropy. Thus the transition takes place through an instability.
This type of development may be observed in scientific systems too. This is not a surprise, since scientific systems are open (they interact with a complex natural and social environment), and they are able to self-organize [99]. In addition, crises exist in these systems, and often these crises are solved by the growth of an appropriate fluctuation that pushes the scientific system to a new state (which can be more or less organized than the state before the crisis). Hence instabilities are important for the evolution of science, and it is extremely important to study the instabilities of scientific (and social) systems [100–102]. The time of instability (crisis) is a critical time, and the regime of instability is a critical regime. The exit from this time and this regime may lead to a new, more organized, and more efficient state of the system or may lead to degradation and even to destruction of the system.
3.1 Financial, Material, and Human Resource Flows Keep Science in an Organized State
Dissipative structures: In order to keep a system far from equilibrium, flows of energy, matter, and information have to be directed toward the system. These flows ensure the possibility for self-organization, i.e., the sequence of transitions toward states of smaller entropy (and larger organization). The corresponding structures are called dissipative structures, and they can exist only if they interact intensively with the environment. If this interaction stops and the above-mentioned flows cease to exist, then the dissipative structures cannot exist, and the system will end at a state of thermodynamic equilibrium where the entropy is at a maximum and organization is at a minimum.
Science structures are dissipative. In order to exist, they need inflows of information (since scientific information becomes outdated relatively fast), people (since the scientists retire or leave and have to be replaced), money (needed for paying scientists, for building and supporting the scientific infrastructure), materials (for running experiments, machines, etc.), etc. The weak point of the dissipative structures is that they can be degraded or even destroyed by decreasing their supporting flows [103]. In science, this type of development to retrograde states may be observed when the flows of financial and material support decrease and flows of information decrease or cease.
3.2 Levels, Characteristic Features, and Evolution of Scientific Structures
Researchers act in two directions: (i) they produce new knowledge and information [104, 105] and decrease the disorder as current knowledge become better organized; (ii) the work of researchers leads to new problems and the possibility for new research directions and thus opens the way to new states with an even higher level of organization. By means of these actions, researchers influence the structure of science. There exist three levels and four characteristic features of the scientific structure [106]. The three levels are:
-
1.
Level of material structure: Here are the scientific institutes, material conditions for scientific work, etc.
-
2.
Level of social structure: This includes the scientists and other personnel as well as the different kinds of social networks connected to scientific organizations.
-
3.
Level of intellectual structure: This includes the structures connected to scientific knowledge and the field of scientific research. There are differences in the intellectual structures connected to the social sciences in comparison to the intellectual structures connected to the natural sciences.
The four characteristic features of the scientific structure are:
-
1.
Dependence on material, financial, and information flows. These flows are directed mainly to the material levels of the scientific structure. They include the flows of money and materials that are needed for the scientific work. But there are also flows to other levels of the scientific structure. An important type of such flows is motivation flows. For example, there exist (i) psychological motivation flow: connected to the social level of the scientific structure. This motivation flow is needed to support each scientist to be an active member of scientific networks and to be an expert in the area of his or her scientific work; (ii) intellectual motivation flow: connected to the intellectual level of the scientific structure. This flow supports scientists to learn constantly and to absorb the newest scientific information from their research area.
-
2.
Cyclical behavior of scientific productivity. At the beginning of research in a new scientific area, there are many problems to be solved, and scientists deal with them (highly motivated, for example, by the intellectual motivation flow and possibly by material flows that the corresponding wise national government assigns to support the research in this area). In the course of time, the simple scientific problems are solved, and what remains are more complex unsolved problems. The corresponding scientific production (the number of publications, for example) usually decreases. Some scientists change their field of research, and then a new scientific area or subarea may arise in this new field of research.
-
3.
Homeostatic feature.
Homeostasis is the property of a system to regulate its variables in such a way that internal conditions remain stable and relatively constant.
-
This feature of science is supported by the system of education, the set of traditions and institutional norms, the books and other material and intellectual tools that ensure the translation of knowledge from one generation of scientists to the next, etc. All this contributes to the stable functioning of scientific systems and helps them to overcome unfavorable environmental conditions.
-
4.
Limiting factors. Limiting factors can be (i) material factors that decrease the intensity of work of the scientific organizations (such as decreased funding, for example); (ii) factors connected to decreasing the speed of the process of exchange of scientific information (closing access to an important electronic scientific journal, for example); (iii) factors that decrease the speed of obtaining new scientific results (for example, the constant pressure to increase the paperwork of scientists).
Scientific structures evolve. This evolution is connected to the evolution of scientific research [107–109]. Usually, the evolution of scientific structures has four stages: normal stage, network stage, cluster stage, specialty stage. Institutional forms of research evolve, for example, as follows. At the normal stage, these forms are informal; then small symposiums arise at the network stage. At the cluster stage, the symposiums evolve to formal meetings and congresses, and at the specialty stage, one observes institutionalization (research groups and departments at research institutes and universities). Cognitive content evolves too. At the normal stage, a paradigm is formulated. At the network stage, this paradigm is applied, and in the cluster stage, deviations from the paradigm (anomalies) are discovered. Then at the specialty stage, one observes exhaustion of the paradigm, and the cycle begins again by formulation of a new paradigm.
Now let us consider a more global point of view on research systems and structures and let us discuss briefly two additional aspects connected to these systems:
-
The place of research in the economic subsystem of society from the point of view of the Triple Helix model of the knowledge-based economy;
-
Relations among different national research systems: we discuss the competition among these systems from the point of view of the concept of the academic diamond.
4 Triple Helix Model of the Knowledge-Based Economy
Research priorities should be selected by taking into account primarily the requirements of the national economics and society, traditions and results previously attained, possible present and future human and financial potential, international relationships, trends in the world’s economic and social growth, and trends of science.
Peter Vinkler
The Triple Helix model of the knowledge-based economy defines the main institutions in this economy as university (academia), industry, and government [110–119]. The Triple Helix has the following basic features:
-
1.
A more prominent role for the university (and research institutes) in innovation, where the other main actors are industry and government.
-
2.
Movement toward collaborative relationships among the three major institutional spheres, in which innovation policy should be increasingly an outcome of interaction rather than a prescription from government.
-
3.
Any of the three spheres may take the role of the other, thus performing new roles in addition to their traditional function. This taking of nontraditional roles is viewed as a major source of innovation.
Organized knowledge production adds a new coordination mechanism in social systems (knowledge production and control) in addition to the two classical coordination mechanisms (economic exchanges and political control). In the Triple Helix model, the economic system, the political system, and the academic system are considered relatively autonomous subsystems of society that operate with different mechanisms. In addition to their autonomy, however these subsystems are interconnected and interdependent. There are amendments in the model of the Triple Helix, and even models of the helix exist with more than three branches [120].
The Triple Helix model allows for the evolution of the branches of the helix. At the beginning of operation of the Triple Helix:
-
1.
Industry operates as a concentration point of production.
-
2.
Government operates as the source of contractual relations and has to be a guarantor for stable interactions and exchange.
-
3.
The academy operates as a source of new knowledge and technology, thus generating the base for establishing a knowledge-based economy.
With increasing time, the place of academia (universities and research institutes) in the helix changes. Initially, the academy is a source of human resources and knowledge, and the connection between academia and industry is relatively weak. Then academia develops organizational capabilities to transfer technologies, and instead of serving only as a source of new ideas for existing firms, academia becomes a source of new firm formation in the area of cutting-edge technologies and in advanced areas of science. Academia becomes a source of regional economic development, and this leads to the establishment of new mechanisms of economic activity and community formation (such as business incubators, science parks, and different kinds of networks between academia and industry). Government supports all this by its traditional regulatory role in setting the rules of the game and also by actions as a public entrepreneur.
The Triple Helix model is a useful model that helps researchers, managers, et al. to imagine the place of research structures in the complex structure of modern economics and society. Let us mention that the Triple Helix can be modeled on the basis of the evolutionary “lock-in” model of innovations [121] connected to the efforts of adoption of competing technologies [122, 123]. And various concepts from time series analysis such as the concept of mutual information [119] can be used to study the Triple Helix dynamics.
5 Scientific Competition Among Nations: The Academic Diamond
It is not enough to do your best. You must know what to do and then do your best
W. Edwards Deming
Globalization creates markets of huge size, and every nation wants to be well represented at these markets with respect to exports of goods, etc. This can happen if a nation has competitive advantages. One important such advantage is the existence of effective national research and development (R & D) systems. Let us note that the scientific production by researchers, research groups, and countries is an object of absolute competition regardless of possible poor equipment, low salaries, or lack of grants for some of the participants in this competition. From this point of view, the evaluation of scientific results may be regarded as unfair if one compares scientists from different nations [4]. Poor working conditions for scientists is clearly a competitive disadvantage to the corresponding nation. In order to export high-tech production, the scientific and technological system of a nation has to work smoothly and be effective enough. A nation that has such a system and uses it effectively for cooperation [124, 125] and competition has a competitive advantage in the global markets. And in order to have such a system, a country should invest wisely in the development of its scientific system and in the processes of strengthening the connection between the national scientific, technological , and business systems and structures [126–130]. In particular, the four parts of the so-called academic diamond [131] should be cultivated.
Each of the four parts of the academic diamond is connected to the other three parts. The parts are:
-
1.
Factor conditions: human resources (quantity of researchers, skills levels [132], etc.), knowledge resources (government research institutes, universities, private research facilities, scientific literature, etc.), physical and basic resources (land, water and mineral resources, climatic conditions, location of the country, proximity to other countries with similar research profiles, size of country, etc.), capital resources (government funding of scientific structures and systems, cost of capital available to finance academia, private funding for research projects, etc.), infrastructure (quality of life, attractiveness of country for skilled scientists, telecommunication systems, etc.).
-
2.
Strategy, structure, and rivalry: goals and strategies of the research organizations (research profile, positioning and key faculties or research areas, internationalization path in terms of staff, campuses, and student body, etc.), local rules and incentives (salaries, promotion system, incentives for publication, etc.), local competition (number of research universities, research institutes, research centers, existing research clusters, territorial dynamics of scientific organizations, etc.).
-
3.
Demand conditions: public and private sectors (demand for training and job positions for researchers, etc.), student population (trained students), other academics in country and abroad (active research scientists outside the government research institutes and universities).
-
4.
Related and supporting industries: publication industry, information technology industry, other research institutions.
In addition, the academic diamond has two more components: chance and government . There are different aspects of chance connected to the research organizations. If we consider chance as the possibility for something to happen, then some countries have elites that ensure a good chance with respect to the positive development of science and technology. Government may contribute to the development of scientific and technological systems of a country. This contribution can be made through appropriate politics with respect to (higher) education; government research institutes; basic research [133, 134]; funding of research and development; economic development; etc.
6 Assessment of Research: The Role of Research Publications
Research is an important process in complex scientific systems. Research production is a result of this process that can be assessed. Quantitative assessment of research (at least of publicly funded basic research) has increased greatly in the last decade [135–138]. Some important reasons for this are economic and societal [134]: constraints on public expenditures, including the field of research and development; growing costs of instrumentation and infrastructure; requirements for greater public accountability; etc. Another reason is connected to the development of information technologies, bibliometrics, and scientometrics in the last fifty years. Several goals of quantitative assessment of research are [4] to obtain information for granting research projects; to determine the quantity and impact of information production for monitoring research activities; to analyze national or international standing of research organizations and countries’ organizations for scientific policy; to obtain information for personnel decisions; etc.
In addition to the rise of quantitative assessment of research, one observes a process of the increasing use of mathematics in different areas of knowledge [139]. This process also concerns the field of knowledge about science. In the process of human evolution, more and more scientific facts have been accumulated, and these facts have been ordered by means of different methods that include also methods of mathematics. In addition, the use of mathematics (which means also the use of mathematical methods beyond the simplest statistical methods) is important and much needed for supporting decisions in the area of research politics.
Many mathematical methods in the area of assessment of research focus on the study of research publications and their citations. This is because publications are an important form of the final results of research work [140–142]. There is a positive correlation between the number of research publications and the meaning that society attaches to the scientific achievements of the corresponding researcher. There exists also a positive correlation between the number of a researcher’s publications and the expert evaluation of his/her scientific work [143]. Senter [144] mentions five factors that may positively influence the research productivity of a researcher:
-
1.
Education level: has important positive impact on productivity;
-
2.
Rank of the scientist: has immediate positive impact on scientific productivity;
-
3.
Years in service: positive impact on productivity but more modest in comparison to the impact of education and rank;
-
4.
Influence of scientist on its research endeavor: positive impact but modest in comparison with the above three factors;
-
5.
Psychological factors: usually they have small effect on productivity (if the problems that influence the psychological condition of the research are not too big).
In recent years, the requirements on the quality of research have increased. Because of this, we shall discuss briefly below several characteristics of quality, performance, quality management systems , and performance management systems, since they are important for the assessment of the quality of the results of basic and applied research [145–148].
7 Quality and Performance: Processes and Process Indicators
Scientific research and its product, scientific information, is multidimensional, and because of this, the evaluation of scientific research must also be multidimensional and based on quantitative indexes and indicators accompanied by qualitative tools of analysis. One important characteristic of research activity is its quality, because the performance of any organization is connected to the quality of its products [149–153]. A simple definition of quality is this: Quality is the ability to fulfill a set of requirements with concrete and measurable actions. The set of requirements can include social requirements, economic requirements, productive requirements, and specific scientific requirements. The set of requirements depends on the stakeholders’ needs and on the needs of producers. These needs should be fulfilled effectively, and an important tool for achieving this is a quality management system. In order to manage quality, one introduces different quality management systems (QMS), which are sets of tools for guiding and controlling an organization with respect to quality aspects of human resources; working procedures, methodologies and practices; technology and know-how.
Research production is organized as a set of processes. A simple definition of a process is as follows: A process is an integrated system of activities that uses resources to transform inputs into outputs [149]. We can observe and assess processes by means of appropriate indicators. An indicator is the quantitative and/or qualitative information on an examined phenomenon (or process or result), which makes it possible to analyze its evolution and to check whether (quality) targets are met, driving actions and decisions [154]. Let us note that we do not need simply to use some indicators. We have to identify the indicators that properly reflect the observed process. These indicators are called key performance indicators.
The main functions of indicators are as follows.
-
1.
Communication. Indicators communicate performance to the internal leadership of the organization and to external stakeholders.
-
2.
Control. Indicators help the leadership of an organization to evaluate and control performance of the corresponding resources.
-
3.
Improvement. Indicators show ways for improvement by identifying gaps between performance and expectations.
Indicators supply us with information about the state, development, and performance of research organizations. Performance measurements are important for taking decisions about development of research organizations [155]. In general, performance measurements supply information about meeting the goals of an organization and about the state of the processes in the organization (for example, whether the processes are in control or there are some problems in their functioning). In more detail, the performance measurement supplies information about the effectiveness of the processes : the degree to which the process output conforms to the requirements, and about efficiency of the processes: the degree to which the process produces the required output at minimal resource cost. Finally, the performance measurements supply information about the need for process improvement.
8 Latent Variables, Measurement Scales, and Kinds of Measurements
Latent features of the studied objects and subjects often are the features we want to measure. One such feature is the scientific productivity of a researcher [156, 157]. Latent features are characterized by latent variables. Latent variables may reflect real characteristics of the studied objects or subjects, but a latent variable is not directly measurable. The indicators are what we measure in practice, e.g., the number of publications or the number of citations. Many latent variables can be operationally defined by sets of indicators. In the simplest case, a latent variable is represented by a single indicator. For example, the production of a researcher may be represented by the number of his/her publications. If we want a more complete characterization of the latent variables, we may have to use more than one indicator for their representation, e.g., one has to avoid (if possible) the reduction of representation of a latent variable to a single indicator. Instead of this, a set of at least two indicators should be used.
A measurement means that certain items are compared with respect to some of their features. There are four scales of measurement:
-
1.
Nominal scale: Differentiates between items or subjects based only on their names or other qualitative classifications they belong to. Examples are language, gender, nationality, ethnicity, form. A quantity connected to the nominal scale is mode: this is the most common item, and it is considered a measure of central tendency.
-
2.
Ordinal scale: Here not only are the items and subject distinguished, but also they are ordered (ranked) with respect to the measured feature. Two notions connected to this scale are mode and median: this is the middle-ranked item or subject. The median is an additional measure of central tendency.
-
3.
Interval scale: For this scale, distinguishing and ranking are available too. In addition, a degree of difference between items is introduced by assigning a number to the measured feature. This number has a precision within some interval. An example for such a scale is the Celsius temperature scale. The quantities connected with the interval scale are mode, median, arithmetic mean, range : the difference between the largest and smallest values in the set of measured data. Range is a measure of dispersion. An additional quantity connected to this kind of scale is standard deviation: a measure of the dispersion from the (arithmetic) mean.
-
4.
Ratio scale: Here in addition to distinguishing, ordering, and assigning a number (with some precision) to the measured feature, there is also estimation of the ratio between the magnitude of a continuous quantity and a unit magnitude of the same kind. An Example of ratio scale measurement is the measurement of mass. If a body’s mass is 10 kg and the mass of another body is 20 kg, one can say that the second body is twice as heavy. If the temperature of a body is 20 \(^{\circ }\)C and the temperature of another body is 40 \(^{\circ }\)C, one cannot say that the second body is twice as warm (because the measure of the temperature in degrees Celsius is a measurement by interval scale and not by ratio scale. The measure of temperature by a ratio scale is the measure in kelvins.
In addition to all quantities connected to the interval scale of measurement, for the ratio scale of measurement one has the following quantities: geometric mean, harmonic mean, coefficient of variation, etc.
With respect to the four scales, there are the following two kinds of measurements:
-
1.
Qualitative measurements: measurements on the basis of nominal or ordinal scales.
-
2.
Quantitative measurements: measurements on the basis of interval or ratio scales.
Before the start of a measurement, a researcher has to perform:
-
1.
qualitative analysis of the measured class of items or subjects in order to select features that are appropriate for measurement from the point of view of the solved problems;
-
2.
choice of the methodology of measurement.
After the measurements are made, it is again time for qualitative analysis of the adequacy of the results to the goals of the study: some measurement can be adequate for one problem, and other measurements can be adequate for another problem. The adequacy depends on the choice of the features that will be measured.
9 Notes on Differences in Statistical Characteristics of Processes in Nature and Society
Let us assume that measurements have led us to some data about a research organization of interest. Research systems are also social systems, and because of this, we have to know some specific features of these systems and especially the characteristics connected to the possible non-Gaussianity of the system.
A large number of processes in nature and society are random. These processes have to be described by random variables. If x is a random variable, it is characterized by a probability distribution that gives the probability of each value associated with the random variable x arising. Probability distributions are characterized by a probability distribution function \(P(x \le X)\) or probability density function \(p(x)=dP/dx\) .
If we want to study the statistical characteristics of some population of items, we study statistical characteristics of samples of the population. We have to be sure that if the sample size is large enough, then the results will be close to the results that would be obtained by studying the entire population.
For the case of a normal (Gaussian) distribution, the central limit theorem guarantees this convergence. For the case of non-Gaussian distributions, however, there is no such guarantee.
Let us discuss this in detail. We begin with the central limit theorem. The central limit theorem of mathematical statistics is the cornerstone of the part of the world described by Gaussian distributions. It is connected to the moments of a probability distribution p(x) with respect to some value X:
The following two moments are of interest for us here:
-
1.
The first moment (\(n=1\)) with respect to \(X=0\): this is the mean value \(\overline{x}\) of the random variable;
-
2.
The second moment (\(n=2\)) with respect to the mean (\(X=\overline{x}\)): dispersion of the random variable (denoted also by \(\sigma ^2\)).
The central limit theorem answers the following question. We have a population of items or subjects characterized by the random variable x. We construct samples from this population and calculate the mean \(\overline{x}\). If we take a large enough number of samples, then what will be the distribution of the mean values of those samples?
The central limit theorem states that if for the probability density function p(x), the finite mean and dispersion exist, then the distribution of the mean values converges to the Gaussian distribution as the number of samples increases. The distributions that have this property are called Gaussian.
But what will be the situation if a distribution does not have the Gaussian property (for example, the second moment of this distribution is infinite)? Such distributions exist [158–160]. They are called non-Gaussian distributions, and some of them play an important role in mathematical models of social systems, and in particular in the models connected to science dynamics. There exists a theorem (called the Gnedenko–Doeblin theorem) that states the central role of one distribution in the world of non-Gaussian distributions. This distribution is called the Zipf distribution. Non-Gaussian distributions (and the Zipf distribution) will be discussed in Part III of this book.
Most distributions that arise in the natural sciences are Gaussian. Many distributions that arise in the social sciences are non-Gaussian (Fig. 1.1). Such distributions arise very often in the models of science dynamics [161, 162]. We do not claim that only Gaussian distributions are observed in the natural sciences and that the distributions that are observed in the social sciences are all non-Gaussian. Non-Gaussian distributions arise frequently in the natural sciences, and Gaussian distributions exist also in the social sciences. The point is that the dominant number of continuous distributions observed in the natural sciences are Gaussian, and many distributions observed in the social sciences are non-Gaussian [163].
Many distributions in the social sciences are non-Gaussian. Several important consequences of this are as follows.
-
1.
Heavy tails. The tails of non-Gaussian distributions are larger than the tails of Gaussian distributions. Thus the probability of extreme events becomes larger, and the moments of the distribution may depend considerably on the size of the sample. Then the conventional statistics based on the Gaussian distributions may be not applicable.
-
2.
The limit distribution of the sample means for large values of the mean is proportional (up to a slowly varying term) to the Zipf distribution (and not to the Gaussian distribution). This is the statement of the Gnedenko–Doeblin theorem.
-
3.
In many natural systems, the distribution of the values of some quantity is sharply concentrated around its mean value. Thus one can perform the transition from a probabilistic description to a deterministic description. This is not the case for non-Gaussian distributions. There is no such concentration around the mean, and because of this, a probabilistic description is appropriate for all problems of the social sciences in which non-Gaussian distributions appear.
There exist differences between the objects and processes studied in the natural and social sciences. Several of these differences are as follows.
-
1.
The number of factors. The objects and processes studied in the social sciences usually depend on many more factors than the objects and processes studied in the natural sciences. Let us connect this to the non-Gaussian distributions in the social sciences [164]. Let y be a variable that characterizes the influences on the studied object. Let n(y)dy be the number of influences in the interval \((y, y+dy)\). Then n(y) is the distribution of the influences. In order to define (a discrete) factor, we separate the area of values of y into subareas each of width \(\varDelta y\). Then if the area of values of y has length L, the number of factors will be \(L/\varDelta y\). Thus n(y) has now the meaning of a distribution of factors. This distribution is Gaussian in most cases in the natural sciences and non-Gaussian in many cases of the social sciences. As we have mentioned above, the non-Gaussian distributions are not very concentrated around the mean value as compared to the Gaussian distributions. In other words, many more factors have to be taken into account when one analyzes items or subjects that are described by non-Gaussian distributions. Thus the analysis of many kinds of social objects or processes must be a multifactor analysis.
-
2.
Dominance of parameters. In the case of systems from the natural sciences, usually there are several dominant latent parameters. In the case of social systems, usually there is no dominant latent parameter. The links among parameters are weak, and in addition, many latent parameters can be important.
-
3.
Subjectivity of the results of measurements. The measurements in the study of social problems must be made very carefully. The main reasons for this are as follows: the measured system often cannot be reproduced; the researchers can easily influence the measurement process; the measurement can be very complicated.
-
4.
Mathematics should be applied with care. The quantities that obey the laws of arithmetic are additive. There are two kinds of measurement scales that are used in the social sciences, and only one of them leads to additive quantities in most cases (i.e., to quantities that can be successfully studied by mathematical methods): closed measurement scales and open measurement scales. The closed measurement scales have a maximum upper value. Such a scale is, for example, the scale of school-children’s grades. Closed scales may lead to nonadditive quantities. The open measurement scales do not have a maximum upper value. Open scales lead in most cases to additive quantities. The measurement scales in the natural sciences are mostly open scales. Thus mathematical methods are generally applicable there. Open scales must be used also in the social sciences if one wants to apply mathematical methods of analysis successfully. The application of mathematical methods (developed for analysis of additive quantities) to nonadditive quantities may be useless. One can also use closed measurement scales, of course. The results of these measurements, however, have to be analyzed mostly qualitatively.
10 Several Notes on Scientometrics, Bibliometrics, Webometrics, and Informetrics
The term scientometrics was introduced in [44]. Scientometrics was defined in [44] as the application of those quantitative methods which are dealing with the analysis of science viewed as an information process. Thus fifty years ago, scientometrics was restricted to the measurement of science communication. Today, the area of research of scientometrics has increased. This can be seen from a more recent definition of scientometrics:
Scientometrics is the study of science, technology, and innovation from a quantitative perspective [165–170].
In several more words, by means of scientometrics one analyzes the quantitative aspects of the generation, propagation, and utilization of scientific information in order to contribute to a better understanding of the mechanism of scientific research activities [171]. The research fields of scientometrics include, for example, production of indicators for support of policy and management of research structures and systems [172–177]; measurement of impact of sets of articles, journals, and institutes as well as understanding scientific citations [178–189]; mapping scientific fields [190–192]. Scientometrics is closely connected to bibliometrics [193–201] and webometrics [202–210]. The term bibliometrics was introduced in 1969 (in the same year as the definition of scientometrics in [44]) as application of mathematical and statistical methods to books and other media of communication [211]. Thus fifty years ago, bibliometrics was used to study general information processes, whereas (as noted above) scientometrics was restricted to the measurement of scientific communication. Bibliometrics has received much attention [212–215], e.g., in the area of evaluation of research programs [216] and in the area of analysis of industrial research performance [217]. Today, the border between scientometrics and bibliometrics has almost vanished, and the the terms scientometrics and bibliometrics are used almost synonymously [218]. The rapid development of information technologies and global computer networks has led to the birth of webometrics. Webometrics is defined as the study of the quantitative aspects of the construction and use of information resources, structures, and technologies on the Web, drawing on bibliometric and informetric approaches [209, 210]. Informetrics is a term for a more general subfield of information science dealing with mathematical and statistical analysis of communication processes in science [219, 220]. Informetrics may be considered an extension of bibliometrics, since informetrics deals also with electronic media and because of this, includes, e.g., the statistical analysis of text and hypertext systems, models for production of information, information measures in electronic libraries, and processes and quantitative aspects of information retrieval [221, 222].
Many researchers have made significant contributions to scientometrics, bibliometrics, and informetrics. We shall mention several names in the following chapters. Let us mention here the name of Eugene Garfield, who started the Science Citation Index (SCI) in 1964 at the Institute for Scientific Information in the USA. SCI was important for the development of bibliometrics and scientometrics and was a response to the information crisis in the sciences after World War II (when the quantity of research results increased rapidly, and problems occurred for scientists to play their main social role, i.e., to produce new knowledge). SCI used experience from earlier databases (such as Shepard’s citations [223, 224]). In 1956, Garfield founded the company Eugene Garfield Associates and began publication of Current Contents, a weekly containing bibliographic information from the area of pharmaceutics and biomedicine (the number of covered areas increased very rapidly). In 1960, Garfield changed the name of the company to Institute of Scientific Information. Let us note that the success of the Current Contents was connected to the use of Bradford’s law for “scattering” of research publications around research journals (Bradford’s law will be discussed in Chap. 4 of the book) [225]. According to the Bradford’s law, the set of publications from some research area can be roughly separated into three subsets: a small subset of core journals, a larger subset of journals connected to the research area, and a large set of journals in which papers from the research area could occur. Bradford’s law was used in the selection of journals contributing to the multidisciplinary index SCI. In the following years, the SCI and ISI became the world leaders in the area of scientific information. This position remained unchallenged for almost fifty years, even after the rise of the Internet.
Below we consider three topics from the area of scientometrics that are of interest for our discussion. These topics are:
-
1.
Quantities that may be analyzed in the process of study of research dynamics;
-
2.
Inequality of scientific achievements;
-
3.
Knowledge landscapes.
10.1 Examples of Quantities that May Be Analyzed in the Process of the Study of Research Dynamics
Below we present a short list of some quantities, kinds of time series, and other units of data that may be used in the process of assessment of research and research organizations. The list is as follows.
-
1.
Time series for the number of published papers in groups of journals (for example in national journals).
-
2.
Time series for the total number and for the percentage of coauthored papers [226]. Coauthorship is an important phenomenon, since the development of modern science is connected to a steady increase in the number of coauthors, especially in the experimental branches of science. Coauthorship contributes to the increase of the length of an author’s publication list, and this length is important for the quality of research [227], for a scientific career, and for the process of approval of research projects.
The percentage of coauthored publications varies in the different sciences. In the social sciences, it is very low, and in the natural sciences it can reach 90 % and even more. There are interesting notes of Price and Rousseau with respect to coauthorship [228, 229]. Price notes that important factors for the growth of coauthorship of publications are (i) the expansion of the material base of scientific research, e.g., new equipment stimulates coauthorship; (ii) in times of expansion, the number of very good scientists increases at a slower rate than the number of scientists. In such conditions, the most productive authors increase their productivity further by becoming leaders of scientific collectives. In these collectives, scientists can be found who want to have publications but are unable to publish alone (because they are inexperienced PhD students, for example). Let us note here that in recent years, one observes frequently the phenomenon of hyperauthorship (a very large number of coauthors of a publication) [230].
-
3.
Network analysis of coauthorship groups [231–240] and especially detection of dense and very productive coauthorship networks : “invisible colleges” [241–246]. An invisible college has a core and periphery. The core usually consists of researchers from the same research structure or from a few research structures, e.g., from the same research institute or from several universities where productive groups exist.
- 4.
-
5.
Time series for the number of patents and discoveries. What can be expected in times of fast growth of the number of scientific discoveries is that their period of doubling is about ten years [249].
-
6.
Distribution of publications among research organizations [250].
-
7.
Distribution of patents and discoveries among the countries from a group of countries (for example, EU countries or the entire world).
-
8.
Statics and dynamics of landscapes of scientific discoveries and engineering patents for different scientific or engineering fields.
-
9.
Time series for the number of scientists (in a country). When a country’s research structures grow, one may expect doubling of the number of researchers every fifteen years. When the scientific structure becomes mature, the growth slows and may come to a halt.
-
10.
Territorial distribution of scientists—national and international [251, 252]. Distribution of scientists with respect to their qualifications.
-
11.
Dynamics of the age structure of scientists at the national level and comparison of the dynamics among countries from a group of countries.
Other kinds of quantities are connected to another important characteristic of research work: the citations of research publications [253–260]. One may analyze:
-
1.
Time series for citations of individual scientists, scientific groups, or scientific organizations [261–268]. We note that the number of citations depends on the number of researchers who work in the corresponding scientific area [267], and there can be also negative citations of the publications of a researcher. Citation analysis allows us to identify different categories of researchers such as identity-creators and image-makers [269]. The number of citations depends on the rate of aging of research information [270]. This rate of aging may be different for different scientific disciplines.
-
2.
Distribution of journals with respect to the citations of the papers in these journals (the impact factor is one possible indicator that can be constructed on the basis of such studies [271]).
-
3.
Distribution of scientific organizations with respect to the citations of the publications of the organization. One has to be very careful here, since in some areas of science there are many more citations than in other areas of science.
-
4.
Citation networks [272–276]. Usually there are subnetworks of leading scientists in some scientific areas, and every leading scientist cites predominantly the other leading scientists. The nonleading scientists cite the leading scientists much more than other nonleading scientists.
-
5.
Distribution of scientists with respect to the number of citations of their publications. Here different possibilities exist, e.g., the study of the distribution of citations of the most cited papers of scientists from a scientific group or scientific organization or the study of the distribution of the number of citations of the papers that contribute to the h-factors or g-indexes of the researchers from the assessed research groups or research organizations.
-
6.
Distribution of publications of a scientific group or scientific organization with respect to the number of citations they have.
-
7.
Distribution of citations among scientific fields [277, 278].
-
8.
Distribution of the time interval between appearance of a publication and its first citation.
-
9.
Landscapes of citations [279, 280] with respect to scientific discipline; countries; kind of publications; research organizations in a country, etc.
-
10.
Distributions of self-citations for scientific disciplines and in research groups and research organizations.
In addition, one may analyze other characteristics of science dynamics such as interdisciplinarity of scientific journals on the basis of the betweenness centrality measure used in social networks analysis [281]; aging of scientific literature [282, 283]; dynamics of scientific communication [284], etc.
10.2 Inequality of Scientific Achievements
Different researchers have different scientific achievements. Many factors influence the achievement of individual researchers or group of researchers. If we consider individual researchers, four main factors may be considered [218]: the subject matter; the author’s age; the author’s social status; the observation period. Experienced researchers usually have larger scientific production and larger scientific achievements in comparison to the newcomers without research experience. Chemists usually have larger research production than mathematicians. An established professor with high social status usually has a much larger network of collaborations and contributes to more scientific achievements in comparison to a young researcher without such social status.
One of the tasks of the assessment of research organizations is the measurement of the inequality of scientific achievements of researchers and research collectives. One may use the notion of Coulter [285 ] that the distribution of some characteristics of research productivity is the division of the units of these characteristics among the components of the corresponding structure of the research organization. Inequality then may be defined as variation of the above division. Observation of large research groups shows that researchers are distributed usually within three classes with respect to a part of their production (the part that can be measured by the number of authored and coauthored papers in the international databases such as ISI Web of Science or SCOPUS): a small class of very productive scientists; a large class of very unproductive scientists; a large class of scientists who fall between the above two extreme classes.
In the study of inequality in research organizations, one may use not only quantitative methods but also qualitative methods such as expert evaluations [286–288], surveys, and content analysis. We shall discuss numerous indexes of inequality of scientific achievements in the following chapters and especially in Chap. 4. Now we shall focus our attention on the concept of knowledge landscape.
10.3 Knowledge Landscapes
An understanding of the evolution of research organizations requires research complementary to mathematical investigations [289–302]. Very useful tool for such research are knowledge maps [303–325] and knowledge landscapes [326–333]. They may be used for identification of potential collaborators [334]; for creation of document-level maps of research fields [335]; for international comparisons of research systems [336]; for modeling of science [337]; etc.
The concept of knowledge landscape is as follows: Describe the corresponding field of science or technology through a function of parameters such as height, weight, size, technical data, etc. Then a virtual knowledge landscape can be constructed from empirical data in order to visualize and understand innovation and other processes in science and technology.
One example of a technological knowledge landscape can be given by the function \(E=E(S,v)\), where E are the expenses for developing a new car, S and v being the size and velocity of the car. The landscape is constructed as the values of S and v are plotted on the horizontal axes, and the values of E are plotted on the vertical axis.
Knowledge landscapes may be used for evaluation and tracking evolution of research structures and systems [338]. Two selected examples are:
-
1.
Application of knowledge landscapes for evaluating national research strategies. The national research structures can be considered to be groups of researchers who compete for scientific results following optimal research strategies. The efforts of each research structure or the efforts of each country become visible, comparable, and measurable by means of appropriate landscapes connected, for example, to the number of publications. The aggregate research strategies of a country can thereby be represented by the distribution of publications in the various scientific disciplines. In so doing, within a two-dimensional space, i.e., axes being the scientific disciplines and number of publications, different countries occupy different locations. Various political discussions can follow, and evolution strategies invented thereafter. In addition, one can track scientific areas of strategic importance on the basis of journal mappings [339] or can construct global maps of science [340] that can be very useful as elements of national research decision support systems.
-
2.
Landscapes for research evaluation based on scientific citations. Citations are important in the retrieval and evaluation of information in scientific communication systems [341–344]. This is based on the objective nature of the citations as components of a global expert evaluation system, as represented by the Science Citation Index. Thus the importance of the citation landscapes increases in the process of formation of a research policy. One example of this is personnel management decisions, which influence individual research careers or investment strategies.
11 Notes on Research Production and Research Productivity
Researchers produce numerous kinds of items as a result of their work: research publications, technical reports, patents, etc. This research production may be counted, and the corresponding numbers can be divided by units of time (month, year, etc.). The obtained numbers are characteristics of the researcher’s research productivity. The values of the characteristics of the research productivity depend on the considered time interval. If the time interval for calculation of the characteristics of the research productivity is the same as the career length of the researcher, then the characteristics of the research productivity coincide with the characteristics of the research production of the researcher.
Research structures and their system of functioning are extremely important for research productivity [345–349]. Research organizations provide the material conditions for research work [350], but they also support research environment that may stimulate or may influence negatively the research work [351]. A part of this environment are the interactions among researchers. These interactions influence the external motivation for research work (external means that the source of motivation is rooted in the actions of other persons). In addition, the above interactions influence the internal motivation for research work (internal means that the source of motivations is rooted in the ideas of the researcher). It is observed that the more productive researchers search for and establish more connections with colleagues. Thus the presence of at least one productive researcher in any of the small units of a research organization (laboratories, sections, or chairs) creates more stimuli for research work.
Research productivity is age-dependent [352–355]. In most cases, the research productivity of an individual researcher decreases as he/she ages and comes close to retirement age. In most cases, research productivity of a research organization decreases with increasing average age of the researchers in the organization (if the average age is large enough). There exist other factors that may affect research productivity [356]. Eleven of these factors [357] are persistence, resource adequacy (adequate funds for research, adequate equipment, etc.), access to literature, professional commitment, intelligence, initiative, creativity, learning capability, concern for advancement, stimulative leadership, and external orientation (adequate contacts with superior researchers and participation in seminars and conferences).
Research production is an important quantity [358] with a complex structure. There exist research collectives, and these research collectives produce publications (but not only publications). Resources are spent for organizing researchers into structures and for supporting a system of functioning for those structures. One result of the actions of researchers is the set of their research publications, which is an important quantitative measure of the scientific productivity. If additional analysis of the content of the publications is made, the set of publications can be used also as a qualitative measure of research productivity. The connection between the research structures and the set of publications is relatively complicated for two reasons: there are scientists who don’t have publications, and some publications are produced by more than one author. It is very interesting [143] that s large number of researchers (about \(50\,\%\)) do not have publications. This will be further discussed in Chap. 4. But the absence of publications does not mean that the corresponding researchers are lazy or incompetent. It may mean that their contribution to the publications is not large enough for their names to be included in the list of coauthors of the publication. In order to evaluate the scientific production of such researchers, one may have to use units that are smaller than a research publication. Such a smaller unit may be, for example, the time spent for support of research work, the number of collaborations with researchers who are authors of publications, etc. The current units used widely for measurement of scientific performance are (i) unit for information: research publication (scientific paper); (ii) unit for impact of the unit for information: a citation of the research publication. It seems that these units are too crude or at least they are too large to measure the research performance of large classes of researchers. Then for the low-productive (in terms of research publications) researchers, one may think about other units for measurement of their performance, and research papers and their citations may be used to measure performance of the top class of researchers. The research community is greatly interconnected, and it needs all kind of “workers”: full “workers” who produce scientific publications and partial “workers” who support the full “workers” in the process of producing the scientific publications.
Statistical methods are much used for evaluation of research work [359]. One has to be careful in applying such methods, because the number of components of a scientific organization may be large, but it is not very large in comparison with the number of molecules in a glass of water, for example. In addition, in contrast to molecules (which are weakly connected), the elements of a research organization may be strongly connected, and these connections are an important part of their structure and system of working. Thus one has to be careful about the direct application of methods from thermodynamics (such as temperature or entropy) to scientific systems and organizations. Such methods may be useful, but they reflect only some of the properties of the research organizations, since a research organization is not only a statistical collection of elements each of which is in random motion. For example, in addition to the random processes in research organizations, there exist also deterministic processes that may lead to the rise of complex research structures.
This book is focused on research publications and their citations. Because of this, we shall discuss below in more detail the importance of research publications for assessment of research production. The importance of citations of research publications for assessment of research production will be demonstrated in Chap. 2, where we shall discuss numerous indexes for evaluation of research production constructed on the basis of citations of research publications.
Research production has the form of books, monographs, reports, theses, articles in journals, etc. For the purposes of scientometric studies, papers published in refereed scientific journals seems to be the most suitable unit (at least today and for some time in the future). Usually, scientific papers are not subdivided into smaller units (this may cause some problems, since as the scientific contribution of large groups of research personnel is not enough to put the names of the corresponding researchers as coauthors of research articles). Other elements used in the evaluation of research are, e.g., the number of citations of a paper, the number of references, the number of coauthors of the paper, etc. The number of publications of a researcher may be used as an indicator of latent characteristics of researchers and scientific organizations such as prestige of the researcher; prestige of the research organization of the researcher; contribution to science of researchers or research organization; productivity of researcher and research organization; eliteness of the researcher or research organization. The speed of increase of the number of research publications can be used as an indicator for the currency of a scientific research area; perspectives of a scientific research area; phase of development of a research group or organization (at the initial phase of development, there is a rapid increase in the number of publications. Then the speed of increase of publications falls, and finally one observes maturity of the research group or organization with an almost constant number of research publications per year).
A significant number of publications are published jointly, and the tendency for joint publications increases with the increasing complexity of scientific research. But how to count joint publications? There are many ways to do this. The three most popular of them are [360]:
-
1.
Normal count: Every author of a publication has received full credit. For an example, if a publication has four authors, it counts as one publication for every author [266].
-
2.
Straight count: Only the first author receives credit for the publication. This count discriminates against the second and subsequent authors.
-
3.
Adjusted count: Every author of a publication receives an equal fraction of the total credit of one unit [361]. For example, if a publication has four authors, every author is given one-fourth credit. By this measure, the relative contribution of every author to the article is ignored.
Several additional formulas for counting joint publications are as follows (N is the number of coauthors in all relationships below and \(S_n\) is the score assigned to the nth author)
-
HCM-count (Howard–Cole–Maxwell) [362 ]
$$\begin{aligned} S_n = \frac{(3/2)^{N-n}}{\sum \limits _{n=1}^N (3/2)^{n-1}}. \end{aligned}$$(1.2) -
EKW-count (Ellwein–Khahab–Waldman) [363 ]
$$\begin{aligned} S_n = \frac{b^{n-1}}{\sum \limits _{n=1}^N b^{n-1}}. \end{aligned}$$(1.3)In [363], \(b=0.8\).
-
LV-count (Lukovits–Vinkler) [364 ]
$$\begin{aligned} S_1 = \frac{N+1}{2 N F}; \ \ S_n = \frac{n + T}{2 n F T}, \end{aligned}$$(1.4)where
$$ T = \frac{100}{A}; \ \ F = \frac{1}{2} \left( \frac{1}{N} + \frac{N-1}{T} + \sum \limits _{n=1}^N \frac{1}{n}\right) $$and A is the authorship threshold (percentage as the lowest share of contribution to a paper: five or ten percent of total credit).
-
TG-count (Trueba–Guerrero ) [365 ]
$$\begin{aligned} S_n = \frac{2(2N-n+2)}{3N(N+1)}(1-f) + C_n f, \end{aligned}$$(1.5)where f is the share for crediting favored authors (usually favored authors are the first, the second, and the last author) \(0< f < 1\); \(C_n\) is the rate of favoring the nth author \(\sum \limits _{i=1}^N C_n=1\).
There are additional systems for distribution of scores among coauthors [366, 367]. The problem of coauthorship will be discussed again in Chap. 2 in connection with the h-index.
The following additional indicators based on publications may be used for assessment of publication activity of single researchers, research groups, or research organizations:
-
1.
Distribution of number of publications in research organizations or in research groups of a research organization.
-
2.
Distribution of publications in research journals.
-
3.
Distribution of researchers with respect of number of publications for the three kinds of counts mentioned above.
-
4.
Distribution of publications with respect to gender of the authors.
-
5.
Distribution of publication with respect to language.
-
6.
Distribution of publications with respect to their kind (articles, papers in conference proceedings, book chapters, books, etc.).
-
7.
Distribution of publications with respect of the kind of research work (theoretical publications, experimental publications, reviews, etc.).
Finally, let us note here an interesting effect occurring when funding is linked to publication counts [368]. In this case, the publication numbers may jump dramatically, but with the highest percentage increase in the lower-impact journals. And the jump is larger as a percentage in universities than in government research institutes, which could mean that there is an unused research potential in universities, whereas research institutes already produce many publications and can’t increase their percentage by as much as is the case for universities.
12 Notes on the Methods of Research Assessment
12.1 Method of Expert Evaluation
Assessment of the production and productivity of researchers can be made by expert evaluation of the work of any researcher, e.g., for the last three or five years. The method of expert evaluation judges mainly the quality of the work of a researcher. One variant of realization of this method is to use a commission of five to seven experts. These experts evaluate a researcher with respect to two criteria: contribution of researchers to the corresponding area of science (external criterion) and importance and usefulness of researchers for his/her scientific organization (internal criterion). The commission of experts ranks researchers with respect to the above two criteria. After the ranking, the ranks of the researchers can be converted to points according to some appropriate conversion schema [143]. Evaluated researchers are divided into groups, since the productivity of a researcher depends on the scientific environment (conditions of work and relationships among the scientists from the scientific group to which the scientists belong) and on its status in the scientific organization. There are five groups of researchers:
-
1.
Researchers without doctoral degree who perform assistant work in the research or applied units of the scientific organization.
-
2.
Researchers without doctoral degree who perform research work in the applied units of the research organization.
-
3.
Researchers without doctoral degree who perform research work in the research units of the research organization.
-
4.
Researchers with doctoral degree who perform research work in the applied units of the research organization.
-
5.
Researchers with doctoral degree who perform research work in the research units of the research organization.
There can be subgroups of these groups. For example, the researchers with doctoral degrees who perform research work in the research units of the research organization can be assistant professors, associate professors, and full professors.
Expert evaluation may be a part of the complex evaluation of a researcher. Such a complex evaluation may contain, for example [369]:
-
1.
Evaluation of production by number of publications, number of internal and external scientific reports, number of developed methods, discovered effects or other significant achievements of the researcher.
-
2.
Further evaluation of production by total number of pages of publications, reports, manuscripts, etc.
-
3.
Evaluation of importance of produced knowledge by number of citations. An important problem here is the scaling of citations from different years or different subject categories [370, 371].
-
4.
Evaluation by prestige of the journals and publishing houses that published the articles, book chapters, or books of the scientist.
-
5.
Evaluation on the basis of obtained national and international awards.
-
6.
Expert evaluation of the significance of the achievements for solving real problems.
-
7.
Expert evaluation on the influence of the scientist on the basis of the following categories: influence in the area of research, influence in the corresponding scientific discipline, influence in other scientific disciplines.
-
8.
Evaluation of influence by the number of colleagues from the scientific organization who think that this researcher is a high-quality scientist.
-
9.
Further evaluation of influence on the basis of number of publications that are judged as important by the colleagues, and number of times each of the important publications is pointed to as important by the colleagues of the researcher.
In addition to the above evaluation, there may also be an evaluation of the distribution of the time spent by researcher as follows [143]: time for the main scientific work (time for the research work of the scientist, time for scientific research of general interest that leads to the solution of large classes of problems, time for scientific research of special interest that leads to the solution of specific problems, time for transfer of scientific research for improving products or processes or for obtaining new kinds of products or processes); time for directing research work of other researchers; time for collaboration with other researchers; time for consultations and expert evaluations; time for pedagogical work; time for administrative work (time for internal administrative work within the scientific unit (laboratory, section or chair), time for communication with the higher levels of the administrative hierarchy, time for relations with other scientific groups and clients). The list of evaluations may be enlarged, e.g., by numerous indexes presented in Chap. 2.
12.2 Assessment of Basic Research
Assessment of basic research is a problem for research administrators, since there are no simple measures of the contribution to scientific knowledge made by researchers. Many partial indicators exist, and each of them accounts for one (or several) factors that influence the basic research and is influenced by other factors that are not connected to the basic research. Thus in order to obtain reliable results, one has to minimize the influence of the factors that are not connected to research [372]. This can be done on the basis of the concept according to which basic research is considered as a process with inputs, process body (scientific production), and outputs [373]. The elements of the process of basic research are
-
1.
Inputs: stock of scientific knowledge and existing techniques; financial resources; institutional scientific resources (scientific instruments, skilled personal, etc.); recruited personnel (untrained students, etc.); environmental conditions (such as diverse natural influences).
-
2.
Scientific production: conceptual, experimental and technical work of scientists; support work by engineers, technicians, etc.; dissemination of research results; education work for development and reproduction of scientific skills (training young scientists, etc.); administrative support work (including organizing adequate inputs).
-
3.
Outputs: scientific contributions to the discipline, new techniques and new scientific knowledge; scientific contributions to other areas of science; educational contribution: trained students, PhD students, and scientists; economic contribution (such as engineers and workers with increased skills for industry, technological spin-offs, commercial benefits for equipment suppliers, etc.; cultural contributions.
In order to minimize the influence of factors not connected to basic research, Martin and Irvin [372] have proposed the method of converging partial indicators for assessment of basic research. This method is based on the following five principles:
-
1.
Indicators are applied to research groups rather than to the individual scientists;
-
2.
Citation-based indicators are seen as reflecting the impact (and not the quality or importance) of the research work [374, 375];
-
3.
A range of indicators is used, each of which focuses on different aspects of the group performance;
-
4.
As far as possible, indicators are applied to a matched group (comparing like with like principle)
-
5.
As the indicators used have an imperfect or partial nature, only in those cases where they yield convergent results can it be assumed that the influence of peripheral factors has been kept small. In these cases, it can be assumed that the indicators provide a reliable estimate of the contribution of the different groups to scientific progress.
An algorithm for scientometric assessment that may be used also for assessment of basic research has been proposed by Moravcsic [376]. This algorithm includes
-
1.
specifying the purpose of the assessment;
-
2.
specifying the system to be assessed;
-
3.
deciding on the level of assessment;
-
4.
setting criteria;
-
5.
selecting methods and specifying indicators for each criterion;
-
6.
determining the links among the components within and outside the system (scientific political issues, type of the subject field and activity, etc.);
-
7.
carrying out measurements;
-
8.
interpreting the results obtained;
-
9.
drawing conclusions of the assessment.
The evaluators should combine the scientometric assessment and peer reviews [377, 378] in order to obtain a complete picture of the evaluated research group or research organization [379–384].
Applications of indicators for assessment of basic research is related to certain problems. Several of these problems are as follows [372]: (i) the contribution of each publication to the scientific knowledge is different; (ii) publication rates are different for different research fields. Four important problems connected to indicators based on citation analysis are [385, 386]: (i) technical problems with databases such as authors with identical names, variation of names, incomplete coverage of journals; (ii) variation of citation rate of a paper during its lifetime; (iii) Presence of critical citations or halo-effect citations (Halo effect means that a researcher’s overall impression about other researchers influences the observer’s feelings and thoughts about the properties of their publications); (iv) self-citations and in-house citations. Finally, there are problems connected to peer evaluation, e.g.: (i) individuals evaluate scientific contributions on the basis of their cognitive and social status (which can be quite different); (ii) perceived implication of results of one’s own center and competitors may affect evaluation; (iii) conformist assessments (for example, triggered by halo effect or by lack of knowledge about the contributions of different centers).
12.3 Evaluation of Research Organizations and Groups of Research Organizations
Evaluation of research and technology programs [387], research organizations, and groups of research organizations becomes increasingly important especially when limited resources for research should be distributed. Below we mention three examples of such systems used in continental Europe.
-
1.
The SEP system of The Netherlands
SEP (Standard Evaluation Protocol) has two key objectives: (i) to improve research quality based on an external peer review, including scientific and societal relevance of research, research policy and management; (ii) to be accountable to the board of the research organization, and toward funding agencies, government, and society. SEP has two levels: evaluation of the research institute as a whole, and evaluation of specific research groups or programs. The criteria for evaluation are quality, productivity, (social) relevance, and vitality and feasibility. Each of these criteria contains several subcriteria.
-
2.
AERES system (France)
The Evaluation Agency for Research and Higher Education (AERES) (www.aeres-evaluation.com) is an independent administrative authority whose task is to evaluate French research organizations and institutions, research and higher education institutions, scientific cooperation foundations and institutions as well as the French National Research Agency. Usually each year, 25 % of all institutions are evaluated, and thus the national evaluation cycle is four years. The evaluation criteria of AERES are: scientific quality and output; academic reputation and drawing power; interactions with the social, economic, and cultural environment; organization and life of the institution; involvement in training by research; strategy and scientific prospects for the next period.
-
3.
Evaluation of national research policy: OECD indicators
Various systems of indicators for evaluation of national research policy can be constructed (for example, see [388, 389]). One example is the system of eight categories of indicators applied by OECD for benchmarking of national research policies [390]. These categories of indicators are: human resources in research and technology development; public and private investment in research and technology development; scientific and technological production; impact of research and technology development on economic competitiveness and employment; human resources, knowledge creation; transmission and application of knowledge; innovation finance, output, and market.
13 Mathematics and Quantification of Research Performance. English–Czerwon Method
Egghe [391] discusses performance in twelve diverse information production systems. Below, we shall focus our attention on a mathematical method for quantification of research performance: the English-Czerwon method that may be used for assessment of research performance within the scope of the following kinds of information production systems discussed in [391]: papers—citations system; authors—publications system; authors—citations system. The English–Czerwon method [392] is an interesting example of a combination of evaluation on the basis of quantitative indicators with evaluation on the basis of peer review. Its description is as follows.
Suppose there are k research units, and let the performance of each unit for the nth year of evaluation be denoted by \(p_i(n)\ge 0\). The performance has two components,
where:
-
\(o_i(n)\) is the “objective” part of the performance, calculated on the basis of quantitative indicators;
-
\(s_i(n)\) is the “subjective” part of performance, calculated on the basis of the judgment of all evaluated units about the research of the ith unit for the nth year of evaluation.
There are two variants of the method: (i) weighting without accounting for the current performance; (ii) weighting with accounting for the current performance.
13.1 Weighting Without Accounting for the Current Performance
In this variant of the method, the “subjective” part of the performance is calculated as follows:
where the weight \(w_j(n)\) is defined as \(w_j(1)=o_j(1)\) for the first year of evaluation, and
for the years \(n>1\). We note that in this variant of the methodology, the performance for the current year \(p_i(n)\) is not taken into account. The methods for calculation of the parameters \(o_j\), \(q_{ij}\), and \(w_j(n)\) will be discussed in a separate paragraph below.
13.2 Weighting with Accounting for the Current Performance
The performance in the current year \(p_i(n)\) is taken into account by defining the weight as follows:
Let us substitute (1.9) in (1.6). Taking into account (1.7), we obtain an equation for \(p_i(n)\) as follows:
We have the following relationship from (1.9):
The substitution of (1.11) in (1.10) leads to the following equation for \(p_i(n)\):
We note that (1.12) defines a system of equations for \(p_i(n)\), and this system still has to be solved.
The solution of (1.12) can be presented in matrix form. Let I be the identity matrix (which contains 1 in all diagonal positions and 0 elsewhere) and Q the matrix whose elements are the evaluations \(q_{ij}\). Let the vectors \(\mathbf {p}(n)\), \(\mathbf {o}(n)\), and \(\mathbf {\overline{w}}(n-1)\) have components \(p_i(n)\), \(o_i(n)\), and \(\overline{w}_i(n-1)\) respectively. Then the solution of (1.12) is
In order to use this solution, we have to impose an additional requirement on the matrix Q. This requirement has to ensure positive values of \(p_i(n)\). The requirement is that the eigenvector of Q with positive elements have an eigenvalue less than n. Then each eigenvalue is less than n (Frobenius–Perron theorem), and the time series expansion of \((nI - Q)^{-1}\) (where the exponent \(-1\) means inversion of the matrix \((nI - Q)\)) converges to a matrix with nonnegative matrix elements. This will ensure positive values of \(p_i(n)\). For small values of the elements of Q, we have the approximation
and the solution (1.13) for \(\mathbf {p}(n)\) becomes
13.3 How to Determine the Values of Parameters
In order to use the two variants of the English–Czerwon method, we have to determine the values \(o_i(n)\) of the “objective” part of the performance as well as the values of the evaluation coefficients \(q_{ij}\) from the “subjective” part of the performance. The simplest way to set the values of \(o_i\) is just to consider the value of one indicator: the number of citations of the papers of the research organization for the current year, for example. Of course, more that one indicator can be incorporated in \(o_i\) by means of appropriate weighting.
The values of \(q_{ij}\) can be set as follows. We note that these values have to be small (in order to satisfy the assumption for small Q on which basis we obtained the relationship (1.15) above). The determination of the values can be made as follows:
-
1.
Evaluation of performance by ranking research organizations.
One takes several (five to ten) leading scientists from each research organization and asks each of them to rank the research organizations with respect to their performance. The ranking is between rank 1 and rank k. Then the average rank \(\overline{r}_{ij}\) is calculated by averaging the assigned ranks to the ith organization from the scientists of the jth organization:
$$\begin{aligned} \overline{r}_{ij} = \frac{1}{L} \sum \limits _{j=1}^{L} r_{ij}, \end{aligned}$$(1.16)where L is the number of evaluating scientists.
-
2.
Calculation of \(q_{ij}\).
The \(q_{ij}\) are calculated as follows:
$$\begin{aligned} q_{ij}=\frac{k-\overline{r}_{ij}}{M}, \end{aligned}$$(1.17)where M is a large number (of order of 5k or larger), which ensures that the values of \(q_{ij}\) are small.
-
3.
Obtaining \(p_i(n)\).
The \(p_{i}(n)\) are then obtained on the basis of (1.15). The final step is to perform a normalization
$$\begin{aligned} \overline{p}_i(n) = \frac{p_i(n)}{\sum \limits _{j=1}^k p_i(n)}, \end{aligned}$$(1.18)and one can rank the institutions with respect to the values of \(\overline{p}_i\) for the corresponding year.
14 Concluding Remarks
At the end of Part I of this book, the reader already may have a impression about the complexity of science and research organizations; about the importance of science for society; about the features of research production and non-Gaussianity of some statistical characteristics of quantities used for assessment of research; about quantities, methods, and systems used for assessment of research and research organizations. Thus the reader is prepared to move to the world of quantities and models used for the study of science dynamics and for assessment of research, researchers, and research organizations.
References
J.L. Heilbron (ed.), The Oxford Companion to the History of Modern Science (Oxford University Press, New York, 2003)
Science. Wikipedia, the free encyclopedia
R.K. Merton, The Sociology of Science: Theoretical and Empirical Investigations (The University of Chicago Press, Chicago, 1973)
P. Vinkler, The Evaluation of Research by Scientometric Indicators (Chandos, Oxford, 2010)
J.D. Bernal, The Social Function of Science (The MIT Press, Cambridge, MA, 1939)
J.D. Bernal, Science and Industry in the Nineteenth Century (Routhledge, New York, 1953)
G. Böhme, N. Stehr, The growing impact of scientific knowledge on social relations, in The Knowledge Society, ed. by G. Böhme, N. Stehr (eds.) (Springer, Netherlands, 1986), pp. 7–29
R. Whitley, The Intellectual and Social Organization of the Sciences (Oxford University Press, Oxford, 2000)
R. Whitley, J. Gläser (eds.), The Changing Governance of the Sciences (Springer, Dordrecht, 2007)
Frascati Manual: Proposed Standard Practice for Surveys on Research and Experimental Development, 6th. edn. (OECD, 2002)
A. Kaufmann, F. Tödtling, Science-industry interaction: the importance of boundary-crossing between systems. Res. Policy 30, 791–804 (2001)
S. Weinberger, The evolving science of war. Nature 505, 156–157 (2014)
R. Lidskog, G. Sundqvist, When does science matter? International relations meets science and technology studies. Glob. Environ. Polit. 15 (in press, 2015). doi:10.1162/GLEP_a_00269
F. Narin, J. Davidson Frame. The growth of Japanese science and technology. Science 245(4918), 600 (1989)
C.S. Wagner, I. Branmakulam, B. Jackson, A. Wong, T. Yoda, Science And Technology Collaboration: Building Capacity in Developing Countries (RAND, 2001). MR-1357.0-WB
B. Parker, The Physics of War: From Arrows to Atoms (Prometheus Books, New York, 2014)
E.R. Gantman, Economic, linguistic, and political factors in the scientific productivity of countries. Scientometrics 93, 967–985 (2012)
M.E.D. Koenig, Information policy—the mounting tension (value additive versus uniquely distributable “public good”). J. Inf. Sci. 21, 229–231 (1995)
P.E. Stephan, The economics of science. J. Econ. Lit. 34, 1199–1235 (1996)
P.E. Stephan, How Economics Shapes Science (Harvard University Press, Cambridge, MA, 2012)
J.N. Cummings, S. Kiesler, Organization theory and the changing nature of science. J. Organ. Des. 3, 1–16 (2014)
M. Hirooka, Nonlinear dynamism of innovation and business cycles. J. Evolut. Econ. 13, 549–576 (2003)
B.R. Martin, The evolution of science policy and innovation studies. Res. Policy 41, 1219–1239 (2012)
F. Narin, K.S. Hamilton, D. Olivastro, Linkage between agency-supported research and patented industrial technology. Res. Eval. 5, 183–187 (1995)
B.R. Martin, P. Nightingale, A. Yegros-Yegros, Science and technology studies: exploring the knowledge base. Res. Policy 41, 1182–1204 (2012)
M. Hirooka, Innovation Dynamism and Economic Growth: A Nonlinear Perspective (Edward Elgar Publishing, Cheltenham, UK, 2006)
P. van den Besselaar, K. Börner, A. Scharnhorst. Science policy and the challenges for modeling science. p.p. 261 – 266 in A. Scharnhorst, K. Börner, P. van den Besselaar (eds.) Models for science dynamics (Springer, Berlin, 2012)
A. Smith, A. Stirling, F. Berkhout, The governance of sustainable socio-technical transitions. Res. Policy 34, 1491–1510 (2005)
A. Stirling, A general framework for analyzing diversity in science, technology and society. J. R. Soc. Interface 4, 707–719 (2007)
F. Berkhout, A. Smith, A. Stirling, Socio-tecgnical regimes and transition contexts, pp. 48–75 in System Innovation and the Transition to Sustainability ed. by B. Elzen, F.W. Geels, K. Green (Edward Elgar, Chentelham, 2004)
D.P. Gaver, V. Srinivasan, Allocating resources between research and development: a macro analysis. Manage. Sci. 18, 492–501 (1972)
A.D. Bender, E.B. Pyle III, W.J. Westlake, B. Douglas, Simulation of R&D investment strategies. Omega 4, 67–77 (1976)
S. Bretschneider, Operations research contributions to evaluation of R&D projects, pp. 122–153 in Evaluating R&D Impacts: Methods and Practice ed. by B. Bozeman, J. Melkers (Springer, US, 1993)
R.L. Schmidt, Recent progress in modelling R&D project-selection process. IEEE Trans. Eng. Manage. 39, 189–201 (1992)
L. Bornmann, Measuring the societal impact of research. EMBO Rep. 13, 673–676 (2012)
C.S. Wagner, The elusive partnership: science and foreign policy. Sci. Public Policy 29, 409–417 (2002)
M.A. Rappa, K. Debackere, Technological communities and the diffusion of knowledge. R&D Manage. 22, 209–220 (1992)
M.A. Rappa, K. Debackere, Technological communities and the diffusion of knowledge: a replication and validation. R&D Manage. 24, 355–371 (1994)
L. Bornmann, What is societal impact of research and how can it be assessed? A literature survey. J. Am. Soc. Inf. Sci. Technol. 64, 217–233 (2013)
A. Verbeek, K. Debackere, M. Luwej, E. Zimmermann, Measuring progress and evolution in science and technology I: the multiple uses of bibliometric indicators. Int. J. Manag. Rev. 4, 179–211 (2002)
A. Verbeek, K. Debackere, M. Luwej, E. Zimmermann, Measuring progress and evolution in science and technology II: the multiple uses of technometric indicators. Int. J. Manag. Rev. 4, 213–231 (2002)
L. Bornmann, W. Marx, How should the societal impact of research be generated and measured? A proposal for a simple and practicable approach to allow interdisciplinary comparison. Scientometrics 98, 211–219 (2014)
A.F.J. van Raan, Measurement of central aspects of scientific research: performance, interdisciplinarity, structure. Meas.: Interdiscip. Res. Perspect. 3, 1–19 (2005)
V.V. Nalimov, G.M. Mulchenko, Naukometriya (Nauka, Moscow, 1969). (in Russian)
C. Michels, U. Schmoch, The growth of science and database coverage. Scientometrics 93, 831–846 (2012)
F. Narin, K.S. Hamilton, D. Olivastro, The increasing linkage between US technology and public science. Res. Policy 26, 317–330 (1997)
J. Anderson, K. Williams, D. Seemungal, F. Narin, D. Olivastro, Human genetic technology: exploring the links between science and innovation. Technol. Anal. Strat. Manag. 8, 135–156 (1996)
F. Narin, D. Olivastro, Status report: linkage between technology and science. Res. Policy 21, 237–249 (1992)
M. Mayer, Tracing knowledge flows in innovation systems. Scientometrics 54, 193–212 (2002)
F. Narin, E. Noma, R. Perry, Patents as indicators of corporate technological strength. Res. Policy 16, 143–155 (1987)
M. Mayer, Patent citation analysis in a novel field of technology: an exploration of nano-science and nano-technology. Scientometrics 51, 163–183 (2001)
A. Verbeek, K. Debackere, M. Luwel, P. Andries, E. Zimmermann, F. Deleus, Linking science to technology: using bibliographic references in patents to build linkage schemes. Scientometrics 54, 399–420 (2002)
S. Bhattacharya, H. Kretdchmer, M. Mayer, Characterizing intellectual spaces between science and technology. Scientometrics 58, 369–390 (2003)
A.L. Porter, A. Thomas Roper, T.W. Mason, F.A, Rossini, J. Banks, Forecasting and Management of Technology (Wiley, New York, 1991)
R.J. Watts, A. Porter, Innovation forecasting. Technol. Forecast. Soc. Chang. 56, 25–47 (1997)
M. Meyer, Measuring science-technology interaction in the knowledge-driven economy: the case of small economy. Scientometrics 66, 425–429 (2006)
J. Dryden, Quantifying technological advance: S&T indicators at the OECD—challenges for the 1990s. Sci. Public Policy 19, 281–290 (1992)
E.C.M. Noyons, A.F.J. van Raan, H. Grupp, U. Schmoch, Exploring the science and technology interface: Inventor-author relations in laser medicine research. Res. Policy 23, 443–457 (1994)
A. Mathieu, M. Mayer, B. van Pottelberghe, de la Potterie, Turning science into business: a case study of a major European Research University. Sci. Public Policy 35, 669–679 (2008)
M. Mayer, Tracing knowledge flows in innovation systems—an infometric perspective of future research on science-based innovation. Econ. Syst. Res. 14, 323–344 (2002)
A. Klitkou, S. Nyagaard, M. Mayer, Tracking techno-science networks: a case study of fuel cells and related hydrogen technology. Scientometrics 70, 491–518 (2007)
A.L. Porter, M.J. Detampel, Technology opportunities analysis. Technol. Forecast. Soc. Chang. 49, 237–255 (1995)
J. McNerney, J. Doyne Farmer, S. Redner, J.E. Trancik, Role of design complexity in technology improvement. Proc. Natl. Acad. Sci. USA 108, 9008–9013 (2011)
G.S. McMillan, F. Narin, D.L. Deeds, An analysis of the critical role of public science in innovation: the case of biotechnology. Res. Policy 29, 1–9 (2000)
D. Hicks, A. Breitzman, K. Hilton, F. Narin, Research excellence and patented innovation. Sci. Public Policy 27, 310–320 (2000)
F. Narin, A. Breitzman, Inventive productivity. Res. Policy 24, 507–519 (1995)
M. Mayer, Does science push technology? Patents citing scientific literature. Res Policy 29, 409–434 (2000)
A. Hullman, M. Meyer, Publications and patents in nanotechnology. Scientometrics 58, 507–527 (2003)
B.G. van Vianen, H.F. Moed, A.F.J. van Raan, An exploraton of the science base of recent technology. Res. Policy 19, 61–81 (1990)
M. Ioannidis, A. Vatikalos, Cutting-edge information and telecommunication technologies meet energy: Energy management systems and smart web platforms, pp. 153 – 162 in Energy-Efficient Computing and Networking, ed. by N. Hatziargyriou, A. Dimeas, T. Tomtsi, A. Weidlich (Springer, Berlin, 2011)
Z.S. Tao, L. Rui, Z. Xia, H.C. Hua, W.Y. Quan, The emerging cutting-edge of virus research. Viral Proteomics. Science China: Life Sci. 65, 502–512 (2011)
C.S. Wagner, S.W. Popper, Identifying critical technologies in the United States: a review of the federal effort. J. Forecast. 22, 113–128 (2003)
M.J. Jackson, Micro and Nanomanufacturing (Springer, New York, 2007)
D.N. Weil, Economic Growth, 3rd edn. (Pearson, Boston, 2013)
L. Leydesdorff, M. Mayer, The decline of university patenting and the end of the Bayh-Dole effect. Scientometrics 83, 355–362 (2010)
M. Sanders, Scientific paradigms, entrepreneurial opportunities and cycles in economic growth. Small Bus. Econ. 28, 339–354 (2007)
H. Grupp (ed.), Dynamics of Science-Based Innovation (Springer, Berlin, 1992)
G.M. Grossman, E. Helpman, Innovation and Growth in the Global Economy (The MIT Press, Cambridge, MA, 1993)
R.J. Barro, X. Sala-i-Martin, Technological diffusion, convergence, and growth. J. Econ. Growth 2, 1–26 (1992)
R.E. Lucas Jr., Lectures on Economic Growth (Harward University Press, Cambridge, MA, 2002)
L. Girifalco, Dynamics of Technological Change (Van Nostrand Reinhold, New York, 1991)
H. Moed, R. de Druin, T.H. van Leeuwn, New bibliometric tools for the assessment of national research performance: database description, overview of indicators and first applications. Scientometrics 33, 381–422 (1995)
H.F. Moed, W.J.M. Burger, J.G. Frankfort, A.F.J. van Raan, The use of bibliometric data for the measurement of university research performance. Res. Policy 14, 131–149 (2002)
A.F.J. van Raan, Fatal attraction: conceptual and methodological problems in the ranking of universities by bibliometric methods. Scientometrics 62, 133–143 (2005)
H. Moed, Measuring China’s research performance using the science citation index. Scientometrics 53, 281–296 (2002)
A.F.J. van Raan, Measuring science pp. 19–50 in Handbook of Quantitative Science and Technology Research, ed. by H.F. Moed, W. Glänzel, U. Schmoch (Springer, Netherlands, 2005)
T.N. van Leeuwen, M.S. Visser, H.F. Moed, T.J. Nederhof, A.F.J. van Raan, The Holly Grail of science policy: exploring and combining bibliometric tools in search for scientific excellence. Scientometrics 57, 281–296 (2002)
J.A.D. Holbrook, Why measure science. Sci. Public Policy 19, 262–266 (1992)
H. Legler, G. Licht, A. Spielkamp, Germany’s Technological Performance. A Study on Behalf of the German Federal Ministry of Education and Research (Physica-Verlag, Berlin, 2000)
E.J. Rinia, Scientometric studies and their role in research policy of two research councils in the Netherlands. Scientometrics 47, 363–378 (2000)
K. Fealing, J. Lane, J. Marburger III, S. Shipp (eds.), The Science of science policy (Stanford University Press, Stanford, CA, A handbook, 2011), p. 2011
L. Bornmann, H.-D. Daniel, Does the \(h\)-index for ranking of scientists really work? Scientometrics 65, 391–392 (2005)
I. Feller, Performance measurement and the governance of American academic science. Minerva 47, 323–344 (2009)
L. Leydesdorrf, L. Bornmann, Integrated impact indicators compared with impact factors: an alternative research design with policy implications. J. Am. Soc. Inf. Sci. Technol. 62, 2133–2146 (2011)
L. Bornmann, L. Leydesdorrf, R. Mutz, The use of percentiles and percentile rank classes in the analysis of bibliometric data: opportunities and limits. J. Infometr. 7, 158–165 (2013)
J.A.D. Holbrook, Basic indicators of scientific and technological performance. Sci. Public Policy 19, 267–273 (1992)
D. Kondepudi, I. Prigogine, Modern Thermodynamics: From Heat Engines to Dissipative Structures (Wiley, New York, 1998)
G. Nicolis, I. Prigogine, Self-Organization in Non-equilibrium Systems (Wiley, New York, 1977)
E.C.M. Noyons, A.F.J. van Raan, Monitoring scientific developments from a dynamic perspective: self-organized structuring to map neural network research. J. Am. Soc. Inf. Sci. 49, 68–81 (1998)
P.A.A. van den Besselaar, L.A. Leydesdorff, Evolutionary Economics and Chaos Theory: New Directions in Technology Studies (Frances Pinter Publishers, 1994)
A. Stirling, Science, precaution, and the politics of technological risk. Ann. N. Y. Acad. Sci. 1128, 95–110 (2008)
A. Smith, A. Stirling, The politics of social-ecological resilience and sustainable socio-technical transitions. Ecol. Soc. 15, Art. No. 11 (2010)
J. Pfeffer, G.R. Salancik, The External Control of Organizations: A Resource Dependence Perspective (Stanford University Press, Stanford, CA, 2003)
M. Gibbons, C. Limoges, H. Nowotny, S. Schwartzman, P. Scott, M. Throw, The New Production of Knowledge: The Dynamics of Science and Research in Contemporary Societies (Sage Publications, London, 1994)
L. Hessels, H. van Lente, Re-thinking new knowledge production: a literature review and a research agenda. Res. Policy 37, 740–760 (2008)
T.R. Blackburn, Information and ecology of scholars. Science 181, 1141–1146 (1971)
N.C. Mullins, Theories and Theory Groups in Contemporary Sociology (Harper & Row, New York, 1973)
E. Jimenez-Contreras, F. de Moya-Anegon, E.D. Lopez-Cozar, The evolution of research activity in Spain: the impact of the national commission for the evaluation of research activity (CNEAI). Res. Policy 32, 123–142 (2003)
B.M. Gupta, C.R. Karisiddappa, Modelling the growth of literature in the area of theoretical population genetics. Scientometrics 49, 321–355 (2000)
H. Etzkowitz, L. Leydesdorff, The Triple Helix: University-industry-government relations: a laboratory for knowledge based economic development. EASST Rev. 14, 14–19 (1995)
L. Leydesdorff, H. Etzkowitz, Emergence of aTriple Helix of university-industry-government relations. Sci. Public Policy 23, 279–286 (1996)
L. Leydesdorff, H. Etzkowitz, The Triple Helix as a model for innovation studies. Sci. Public Policy 25, 195–203 (1998)
H. Etzkowitz, L. Leydesdorff, The endless transition: a ‘Triple Helix’ of university industry government relations. Minerva 36, 203–208 (1998)
H. Etzkowitz, L. Leydesdorff, The dynamics of innovation: from National Systems and ‘Mode 2’ to a Triple Helix of university-industry-government relations. Res. Policy 29, 109–123 (2000)
L. Leydesdorff, G. Zawdie, The Triple Helix perspective of innovation systems. Technol. Anal. Strat. Manag. 22, 789–804 (2010)
L. Leydesdorff, The knowledge-based economy and the Triple Helix model. Annu. Rev. Inf. Sci. Technol. 1, 365–417 (2010)
H. Etzkowitz, The Triple Helix: University-Industry-Government Innovation in Action (Routledge, New York, 2008)
L. Leydesdorff, E. Perevodchikov, A. Uvarov, Measuring Triple-Helix synergy in the Russian innovation system at regional, provincial, and national levels. J. Assoc. Inf. Sci. Technol. 66, 1229–1238 (2015)
L. Leydesdorff, The mutual information of university-industry-government relations: an indicator of the Triple Helix dynamics. Scientometrics 58, 445–467 (2003)
L. Leydesdorff, The Triple Helix, Quadruple Helix, ..., and N-tuple of helixes: explanatory models for analyzing the knowledge-based economy? J. Knowl. Econ. 3, 25–35 (2012)
L. Leydesdorff, The Triple Helix: an evolutionary model of innovations. Res. Policy 29, 243–255 (2000)
W.B. Arthur. Competing technologies. pp. 590–607, in Technical Change and Economic Theory, ed. by G. Dosi, C. Frenman, R. Nelson, G. Silverberg, L. Soete (Pinter, London, 1988)
W.B. Arthur, Competing technologies, increasing returns, and lock-in by historical events. Econ. J. 99, 116–131 (1989)
M. Zitt, E. Bassecoulard, Y. Okubo, Shadows of the past in international cooperation: collaboration profiles of the top five producers of science. Scientometrics 47, 627–657 (2000)
D.A. King, The scientific impact of nations. What different countries get for their research spending. Nature 430, 311–316 (2004)
M.E. Porter, The Competitive Advantage of Nations (Basingstoke, New York, Palgrave MacMillan, 1990)
R.M. May, The scientific wealth of nations. Science 275, 793–795 (1977)
A.L. Porter, J. David Roessner, X.-Y. Jin, N.C. Newman, Measuring national ‘emerging technology’ capabilities. Sci. Public Policy 29, 189–200 (2002)
J.-Y. Choung, H.-R. Hwang, National systems of innovation: Institutional linkages and performances in the case of Korea and Taiwan. Scientometrics 48, 413–426 (2000)
P. Zhou, L. Leydesdorff, The emergence of China as leading nation in science. Res. Policy 35, 83–104 (2006)
A.-W. Harzing, A. Giroud, The competitive advantage of nations: An application to academia. J. Infometr. 8, 29–42 (2014)
T.-E. Sandberg Hannsen, F. Jørgensen, The value of experience in research. J. Infometr. 9, 16–24 (2015)
B.R. Martin, J. Irvine, Assessing basic research: some partial indicators of scientific progress in radio astronomy. Res. Policy 12, 61–90 (1983)
B.R. Martin, The use of multiple indicators in the assessment of basic research. Scientometrics 36, 343–362 (1996)
D. Hicks, Performance-based university research funding systems. Res. Policy 41, 251–261 (2012)
S. Hornbostel, S. Böhmer, B. Klingsporn, J. Neufeld, M. von Ins, Funding of young scientist and scientific excellence. Scientometrics 79, 171–190 (2009)
M. Lamont, How Professors Think: Inside the Curious World of Academic Judgment (Harvard University Press, Cambridge, MA, 2009)
D. Hicks, J.S. Katz, Equity and excellence in research funding. Minerva 49, 137–151 (2011)
V.V. Nalimov, Faces of Science (ISI Press, Philadelphia, 1981)
L. Esterle, M. Zitt. Observation of scientific publications in astronomy/astrophysics. pp. 91–109 in Organizations and Strategies in Astronomy, ed. by A. Heck (Kluwer, Dordrecht, 2000)
M. Crosland, Scientific credentials: record of publications in the assessment of qualifications for election to the French Acad\(\acute{\rm {e}}\)mie des Sciences. Minerva 19, 605–631 (1981)
G.J. Feist, Quantity, quality, and depth of research as influences on scientific eminence: is quantity most important? Creat. Res. J. 10, 325–335 (1997)
D.C. Pelz, F.M. Andrews, Scientists in Organizations. Productive Climates for Research and Development (Wiley, New York, 1966)
R. Senter Jr., A causal model of productivity in a research facility. Scientometrics 10, 307–328 (1986)
T. Luukkonen, B. Stahle, Quality evaluations in the management of basic and applied research. Res. Policy 19, 357–368 (1990)
B. Kim, H. Oh, An effective R&D performance measurement system: survey of Korean R&D researchers. Omega 30, 19–31 (2002)
P. Dahler-Larsen, Constitutive effects of performance indicators. Public Manag. Rev. 16, 969–986 (2014)
P. Dahler-Larsen, The Evaluation Society (Stanford Business Books, Stanford, CA, 2012)
ISO 9000:2000, Quality Management Systems—Fundamentals and Vocabulary (ISO, Geneva)
S. Helmin, Scientific quality in the eyes of the scientists. A questionnaire study. Scientometrics 27, 3–18 (1993)
J.M. Pastor, L. Serrano, I. Zaera, The research output of European higher education institutions. Scientometrics 102, 1867–1893 (2015)
U. Schmoch, T. Schubert, Are international co-publications an indicator for quality of scientific research? Scientometrics 74, 361–377 (2008)
D. Cutla\(\breve{\rm c}\)a, D. Babi\(\acute{\rm c}\), I. \(\breve{\rm Z}\)ivkovi\(\acute{\rm c}\), D. \(\breve{\rm S}\)trbac, Analysis of qualitative and quantitative indicators of SEE countries scientific output. Scientometrics 102, 247–265 (2015)
F. Franceschini, M. Galetto, D. Maisano, Management by Measurement (Springer, Berlin, 2007)
F. Welter, S. Schröder, I. Leisten, A. Richert, S. Jeschke, Scientific performance indicators - empirical results from collaborative research centers and clusters of excellence in Germany, pp. 203–220 in Automation, Communication and Cybernetics in Science and Engineering 2013/2014, ed. by S. Jeschke, I. Insenhardt, F. Hees, K. Henning (Springer International Publishing, Switzerland, 2014)
S.D. Haitun, The problem of indicator-latent relationship in metric models I: statement and general solution. Scientometrics 23, 335–351 (1992)
S.D. Haitun, The problem of indicator-latent relationship in metric models II: metric models with a priori latent assignment. Scientometrics 24, 221–235 (1992)
G. Samorodnitsky, M.S. Taqqu, Non-Gaussian Random Processes. Stochastic Models with Infinite Variance. (Chapmann & Hall, Boca Raton, 1994)
F.E. Beth, J. Kallsen, T. Meyer-Brandis, A non-Gaussian Ornstein—Uhlenbeck process for electricity spot price modeling and derivatives pricing. Appl. Math. Finan. 14, 153–169 (2007)
O.E. Bandorff-Nielsen, N. Sheppard, Non-Gaussian Ornstein-Uhlenbeck-based models and some of their uses in financial economics. J. R. Stat. Soc. B 63, 167–241 (2001)
S.D. Haitun, The “rank-distortion” effect and non-Gaussian nature of scientific activities. Scientometrics 5, 375–395 (1983)
S.D. Haitun, Stationary scientometric distributions. II. Non-Gaussian nature of scientific activities. Scientometrics 4, 89–104 (1982)
M.G. Kendall, Natural law in the social sciences. J. R. Stat. Soc. A 124, 1–16 (1961)
S.D. Haitun, Scientometrics: State and Perspectives (Nauka, Moscow, 1983). (in Russian)
L. Leydesdorff, The Challenge of Scientometrics: The Development, Measurement, and Self-organization of Scientific Communications (DSWO Press, Leiden, 1995)
W. Glänzel, U. Schoepflin, Little scientometrics, big scientometrics.. and beyond? Scientometrics 30, 375–384 (1994)
L. Leydesdorff, S. Milojevich, Scientometrics. 1208, 4566 (2012)
L. Bornmann, L. Leydesdorff, Scientometrics in a changing research landscape. EMBO Rep. 15, 1228–1232 (2014)
A. Schubert, Scientometrics: the research field and its journal, pp. 179–195 in Organizations and Strategies in Astronomy II, ed. by A. Heck (Kluwer, Dordrecht, 2001)
L. Leydesdorff, P. van den Besselaar, Scientometrics and communication theory: towards theoretically informed indicators. Scientometrics 38, 155–174 (1997)
T. Braun, E. Bujdodo, A. Schubert, Literature of Analytical Chemistry: A scientometric evaluation (CRC Press, Boca Raton, FL, 1987)
W. Glänzel, A. Schubert, A new classification scheme of science fields and subfields designed for scientometric evaluation purposes. Scientometrics 56, 357–367 (2003)
M. Zitt, E. Bassecoulard, Challenges for scientometric indicators: data demining, knowledge-flow measurements and diversity issues. Ethics Sci. Environ. Polit. 8, 49–60 (2008)
T. Braun, A. Schubert, Scientometric versus socio-economic indicators. Scatter plots for 51 countries. 1978–1980. Scientometrics 13, 3–9 (1988)
F. Narin, M.B. Albert, V.M. Smith, Technology indicators and strategic planning. Sci. Public Policy 19, 369–381 (1992)
E. Bassecoulard, M. Zitt, Indicators in a research institute: a multi-level classification of scientific journals. Scientometrics 44, 325–345 (1999)
A. Schubert, S. Zsindely, T. Braun, Scientometric analysis of attendance at international scientific meetings. Scientometrics 5, 177–187 (1983)
A. Schubert, W. Glänzel, T. Braun, Scientometric datafiles. A comprehensive set of indicators on 2649 journals and 96 countries in all major science fields and subfields 1981–1985. Scientometrics 16, 3–478 (1989)
T. Braun, W. Glänzel, A. Schubert, Publication productivity: from frequency distributions to scientometric indicators. J. Inf. Sci. 16, 37–44 (1990)
T.A. Brooks, Private acts and public objects: an investigation of citer motivations. J. Am. Soc. Inf. Sci. 36, 223–229 (1985)
M.J. Moravcsik, P. Murugesan, Some results on the function and quality of citations. Soc. Stud. Sci. 5, 86–92 (1975)
T.A. Brooks, Evidence of complex citer motivations. J. Am. Soc. Inf. Sci. 37, 34–36 (1986)
D.W. Aksnes, Characteristics of highly cited papers. Res. Eval. 12, 159–170 (2003)
D.W. Aksnes, G. Sivertsen, The effect of highly cited papers on national citation indicators. Scientometrics 59, 213–224 (2004)
B. Cronin, The need of a theory of citing. J. Doc. 37, 16–24 (1981)
B. Cronin, The Citation Process. The Role and Significance of Citations in Scientific Communication (Taylor Graham, London, 1984)
B. Cronin, Norms and functions in citation: the view of journal editors and referees in psychology. Soc. Sci. Inf. Stud. 2, 65–78 (1982)
V. Cano, Citation behavior: classification, utility, and location. J. Am. Soc. Inf. Sci. 40, 284–290 (1989)
A. Schubert, The web of scientometrics. Scientometrics 63, 3–20 (2002)
F. Janssens, J. Leta, W. Glänzel, B. de Moor, Towards mapping library and information science. Inf. Process. Manag. 42, 1614–1642 (2006)
H. Small, K.W. Boyack, R. Klavans, Identifying emerging topics in science and technology. Res. Policy 43, 1450–1467 (2014)
H. Small, Interpreting maps of science using citation context sentiments: a preliminary investigation. Scientometrics 87, 373–388 (2011)
F. Narin, Bibliometrics. Annu. Rev. Inf. Sci. Technol. 35–58 (1977)
M.J. Kurtz, J. Bollen, Usage bibliometrics. Annu. Rev. Inf. Sci. Technol. 44, 1–64 (2010)
H.D. White, K.W. McCain, Bibliometrics. Annu. Rev. Inf. Sci. Technol. 24, 119–186 (1989)
F. Narin, Evaluative bibliometrics. Computer Horizons, Inc. Project No. 704R (1996)
B. Cronin, Bibliometrics and beyond: some thoughts on the web-based citation analysis. J. Inf. Sci. 27, 1–7 (2001)
J. Nicolaisen, The scholarliness of published peer reviews: a bibliometric study of book reviews in selected social science fields. Res. Eval. 11, 129–140 (2002)
A.J. Nederhof, A.F.J. van Raan, A bibliometric analysis of six economics research groups: A comparison with peer review. Res. Policy 22, 353–368 (1993)
A.J. Nederhof, Bibliometric monitoring of research performance in the social sciences and the humanities: A review. Scientometrics 66, 81–100 (2006)
F. Narin, Patent bibliometrics. Scientometrics 30, 147–155 (1994)
L. Björneborn, P. Ingwersen, Towards a basic framework for webometrics. J. Am. Soc. Inf. Sci. Technol. 55, 1216–1227 (2004)
M. Thelwall, L. Vaughan, L. Björnedborn, Webometrics. Annu. Rev. Inf. Sci. Technol. 39, 81–135 (2005)
L. Björneborn, P. Ingwersen, Perspectives of webometrics. Scientometrics 50, 65–82 (2001)
C. Borgman, J. Furner, Scholarly communication and bibliometrics. Annu. Rev. Inf. Sci. Technol. 36, 3–72 (2002)
M. Thelwall, Introduction to webometrics: quantitative web research for the social sciences. Synth. Lect. Inf. Concepts, Retr., Serv. 1, 1–116 (2009)
M. Thelwall, Bibliometrics to webometrics. J. Inf. Sci. 34, 605–621 (2008)
T.C. Almind, P. Ingwersen, Informetric analyses of the World Wide Web: methodological approaches to ‘webometrics’. J. Doc. 53, 404–426 (1997)
L. Björneborn, Small-world link structures across an academic Web space: a library and information science approach. Doctoral dissertation (Royal School of Library and Information Science, Copenhagen, Denmark, 2004)
P. Ingwersen, L. Björneborn, Methodological issues of webometric studies, pp. 339–369 in Handbook of Quantitative Science and Technology Research, ed. by H.F. Moed, W. Glänzel, U. Schmoch (Kluwer, New York, 2004)
A. Pritchard, Statistical bibliography or bibliometrics? J. Doc. 24, 348–349 (1969)
K. Debackere, W. Glänzel, Using a bibliometric approach to support research policy making: the case of the Flemish BOF-key. Scientometrics 59, 253–276 (2004)
T.N. van Leeuwen, M.S. Visser, H.F. Moed, T.J. Nederhof, A.F. van Raan, The Holy Grail of science policy: exploring and combining bibliometric tools in search of scientific excellence. Scientometrics 57, 257–280 (2003)
D.W. Aksnes, R.E. Taxt, Peer reviews and bibliometric indicators: a comparative study at a Norwegian university. Res. Eval. 13, 33–41 (2004)
A. Andres, Measuring Academic Research. How to Undertake a Bibliometric Study (Chandos Publishing, Oxford, 2009)
F. Narin, Bibliometric techniques in the evaluation of research programs. Sci. Public Policy 14, 99–106 (1987)
F. Narin, R.P. Rozek, Bibliometric analysis of the US pharmaceutical industry research performance. Res. Policy 17, 139–154 (1988)
W. Glaänzel, Bibliometrics as a research field. A course on theory and application of bibliometric indicators (Ungarische Akademie der Wissenschaften, Budapest, 2003)
V.I. Gorkova, Informetrics (quantitative methods in scientific and technical information). Itogi Nauki i Tekhniki. Ser. Informatika 10, 328 (1988). (in Russian)
M.S. Galyavieva, On the formation of the concept of informtrics (Review). Sci. Tech. Inf. Process. 40, 89–96 (2013)
C.S. Wilson, Informetrics. Annu. Rev. Inf. Sci. Technol. 34, 107–247 (1999)
J. Bar-Ilan, Informetrics at the beginning of the 21st century: a review. J. Informetr. 2, 1–52 (2008)
W.C. Adair, Citation indexes for scientific literature? Am. Doc. 6, 31–32 (1955)
E. Garfield, Citation indexes for science. Science 122(3159), 108–111 (1955)
E. Garfield, The Mystery of the transposed journal lists—wherein Bradford’s law of scattering is generalized according to Garfield’s law of concentration. Curr. Contents 17, 222–223 (1971)
O. Persson, Studying research collaboration using co-authorships. Scientometrics 36, 363–377 (1996)
M.A. Abbas, Weighted indexes for evaluating the quality of research with multiple authorship. Scientometrics, 88, 107–131 (2011)
D. De Solla Price, Little Science, Big Science (Columbia University Press, New York, 1963)
R. Rousseau, Why am I not cited or, why are multi-authored papers more cited than others? J. Doc. 48, 79–80 (1992)
B. Cronin, Hyperauthorship: a postmodern perversion or evidence of a structural shift in scholarly communication practices? J. Am. Soc. Inf. Sci. Technol. 52, 558–569 (2001)
C.S. Wagner, L. Leydesdorff, Network structure, self-organization, and the growth of international collaboration in science. Res. Policy 34, 1608–1618 (2005)
A.-L. Barabasi, H. Jeong, Z. Neda, E. Ravasz, A. Schubert, T. Vicsek, Evolution of the social network of scientific collaborations. Physica A 311, 590–614 (2002)
W. Glänzel, A. Schubert, Analysing scientific networks through co-authorship, pp. 257–276 in Handbook of Quantitative Science and Technology Research, ed. by H.F. Moed, W. Glänzel, U. Schmoch (Springer, Netherlands, 2005)
A. Schubert, T. Braun, International collaborations in the sciences 1981–1985. Scientometrics 19, 3–10 (1990)
A. Prarianes-Rodriguez, C. Olmeda-Gomez, F. Moya-Anegon, Detecting, identifying and visualizing research groups in co-authorship networks. Scientometrics 82, 307–319 (2010)
A. Bookstein, H. Moed, M. Yitzahki, Measures of international collaboration in scientific literature: Part I. Inf. Process. Manag. 42, 1408–1421 (2006)
S. Lehmann, B. Lautrup, A.D. Jackson, Citation networks in high energy physics. Phys. Rev. E 68, Art. No. 026113 (2003)
A. Bookstein, H. Moed, M. Yitzahki, Measures of international collaboration in scientific literature: Part II. Inf. Process. Manag. 42, 1422–1427 (2006)
M.J. Mulkay, G.N. Gilbert, S. Woolgar, Problem areas and research networks in science. Sociology: J. Brit. Soc. Assoc. 9, 187–203 (1975)
C.S. Wagner, Measuring the network of global science: comparing international co-authorships from 1990 to 2000. Int. J. Technol. Glob. 1, 185–208 (2005)
C.S. Wagner, The New Invisible College: Science for Development (The Brookings Institution, 2008)
D. Crane, Invisible Colleges: Diffusion of Knowledge in Scientific Communities (The University of Chicago Press, Chicago, 1972)
D.J. de Solla Price, D.B. Beaver, Collaboration in an invisible college. Am. Psychol. 21, 1011–1018 (1966)
A. Zuccala, Modeling the invisible college. J. Am. Soc. Inf. Sci. Technol. 57, 152–168 (2006)
A.A. Zuccala, Revisiting the invisible college: a case study of the intellectual structure and social process of singularity theory research in mathematics. Ph.D. thesis, University of Toronto, 2004
B. Cronin, Invisible colleges and information transfer. A review and commentary with particular reference to the social sciences. J. Doc. 38, 212–236 (1982)
H. Small, B.G. Griffith, The structure of scientific literatures I: identifying and graphic specialities. Sci. Stud. 4, 17–40 (1974)
B.G. Griffith, H.G. Small, J.A. Stonehill, S. Dey, The structure of scientific literatures II: toward a macro- and microstructure for science. Soc. Stud. Sci. 4, 339–365 (1974)
P. Auger, Tendances Actuelles de la Recherche Scientifique (UNESCO, 1961)
H.F. Moed, F. de Moya-Anegon, C. Lopez-Illescas, M. Visser, Is concentration of university research associated with better research performance? J. Infometr. 5, 649–658 (2011)
H.F. Moed, G. Halevi, A bibliometric approach to tracking international scientific migration. Scientometrics 101, 1987–2001 (2014)
H.F. Moed, M. Aisati, A. Plume, Studying scientific migration in Scopus. Scientometrics 94, 929–942 (2013)
E. Garfield, Citation analysis as a tool in journal evaluation. Science 178, 471–479 (1972)
E. Garfield, Citation indexing for studying science. Nature 227, 669–671 (1970)
L. Bornmann, H.-D. Daniel, What do citation counts measure? A review of studies on citing behavior. J. Doc. 64, 45–80 (2008)
R. Plomp, The highly cited papers of professors as an indicator of a research group’s scientific performance. Scientometrics 29, 377–393 (1994)
H. Small, Co-citation in the scientific literature: a new measure of the relationship between two documents. J. Am. Soc. Inf. Sci. 24, 265–269 (1973)
B. Cronin, H. Snyder, H. Atkins, Comparative citation rankings of authors in monographic and journal literature: a study of sociology. J. Doc. 53, 263–273 (1997)
L. Bornmann, R. Mutz, C. Neuhaus, H.-D. Daniel, Citation counts for research evaluation: standards of good practice for analyzing bibliometric data and presenting and interpreting results. Ethics Sci. Environ. Polit. 8, 93–102 (2008)
L. Bornmann, H.-D. Daniel, The citation speed index: a useful bibliometric indicator to add to the \(h\)-index. J. Infometr. 4, 444–446 (2010)
L. Leydesdorff, O. Amsterdamska, Dimensions of citation analysis. Sci. Technol. Hum. Values 15, 305–335 (1990)
H. Moed, Citation Analysis in Research Evaluation (Springer, Netherlands, 2005)
W. Glänzel, B. Dchlemmer, B. Thijs, Better late than never? On the chance to become highly cited only beyond the standard bibliometric time horizon. Scientometrics 58, 571–586 (2003)
H.F. Moed, M. Vriens, Possible inaccuracies occurring in citation analysis. J. Inf. Sci. 15, 95–107 (1989)
H.F. Moed, Citation analysis of scientific journals and journal impact measures. Curr. Sci. 89, 1990–1996 (2005)
B. Cronin, K. Overfelt, Citation—based auditing of academic performance. J. Am. Soc. Inf. Sci. 45, 61–72 (1994)
G. Lewison, The frequencies of occurrence of scientific papers with authors of each initial letter and their variation with nationality. Scientometrics 37, 401–416 (1996)
H.F. Moed, New developments in the use of citation analysis in research evaluation. Arch. Immunol. Ther. Exp. 57, 13–18 (2009)
B. Cronin, D. Shaw, Identity-creators and image-makers: using citation analysis and thick description to put authors in their place. Scientometrics 54, 31–49 (2002)
L. Egghe, R. Rousseau, Aging, obsolescence, impact, growth, and utilization: definitions and relations. J. Am. Soc. Inf. Sci. 51, 1004–1017 (2000)
E. Archambault, V. Lariviere, History of the journal impact factor: contingencies and consequences. Scientometrics 7, 635–649 (2009)
N. Scibata, Y. Kajikawa, K. Matsushima, Topological analysis of citation networks to discover the future core articles. J. Assoc. Inf. Sci. Technol. 68, 872–882 (2007)
E. Otte, R. Rousseau, Social network analysis: a powerful strategy, also in the information sciences. J. Inf. Sci. 28, 441–453 (2002)
M. Girvan, M.E.J. Newman, Community structure in social and biological networks. PNAS 99, 7821–7826 (2002)
N. Shibata, Y. Kajikawa, Y. Takeda, K. Matsushima, Detecting emerging research fronts based on topological measures in citation networks of scientific publications. Technovation 28, 758–775 (2008)
N.P. Hummon, P. Dereian, Connectivity in a citation networks: the development of DNA theory. Soc. Netw. 11, 39–63 (1989)
M. Zitt, E. Bassecoulard, Delineating complex scientific fields by an hybrid lexical-citation method: an application to nanosciences. Inf. Process. Manag. 42, 1513–1551 (2006)
M. Zitt, S. Ramanana-Rahary, E. Bassecoulard, Relativity of citation performance and excellence measures: from cross-field to cross-scale effects of field-normalization. Scientometrics 63, 373–401 (2005)
H. Small, Tracking and predicting growth areas in science. Scientometrics 68, 595–610 (2006)
H. Small, Paradigms, citations, and maps of science: a personal history. J. Am. Soc. Inf. Sci. Technol. 54, 394–399 (2003)
L. Leydesdorff, Betweeness centrality as an indicator of the interdisciplinarity of scientific journals. J. Am. Soc. Inf. Sci. Technol. 58, 1303–1319 (2007)
W. Glänzel, U. Schoepflin, A bibliometric study on ageing and reception processes of scientific literature. J. Inf. Sci. 21, 37–53 (1995)
T. Pollmann, Forgetting and the ageing of scientific publications. Scientometrics 47, 43–54 (2000)
I. Rafols, L. Leydesdorff, Contents-based and algorithmic classification of journals: perspectives on the dynamics of scientific communication and indexer effects. J. Am. Soc. Inf. Sci. Technol. 60, 1823–1835 (2009)
P.B. Coulter, Measuring Inequality (Westview Press, Boulder, CO, 1989)
M.E.D. Koenig, Determinants of expert judgment of research performance. Scientometrics 4, 361–378 (1982)
M.E.D. Koenig, Bibliometric indicators versus expert opinion in assessing research performance. J. Am. Soc. Inf. Sci. 34, 136–145 (1983)
L. Langfeldt, Decision-making and sources of bias. Expert panels evaluating research. Res. Eval. 13, 51–62 (2004)
A. Bryman, Quantity and Quality in Social Research (Unwin Hyman, London, 1988)
L. Leydesdorff, Various methods for the maping of science. Scientometrics 11, 295–324 (1987)
L. Leydesdorrf, T. Schank, Dynamic animations of journal maps: indicators of structural changes and interdisciplinary developments. J. Am. Soc. Inf. Sci. Technol. 59, 1810–1818 (2008)
L. Leydesdorff, I. Rafols, Local emergence and global diffusion of research technologies: an exploration of patterns of network formation. J. Am. Soc. Inf. Sci. Technol. 62, 846–860 (2011)
E. Bassencoulard, A. Lelu, M. Zitt, Mapping nanosciences by citation flows: a preliminary analysis. Scientometrics 70, 859–880 (2007)
H.P.F. Peters, A.F.J. van Raan, Representations by direct multidimensional scaling. Co-word-based science maps of chemical engineering. Part I. Res. Policy 22, 23–45 (1993)
H.P.F. Peters, A.F.J. van Raan, Representations by combined clustering and multidimensional scaling, Co-word-based science maps of chemical engineering. Part II. Res. Policy 22, 47–71 (1993)
M. Zitt, R. Barre, A. Sigogneau, F. Laville, Territorial concentration and evolution of science and technology activities in the European Union: A descriptive analysis. Res. Policy 28, 545–562 (1999)
C. Chen, Visualising semantic spaces and author co-citation networks in digital libraries. Inf. Process. Manag. 35, 401–420 (1999)
L. Kay, N. Newman, J. Youtie, A.L. Porter, I. Rafols, Patent overlay mapping: visualizing technological distance. J. Assoc. Inf. Sci. Technol. 65, 2432–2443 (2014)
C. Chen, CiteSpace II: detecting and visualizing emerging trends and transient patterns in scientific literature. J. Am. Soc. Inf. Sci. Technol. 57, 359–377 (2006)
M. Zitt, S. Ramanana-Rahary, E. Bassencoulard, F. Laville, Potential science - technology spillovers in regions: an insight on geographic co-location of knowledge activities in the EU. Scientometrics 57, 295–320 (2003)
C. Chen, Information Visualization: Beyond the Horizon (Springer, London, 2006)
L. Leydesdorff, Clusters and maps of science journals based on bi-connected graphs in Journal Citations Reports. J. Doc. 60, 371–427 (2004)
P. van den Besselaar, G. Heimeriks, Mapping research topics using word-reference co-occurrences: a method and an exploratory case study. Scientometrics 68, 377–399 (2006)
K. Börner, D.E. Poley, Visual Insights: A Practical Guide to Making Sense of Data (MIT Press, Cambridge, MA, 2014)
K. Börner, Atlas of Knowledge: Anyone can Map (MIT Press, Cambridge, MA, 2014)
K. Börner, T.N. Theriault, K.W. Boyack, Mapping science introduction: past, present and future. Bull. Am. Soc. Inf. Sci. Technol. 41, 12–16 (2015)
F. de Moya-Anegon, B. Vargas-Quesada, V. Herrero-Solana, Z. Chinchilla-Rodriguez, E. Corera-lvarez, F.J. Munoz-Fernande, A new technique for building maps of large scientific domains based on the cocitation of classes and categories. Scientometrics 61, 129–145 (2004)
R.R. Braam, H.F. Moed, A.F.J. van Raan, Mapping of science by combined co-citation and word analysis I. Structural aspects. JASIS 42, 233–251 (1991)
R.R. Braam, H.F. Moed, A.F.J. van Raan, Mapping of science by combined co-citation and word analysis II. Dynamical aspects. JASIS 42, 252–266 (1991)
A.M. Zoss, K. Börner, Mapping interactions within the evolving science of science and innovation policy community. Scientometrics 91, 631–644 (2011)
B. Vargas-Quesada, F. de Maoya-Anegon, Visualizing the Structure of Science (Springer, Berlin, 2007)
K.W. Boyack, R. Klavans, K. Börner, Mapping the backbone of science. Scientometrics 64, 351–374 (2005)
K.W. Boyack, D. Newman, R.J. Duhon, R. Klavans, M. Patek, J.R. Biberstine, B. Schijvenaars, A. Skupin, N. Ma, K. Börner, Clustering more than two million biomedical publications: comparing the accuracies of nine text-based similarity approaches. PloS One 6, e18029 (2011)
G. Heimeriks, M. Hoerlesberger, P. van den Besselaar, Mapping communication and collaboration in heterogeneous research networks. Scientometrics 58, 391–413 (2003)
K. Börner, Plug-and-play macroscopes. Commun. ACM 54, 60–69 (2011)
Y.W. Chen, S. Fang, K. Börner, Mapping the development of scientometrics: 2002–2008. J. Libr. Sci. China 3, 131–146 (2011)
K. Börner, C. Chen, K.W. Boyack, Vizualizing knowledge domains. Annu. Rev. Inf. Sci. Technol. 37, 179–255 (2003)
F. de Moya-Anegon, B. Vargas-Quesada, Z. Chinchilla-Rodriguez, E. Corera-lvarez, F.J Munoz-Fernandez, V. Herrero-Solana, Visualizing the marrow of science. J. Am. Soc. Inf. Sci. Technol. 58, 2167–2179 (2007)
R.J.W. Tijsen, A.F.J. van Raan, Mapping changes in science and technology bibliometric co-occurrence analysis of the R&D literature. Eval. Rev. 18, 98–115 (1994)
R. Klavans, K. Boyack, Toward a consensus map of science. J. Am. Soc. Inf. Sci. Technol. 60, 455–476 (2009)
A. Quirin, O. Cordon, J. Santamaria, B. Vargas-Quesada, F. de Moya-Anegon, A new variant of the pathfinder algorithm to generate large visual science maps in cubic time. Inf. Process. Manag. 4, 1611–1623 (2008)
K.W. Boyack, Mapping knowledge domains: characterizing PNAS. PNAS 101(Supplement 1), 5192–5199 (2004)
H. Small, A SCI-MAP case study: building a map of AIDS research. Scientometrics 30, 229–241 (1994)
P. van den Besselaar, L. Leydesdorff, Mapping change in scientific specialities: a scientometric reconstruction of the developing of artificial intelligence. J. Am. Soc. Inf. Sci. 47, 415–436 (1996)
A. Perianes-Rodriguez, C. Olmeda-Gomez, F. Moya-Anegon, Detecting, identifying and visualizing research groups in co-authorship networks. Scientometrics 82, 307–319 (2010)
S. Wright, The roles of mutation, inbreeding, crossbreeding and selection in evolution, in Proceedings of the Sixth International Congress on Genetics vol. 1, pp. 356–366 (1932)
H. Small, Update on science mapping: creating large document spaces. Scientometrics 38, 275–293 (1997)
H. Small, A general framework for creating large-scale maps of science in two or three dimensions: The SciViz system. Scientometrics 41, 125–133 (1998)
L. Bornmann, L. Waltman, The detection of “hot regions” in the geography of science—a visualization approach by using density maps. J. Informetr. 5, 547–553 (2011)
L. Leydesdorff, I. Rafols, A global map of science based on the ISI subject categories. J. Am. Soc. Inf. Sci. Technol. 60, 348–362 (2009)
I. Rafols, A.L. Porter, L. Leydesdorff, Science overlay maps: a new tool for research policy and library management. J. Am. Soc. Inf. Sci. Technol. 61, 1871–1887 (2010)
C. Wagner, L. Leydesdorff, Mapping the network of global science: comparing international co-authorships from 1990 to 2000. Int. J. Technol. Glob. 1, 185–208 (2005)
E.C.M. Noyons, A.F.J. van Raan, Advanced mapping of science and technology. Scientometrics 41, 61–67 (1998)
K. Boyack, Using detailed maps of science to identify potential collaborators. Scientometrics 57, 27–44 (2008)
R. Klavans, K.W. Boyack, Using global mapping to create more accurate document-level maps of research fields. J. Am. Soc. Inf. Sci. Technol. 62, 1–18 (2011)
M. Zitt, S. Ramanana-Rahary, E. Bassecoulard, Correcting glasses help fair comparisons in international science landscape: country indicators as a function of ISI database delineation. Scientometrics 56, 259–282 (2003)
K.W. Boyack, R. Klavans, Creation of a highly detailed, dynamic, global model and map of science. J. Am. Soc. Inf. Sci. Technol. 65, 670–685 (2014)
A. Poter, I. Rafols, Is science becoming more interdisciplinary? Measuring and mapping six research fields over time. Scientometrics 81, 719–745 (2009)
L. Leydesdorff, S. Cozzen, P. van den Besselaar, Tracking areas of strategic importance using scientometric journal mapings. Res. Policy 23, 217–229 (1994)
L. leydesdorff, S. Carley, I. Rafols, Global maps of science based on the new Web-of-Science categories. Scientometrics 94, 589–593 (2013)
A. Scharnhorst, Citation networks, science landscapes and evolutionary strategies. Scientometrics 43, 95–106 (1998)
G. Krampen, R. Becker, U. Wahner, L. Montada, On the validity of citation counting in science evaluation: content analyses of references and citations in psychological publications. Scientometrics 71, 191–202 (2007)
S.M. Lawani, A.E. Bayer, Validity of citation criteria for assessing the influence of scientific publications: new evidence with peer assessment. J. Am. Soc. Inf. Sci. 34, 59–66 (1983)
L. Egghe, Mathematical theory of citation. Scientometrics 43, 57–62 (1990)
Y.A. Shenhav, Y. Haberfield, The various faces of scientific productivity: a contingency analysis. Qual. Quant. 22, 365–380 (1988)
B.F. Reskin, Scientific productivity and the reward structure of science. Am. Sociol. Rev. 42, 419–504 (1977)
P.D. Allison, Inequality and scientific productivity. Soc. Stud. Sci. 10, 163–179 (1980)
S. Cole, T.J. Phelan, The scientific productivity of nations. Minerva 37, 1–23 (1999)
S. Lee, B. Bozeman, The impact of research collaboration on scientific productivity. Soc. Stud. Sci. 35, 673–702 (2005)
L. Meitzer, Scientific productivity in organizational settings. J. Soc. Issues 12, 32–40 (1956)
B.A. Jacob, L. Lefgren, The impact of research grant funding on scientific productivity. J. Public Econ. 95, 1168–1177 (2011)
A. Bonaccorsi, C. Daraio, Age effects in scientific productivity. Case of Italian National Research Council (CNR). Scientometrics 58, 49–90 (2003)
D.K. Simonton, Creative productivity and age: a mathematical model based on a two-step cognitive process. Dev. Rev. 4, 77–111 (1984)
A. Diamond, An economic model of the life-cycle research productivity of scientists. Scientometrics 6, 189–196 (1984)
S. Kyvik, Age and scientific productivity. Differences between fields of learning. High. Educ. 19, 37–55 (1990)
P. Seglen, D. Aksnes, Scientific productivity and group size: a bibliometric analysis of Norwegian microbiological research. Scientometrics 49, 125–143 (2000)
A. Ramesh Babu, Y.P. Singh, Determinants of research productivity. Scientometrics 43, 309–329 (1998)
K. Jaffe, M. Caicedo, M. Manzaranes, M. Gil, A. Rios, A. Florez, C. Montoreano, V. Davilla, Productivity in physical and chemical science predicts the future economic growth of developing countries better than other popular indices. PLOS ONE, e66239 (2013)
A. Bonaccorsi, C. Daraio, A robust nonparametric approach to the analysis of scientific productivity. Res. Eval. 12, 47–69 (2003)
D. Lindsey, Production and citation measures in the sociology of science: the problem of multiple authorship. Soc. Stud. Sci. 10, 145–162 (1980)
H.F. Moed, Bibliometric indicators reflect publication and management strategies. Scientometrics 47, 323–346 (2000)
G.S. Howard, D.A. Cole, S.E. Maxwell, Research productivity in psychology based on publication in the journals of the American Psychological Association. Am. Psychol. 42, 975–986 (1987)
L.B. Ellwein, M. Khachab, R.H. Waldman, Assessing research productivity: evaluating journal publication across academic departments. Acad. Med. 64, 319–325 (1989)
I. Lukovits, P. Vinkler, Correct credit distribution: a model for sharing credit among coauthors. Soc. Indic. Res. 36, 91–98 (1995)
F.J. Trueba, H. Guerrero, A robust formula to credit authors for their publications. Scientometrics 60, 181–204 (2004)
L.B. Dizon, M.S.M. Sadorra, Patterns of publication by the staff of an international fisheries research center. Scientometrics 32, 67–75 (1995)
P. Vinkler, Evaluation of the publication activity of research teams by means of scientometric indicators. Curr. Sci. 79, 602–612 (2000)
L. Butler, What happens when funding is linked to publication counts? pp. 389–405 in. H.F. Moed, W. Glänzel, U. Schmoch, Handbook of Quantitative Science and Technology Research (Springer, Netherlands, 2005)
G.G. Dyumenton, Networks of Scientific Communications and Organization of Fundamental Research (Nauka, Moscow, 1987) (in Russian)
G. Abramo, T. Cicero, C.A. D’Angelo, How important is choice of the scaling factor in standardizing citations? J. Informetr. 6, 645–654 (2012)
G. Abramo, T. Cicero, C.A. D’Angelo, Revisiting the scaling of citations for research assessment. J. Informetr. 6, 470–4479 (2012)
B.R. Martin, J. Irvine, Assessing basic research. Some partial indicators of scientific progress in radioastronomy. Res. Policy 12, 61–90 (1983)
M.J. Moravcsik, Progress report on quantification of science. J. Sci. Ind. Res. 36, 195–203 (1977)
D. Lindsey, Using citation counts as a measure of quality in science: measuring what’s measurable rather than what’s valid. Scientometrics 15, 189–203 (1989)
D. Lindsey, The corrected quality ratio: a composite index of scientific contribution to knowledge. Soc. Stud. Sci. 8, 349–354 (1978)
M.J. Moravcsik, Some contextual problems of science indicators. pp. 11 – 30 in Handbook of Quantitative Studies of Science and Technology, ed. by A.F.J. Van Raan (Elsevier, Amsterdam, 1988)
P. Abelson, Mechanisms for evaluating scientific information and the role of peer review. J. Am. Soc. Inf. Sci. 41(3), 216–222 (1990)
G. Abramo, C.A. D’Angelo, Evaluating research: from informed peer review to bibliometrics. Scientometrics 87, 499–514 (2011)
A. Pouris, Evaluating academic science institutions in South Africa. J. Am. Soc. Inf. Sci. 40, 269–372 (1989)
R. Miller, The influence of primary task on R&D laboratory evaluation: a comparative bibliometric analysis. R&D Manag. 22, 3–20 (1992)
P. Vinkler, General performance indexes calculated for research institutes of the Hungarian Academy of Sciences based on scientometric indicators. Scientometrics 41, 185–200 (1998)
T.N. van Leeuwen, L.J. van der Wurff, A.F.J. van Raan, The use of combined bibliometric methods in research funding policy. Res. Eval. 10, 195–201 (2001)
A. Watson, UK research funding—Universities raise their game, but the money doesn’t flow. Science 294, 2448–2449 (2001)
H.F. Moed, UK research assessment exercises: informed judgments on research quality or quantity? Scientometrics 74, 153–161 (2008)
M.H. MacRoberts, B.R. MacRoberts, Problems of citation analysis: a critical review. J. Am. Soc. Inf. Sci. 40, 342–349 (1989)
J. Nicolaisen, Citation analysis. Ann. Rev. Inf. Sci. Technol. 41, 609–642 (2007)
L. Georghiou, D. Roessner, Evaluating technology programs: tools and methods. Res. Policy 29, 657–678 (2000)
M. Marzolla, Quantitative analysis of the Italian national scientific qualification. J. Infometr. 9, 285–316 (2015)
K. Rørstad, D.W. Aksness, Publication rate expressed by age, gender and academic position—A large scale analysis of Norwegian academic staff. J. Infometr. 9, 317–333 (2015)
R. Barre, S&T indicators for policy making in a changing science-society relationship, pp. 115–131 in Handbook of Quantitative Science and Technology Research. The Use of Publication and Patent Statistics in Studies of S&T Systems, ed. by H.F. Moed, W. Glänzel, U. Schmoch (Springer, Netherlands, 2005)
L. Egghe, Performance and its relation with productivity in Lotkaian systems. Scientometrics 81, 567–585 (2009)
H. English, H.-J. Czerwon, Quantification of the performance of research units: a simple mathematical model. Res. Polit. 19, 477–480 (1990)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Vitanov, N.K. (2016). Science and Society. Assessment of Research. In: Science Dynamics and Research Production. Qualitative and Quantitative Analysis of Scientific and Scholarly Communication. Springer, Cham. https://doi.org/10.1007/978-3-319-41631-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-41631-1_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-41629-8
Online ISBN: 978-3-319-41631-1
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)