Abstract
The very nature of scientific activity and information, which are meant to be shared, is the starting point in defining a scientific journal, and the criteria according to which its value and role are determined. The authors aim at analysing some criteria that define the quality of scientific journals considering their visibility and impact. The concept of open access for journals is analysed in point of its advantages and disadvantages, since it differs greatly from the subscription-based access, whether we talk about institutional or individual subscriptions. The authors are in favour of the concept of public access, considering that it gives a journal more visibility, on the condition that article processing charges are reduced. The essential condition for a journal to become renowned is to be as visible as possible. The concept of open access is beneficial, supports instruction through and for scientific research, regardless of educational level. The aim of this paper is to emphasise the modalities, specificities and bibliometric performances (percent of citable documents, impact factor and immediacy index) of open access versus subscription-based access, as well as to investigate whether the use of open access concept determines an increase of the journals’ quality, study applied to the analysis of Hindawi Publishing Company journals and Multidisciplinary Digital Publishing Institute (MDPI) journals.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Within an economy based on knowledge, all types of information (scientific, technical, medical, socio-human) have an essential role in the society development. Identifying and assessing the consequences of scientific information lack haven’t always been a concern to all policy-makers, though they are fully aware that economic development and social welfare are closely connected to the existence of a well-informed intellectual élite.
Innovation, as a process consisting in extracting social and economic values out of knowledge, generates and develops new ideas, and, consequently, requires not only use, but also dissemination. Researchers and innovators in the domains of various sciences and technologies can expand the frontiers of knowledge only if good and rapid information is ensured. The market success of researchers and the implicit economic development are closely connected to the access to information, especially to highly specialised information, and the benefits of research depend mainly on the access to information. Scientific activity is based on a series of vehicles of communication, which have specific objectives.
Scientific articles as vehicles of communicating research results appeared for the first time in the early scientific journals of the 17-th century: Philosophical Transactions of the Royal Society—1665 in England, Journal des Sçavans—1665 and Comptes Rendus de l’Académie des Sciences—1835 in France. From the beginning, the journal Philosophical Transactions acquires a significant prestige, since among those who published in it there were Isaac Newton, Charles Darwin, Michael Faraday, William Herschel and many others. Philosophical Transactions established the important principles of scientific priority and peer review, which have become the central foundations of scientific journals ever since (Philosophical Transactions 2015). Le Journal des Savants publishes original articles which present examples of notable progress in various domains (Journal des Savants 2015). Comptes Rendus ensures the dissemination of scientific information, keeping the scientific community informed in a rapid way about the newest and most significant published research results (Institut de France, Académie des Sciences 2015).
Scientific journals are the basic instruments in the process of knowledge dissemination, at the same time, being true knowledge archives, connecting domains, thematic areas, authors, research entities, quotations, examples of cooperation etc. (Ardelean et al. 2014).
As science is defined using terms belonging to the domain of scientific literature, it results that the notions of knowledge and scientific writings are quasi-synonymous. This equivalence was made for the first time by John Derek de Solla Price (1964), who, making use of Lotka and Bradford’s studies (Lotka 1926; Bradford 1934), stated the principle of using the number of scientific articles as a quantitative index of research activity: science is what is published in scientific journals, articles, and presentations. He hypothesised that the number of scientific articles published every year represents an index of the general activity or of a certain research domain, pointing out that by measuring the science development in this way, the result is an exponential growth following a period of linear growth.
Exponential growth has characterised the development of western science beginning with the 17-th century, and the law describing this growth is expressed by the equation:
where F(t) represents the variation in the scientific literature corpus in time; a represents the initial dimension of the scientific literature corpus at the moment t = 0; b represents the continuous level of knowledge, i.e. the percentage of yearly scientific literature increase.
This law of the exponential growth of sciences stated by Price, was established after accounting the number of scientific articles in the series Philosophical Transaction of the Royal Society of London, starting with the year 1665, and Physics Abstracts (between 1898 and 1950). The law was verified on the articles in the domain of mathematics referring to determinants and matrices, between 1693–1919 and 1920–1949, Price making the following observation: if a significantly large segment of science is measured by a reasonable means, its natural growth is exponential (Ardelean et al. 2014).
An article, irrespective of its nature, is regarded as a message, as a summum of representations, but also as a sum of specific elements which can be analysed by using different methods and interpreted accordingly. An article published in a journal meets one of the final purposes of research, i.e. its acknowledgement by the scientific community. Periodicals are classified according to a variety of criteria: the most cited in the form of the absolute value; those containing the highest number of articles cited on the average, globally, or considering the domain; those which publish the highest number of articles; those which are cited by prestigious journals etc. They can be assessed according to objective criteria.
According to the data published on the World Bank webpage, in 2011 there were published 582,012 articles in the following domains: physics, biology, chemistry, mathematics, clinical medicine, biomedical research, engineering and technology, space and earth sciences, among which 176,189 in the euro zone.
Therefore, the connection between journals and the articles they publish is obvious. The main issue raised by both authors and editors is the value level of a certain journal, since the latter represents a real knowledge archive that puts together the various vuepoints of the researchers.
This paper aims to emphasise some bibliometric indicators of several open access journals belonging to Hindawi Publishing Company and Multidisciplinary Digital Publishing Institute (MDPI), and also, relying on the analysed bibliometric performances (percent of citable documents, impact factor and immediacy index), to establish whether a visibility increase would lead to a quality increase, expressed by the impact factor.
The bibliometric assessing of journals
The proposing authors submit their scientific articles, mainly depending on their research value (established by self-assessment) and on the value of the journal. The latter is based on economic and intellectual criteria. Among those belonging to the former category, we can mention: editing expenses, peer-review expenses, the number of subscriptions etc. Intellectual criteria include the quality of the journal, its visibility and number of citations. It is obvious that there is a close interdependence among these criteria. In the last decades, journals have been indexed in data bases and the indexation is determined by their meeting well-defined and quantifiable criteria.
The quality level of journals and the method of establishing it, have been a permanent concern of both editors and scientists (Rousseau 2002; Braun 2007).
Haensly et al. (2009) propose to establish the quality level of a journal in two ways:
-
based on a survey made by means of a questionnaire asking the participants to rate a list of journals;
-
based on the number of citations.
Rousseau summarised the ten characteristics of a quality journal by reviewing Zwemer (1970), and Garfield (1990). These ten criteria are (Vishwakarma and Mukherjee 2014):
-
1.
Acceptance and rejection rates;
-
2.
Subject and geographical representativeness of the editorial board;
-
3.
Use of a critical refereeing system;
-
4.
Promptness of publication;
-
5.
Coverage by major abstracting and indexing services;
-
6.
High confidence level of scientists using the journal in its contents;
-
7.
High frequency of citation by other journals (impact);
-
8.
Inclusion of abstracts/summaries in English;
-
9.
Providing author(s) addresses (author reputation score); and
-
10.
Providing complete bibliographic information.
We will further deal with criteria number 5 and 7, because we aim at the exclusive analysis of the main factors determining journal impact, the non-analysed criteria exerting a smaller influence on the impact. Data base inclusion is extremely important and, as we have already stated, it depends on meeting a series of sine qua non conditions. Of course, data bases have also other criteria to include a journal, among which we mention: regular journal issues, the editing quality, the uniform system of publishing papers, the uniform structure of the article etc. Data bases present and/or describe the content of the document by means of indexation. A data base is a corpus of information, belonging to one or more domains of knowledge, stored within a certain entity and organised so that the information can be accessed. Indexation is the process of analysing a document and assigning it some terms in order to be better described.
The criterion “Coverage by major abstracting and indexing services” is important only to the extent to which these services ensure articles/journals high visibility, thus making them acquire popularity. The indexation categories are called descriptors and they can be:
-
a single key word;
-
2–3 key words;
-
classifying codes, for instance those in Chemical Abstracts, those from Derwent World Patents Index etc.
Today, most data bases are made in electronic form by various entities, beginning with scientific publishing houses which issue scientific periodicals. These have their own data bases, usually based on subscription, i.e. on paying a fee in order to gain access. Few data bases are fully available through free access system.
The impact, generally defined as the effect or the influence produced by something, acquires a special significance with reference to articles and journals. In this respect, one can state that a journal is more or less important than other journals, in the same way in which an article can be more or less important than another.
The most widely used measure of the impact is the number of citations of the articles of a journal in another journal. The number of citations is a measure which results in the impact factor for journals (Okubo 1997; Moed 2005), the ratio between the number of citations and the number of articles for a period of 2 years. The impact factor was introduced by Garfield (1955, 1972, 1979). For each journal, the impact factor is calculated and published in the Journal of Citation Reports (JRC), resulting in various comparisons (Kulkarni et al. 2009) or a journal hierarchy (Annual Reviews rankings in Thomson Reuters Journal Citation Reports 2015). The JRC includes more than 11,000 journals in 237 disciplines, from 82 countries. The 2015 Edition of the Journal Citation Reports (JRC) provides a combination of impact and influence metrics from 2014 data. Then, other impact indices are introduced, all of them being exclusively linked to the number of citations:
-
Hirsh index introduced by J.E. Hirsch, seeks to establish a bibliometric criterion of quantifying the impact and relevance of research (Hirsch 2005) and also of journals (Braun et al. 2006);
-
eigenfactor (Bergstrom 2007; Bergstrom et al. 2008; Eigenfactor 2016); Eigenfactor.org contains 115,000 reference items;
-
article influence score (AIS) the Eigenfactor and the AIS are calculated considering the citations received over a 5 year period and can be consulted freely on the web page of eigenfactor.org (Ardent 2010);
-
g index (Egghe 2006);
-
e index (Zhang 2009); g-index and e-index are used in the assessment made by Publish or Perish (Harzing 2007);
-
SCImago Journal Rank, developed by Scopus data base (Elsevier) (SCImago Journal and Country Rank 2016);
-
source normalized impact per paper (SNIP) (Moed 2010) etc.
For calculating the Hirsch index, Thomson Reuters and other data bases such as Google Scholar consider self-citations. Studies on the topic (Glänzel and Thijs 2004) proved that “at a macro level” it is not necessary to exclude self-citations.
This paper does not intend to analyse all indexes, but just to mention them, considering that the impact is determined by the number of citations, these being numbered by three extremely important data bases: Web of Science, Google Scholar (through Publish or Perish) and Scopus (through SciMago). The reviewing concerns the journals indexed in their own data bases, except for Google Scholar, who reviews all web published publications. Sometimes, there are differences among the data offered by the above mentioned data bases (Garfield 1979; Archambault et al. 2009; Falagas et al. 2008; Abrizah et al. 2013; Bar-Illan et al. 2007).
Unlike Web of Science and Scopus, Google Scholar information access is free. The number of scholarly documents published in English on the web exceeds 114 million, of which about 100 million on Google Scholar and at least 27 million being free available documents (Khabsa and Giles 2014) (Fig. 1).
The open access concept
Free access to scientific information is a fundamental problem of contemporary society and, at the same time, a challenge of the 21-st century. Scientific and technical information, which is a type of commodity from the economic point of view, being integrated within the production and/or social area, allows a reasonable use of the means of production, improves goods and services, proposes new ones and takes part in the globalization phenomenon. Being in a competitional environment, enterprises must continuously innovate, improving their productivity, offering new products and services, entering new markets, which cannot be done in the absence of scientific and technical information, or of the results of research and development. Productivity and research progress (at country, region, institution or researcher levels) can be measured by using indices such as: the number of publications/letters patents, the number of citations, Hirsch index etc. The spread of scientific information, the dissemination of the results of research, was done, till recently, almost exclusively, by means of specialised journals, each having its own impact. Spreading scientific information involves two factors: time, determined by the distribution speed, and space, in the form of the distribution area. Scientific journals, privileged instruments of knowledge circulation, are, on the one hand, indispensable elements for a researcher’s visibility and acknowledgement, and, on the other hand, a significant source of profit for editors, in case of subscription. Scholarly journals have become ever more expensive and, economically, they can be less and less affordable for libraries and research institutions. Thus, a university cannot make subscriptions to thousands of journals, pertaining to various domains of knowledge, and, consequently, the effects are a more reduced visibility of sciences in general, and a corresponding level of information. There is an opinion that the system based on subscription to thousands of journals does not work in the best interest of researchers, but in that of the editors who aim at improving their market position. For instance, in the financial year 2013–2014, Canadian universities spent over 360 million dollars to Library acquisitions (Statistics Canada 2015).
Walters (2008) estimates that the fees will increase each year with 1.4 % for books, and with 8.5 % for journals. The author, taking the year 2007 as reference (to which he assigns 100 %), considers that in 2025 the fees for journals will become with almost 500 % greater. Considering an annual increase of the budget of 3 %, for these price increases, the number of journal titles acquired will diminish till 2025 to less than one half of those acquired in 2007.
It is obvious that, at the same time, the editors are confronted with ever increasing costs caused by “peer review”, editing and distribution. The notion of open access (OA) appeared in 2002, to solve the relation between the access to scientific information, on the one hand, and publishing, on the other. The notion was mentioned in The Budapest Declaration, following the initiative of The Open Society Institute (Budapest Open Access Initiative 2002), and had the declared purpose to eliminate the obstacles blocking the access to scientific information. In this statement, open access is defined as “making scientific literature available on line, freely and without restrictions”. Among the supporters of the concept of open access, we mention: Bethesda Statement on Open Access Publishing—2003, Access to scientific information-Interacademy panel Mexic—2003, Berlin Declaration on Open Access to Knowledge in the Sciences and Humanities—2003, Statement on open Access-International Federation of Library Associations—2004, and, more recently, The Joint Declaration on Open Science-ALEEA and European Commission—2012, The Aachen Declaration-The Annual Meeting of the General Medical Library Association, Germany—2012 etc.
In Mexico, on December 4, 2003, the Science Academies of several countries made their position known referring to the access to scientific information. Among other things, they recommended:
-
electronic access to journals must become free all over the world, a year after the publication for the researchers from developed countries and immediately for the researchers from the developing countries;
-
the content of the journals and data must be presented in a standardised distribution format to facilitate the use.
According to the data published on the web, in January 2015, the number of open access journals in DOAJ (Directory of Open Access Journals) amounts to 10,128, BioMed Central editor includes about 270 journals, Hindawi 405 journals, Science Direct offers about 290 journals, De Gruyter Open about 350 journals, SpringerOpen about 200 journals etc. In 2012, Morisson opinionated that about 30 % of scholarly peer-reviewed journals are included in the open access system. At the same time, the number of Institutional Repositories which publish doctoral theses, technical reports, etc. has increased, and so has self-archivation.
The European Commission defines the two basic open access models in the following way [COM(2012) 401 final]:
-
“Gold open access (open access publishing): payment of publication costs are shifted from readers (via subscriptions) to authors. These costs are usually borne by the university or research institute to which the researcher is affiliated, or by the funding agency supporting the research”;
-
“Green open access (self-archiving): the published article or the final peer-reviewed manuscript is archived by the researcher in an online repository before, after or alongside its publication. Access to this article is often delayed (‘embargo period’) at the request of the publisher so that subscribers retain an added benefit”. The EC document also stipulates that the “embargo period” of “green open” model and the archived version at various moments in time, may depend on the agreement between publishers and authors.
The European Commission mentions that “as of 2014, all articles produced with funding from Horizon 2020 will have to be accessible:
-
the articles will either immediately be made accessible online by the publisher (‘Gold’ open access)—up-front publication costs can be eligible for reimbursement by the European Commission; or
-
researchers will make their articles available through an open access repository no later than 6 months (12 months for articles in the fields of social sciences and humanities) after publication (‘Green’ open access)”.
Advantages of using open access
-
Scientific articles become available all over the world, in various formats (tex, pdf), by means of a system which is secured and rather stable technically (Ardelean et al. 2014);
-
Scientific information spread becomes universal, which has an immediate impact on the author. The impact factor multiplies in the case of free access articles. Lawrence (2001) has proven that free access increases the number of citations. From the data presented, it is obvious that the higher the number of articles available on the internet, the higher the number of citations. Lawrence (2001) calculated the mean of offline article citations and the result was 2.74, whereas the mean of on line article citation is 7.03, which means a 157 % increase. Thus, not only dissemination but also the visibility of researchers and that of their institutions, increase. Later on, Moed (2005) confirmed that. In time, the number of studies in a certain field grows, numerous articles studying the relation between downloads and citations (Morrison 2012; Harnad and Brody 2004; Zhang 2006; Eysenbach 2006; Koler-Povh et al. 2014; Gunasekaran and Arunachalam 2014; Schlögl et al. 2014; Bazrafshan et al. 2015), most of which prove the increase of visibility. Furthermore, Public Library of Science One Journal shows that in 2013, over a million articles were downloaded every month, and, consequently, about 20,000 articles have over 10 citations.
-
Free access facilitates connections among researchers, laboratories, allowing direct links through e-mail, and also group discussions based on materials available to everybody. If the case may be, the authors can modify the preprint or they can withdraw it from the scientific circuit. The articles are indexed and various connections among data bases can be made, considering common keywords, the same authors, the same topics etc.
-
Archives constitute stores which memorise researches at much lower costs than those of the traditional journals and, even if there are journals that require a fee to grant access, institutions can afford the costs.
-
The period of time necessary to perform peer-review evaluation is shorter for open access journals. We give the following examples: Molecules (the first decision within 24 days after submission), International Journal of Molecular Sciences (the first decision within 24 days after submission), Oxidative Medicine and Cellular Longevity (within 57 days till the publication), Journal of Analytical Methods in Chemistry (within 33 days), Neoplasia (within 21 days), Journal of Diabetes Research (within 45 days till the final decision), International Journal of Nanomedicine (within 17 days from submission to the first editorial decision) etc.
-
Electronic publication is rapid and the distribution space is enormous compared to the space covered by traditional journals.
-
The data presented on paper do not meet the handling exigency, especially when a lot of information needs to be analysed, assessed and processed.
-
Reputed journals which publish according to the open access system don’t sacrifice quality, a proof being their ratios of acceptance, defined as the number of articles accepted for publication related to the number of submitted articles. It ranges between 3 and about 70 %, similar to that of the subscription-based journals. Acceptance ratios considered high by the reader don’t necessarily lead to the impact factor decrease, i.e. of the scientific impact, and, consequently, we haven’t been able to establish a clear correlation between the acceptance ratio percentage and impact factors. Some examples referring to the acceptance percentage, according to the web pages of the journals under discussion, are presented in Table 1.
Acceptance ratios have been proposed as another objective method for measuring the journal quality, but they are an imprecise measure when they are applied to a single journal, without considering the field and its journals, or the method of evaluating the articles (blind review or editorial review) (Haensly et al. 2009).
-
Open access publishing journals have the possibility to inform their readers and authors regarding article metrics. Thus, Public Library of Science established that the following article information is relevant for its journals: viewed, cited, saved, discussed, recommended.
-
In order to argue in favour of their research, in their articles, the authors can use texts graphs, tables etc. taken from open access journals, citing the source, because the journals observe the Creative Commons Attribution Licence, CC BY NC ND 3.0 or CC BY 4.0 (Creative Commons 2015).
-
The concept ensures free and easy access to scientific information and to the first line of knowledge to poor countries and institutions, which cannot afford to subscribe to the numerous journals all over the world. We refer mainly to the scientific information in the medical and biomedical domains and also to the impressive number of open access journals (Kurata et al. 2013).
-
The open access journals provide the students with a rich and relevant documentation, and represent a viable support in elaborating works and in performing researches, ensuring in this manner the instruction through and for scientific research.
The disadvantages of using open access
-
The publication fees (article processing charges-APC) required for an article by journals are extremely high and poor institutions (universities, research entities) cannot afford to provide financial support to all potential authors (Manista 2012). Nevertheless, there is also the variant of paying the fees from the budget of research projects or grants. We also mention the praise-worthy case of the non-profit fundation Beilstein-Institut in Germany, who is the editor of two journals which are tax-free for the authors and their affiliations. Beilstein Journal of Organic Chemistry and Beilstein Journal of Nanotechnology are financed by the fundation.
The analysis of the fees shows a certain direct proportional relationship between them and the impact factor. Table 2 presents a list of the publication fees of some journals belonging to Hindawi Publishing and to Multidisciplinary Digital Publishing Institute, but also to other reputed publishers.
Some publishing houses make deductions for the authors from low-and middle-income countries and ensure publication fee assistance, e.g. Public Library of Science. Björk and Solomon (2012) establish correlations between citation averages for open access journals using article processing charges (APCs) versus those that are free to publish for authors, compared to impact factors for subscription journals and draw the conclusion that “APC funded OA journals” average impact increased markedly in the period 1996–2001 and to a lesser extent in 2002–2011, nearly reaching the same level as subscription journals at about 3.2.
-
Publication fees resulted in the appearance of numerous journals, some being quite old, but which haven’t been able to impose themselves in point of their visibility and prestige, thus, not being included in Journal of Citation Reports.
-
The researchers of the phenomenon consider that the peer–review process may undergo a decrease in exigency, the danger being that of lowering standards in favour of the profit (Kurata et al. 2013).
-
There is also the opinion that open access does not determine an increase of the citation number (Craig et al. 2007; Davis et al. 2008; Davis 2009). Thus, Craig et al. (2007) show that the journals published according to the open access system can attract a higher number of readers, compared to subscription-based journals, but did not mention any advantage of open access regarding the number of citations during the first year after the publishing. The same conclusion is also obtained as a result of the analysis of a number of 36 journals, covering the biological, medical, and multidisciplinary sciences, social sciences, and humanities (Davis 2011). Craig et al. (2007) point in fact that, apart from the accessibility to the paper, the number of citations is also decisively influenced by the scientific quality of the work, through its relevance and importance.
Ennas and Di Guardo (2015) have analysed a set of 1910 gold open access journals, by using Scopus SJR2012, DOAJ data base, and data provided by previous studies, concluding that the factors that influence the visibility are: language, country, years of activity and years in the DOAJ repository, publication fee, field of study, whether the journal has been launched as open access or converted, and type of publisher.
Wray (2016) appreciates that for social sciences and humanities, there are no data that could lead to a decisive conclusion that open access determines an increase of the citations number. The author cites 15 studies on this theme, from which only 7 present open access as an advantage for the increase of the citation number.
An analysis of the bibliometric performances of the journals published by Hindawi Publishing Corporation and Multidisciplinary Digital Publishing Institute (MDPI)
Our analysis is focused on journals belonging to Hindawi Publishing Corporation and Multidisciplinary Digital Publishing Institute (MDPI) considered as illustrative models for open access.
Hindawi publishes 405 journals, in fields like medicine, science, technology, social sciences, among which 51 are indexed in Journal Citation Reports (2015) and 65 in Emerging Sources Citation Index (ESCI 2015). The latter corresponds to a noteworthy percent of 16.05 % from the total number of journals.
We consider as main journal performance the number of citations (that allows the calculation of some scientometric indicators). From this standpoint, we will representatively analyse the performances of the journals belonging to Hindawi Publishing Corporation indexed JCR with calculated impact factor, and that also have data referring to the number of cited papers on 3 years intervals for each paper, for the period 2008–2013, calculated by SCImago (see Table 3).
Multidisciplinary Digital Publishing Institute (MDPI) publishes 160 journals, among which 30 are indexed in Web of Science (28 in Science Citation Index Expanded, one in Art and Humanities Citation Index, and one in both Social Sciences Citation Index and Science Citation Index Expanded) and 51 in Emerging Sources Citation Index. One should remark the admission of 51 journals in Emerging Sources Citation Index, that represents almost 32 % from the total numer of journals, and 39 % from the total number of journals non-indexed in Web of Science (130). In Table 4, the main bibliometric performances of MDPI journals are presented. For the analysis, we considered only the journals for which data could be collected, on at least the last 4 years. The sources of these data are: Journal Database, citefactor.org, researchgate.net and SCImago Journal and Country Rank.
The data presented in Table 3 show:
-
the number of cited documents is in general strongly related to the impact factor, although the number of cited documents does not represent the number of citations (a paper can receive more than one citation). So, if the number of cited documents increases, the impact factor will also increase and vice versa (see positions 2, 3, 4, 10 etc.);
-
a mild increasing trend of the number of cited documents in time but also of the impact factor, can be noticed (see positions 4, 5, 9, 10, 11, 15, 18,19, 20, 22-25, 28-30, 35, 36, 42). This moderate increase can be explained by the enhanced visibility ensured by open access;
-
the biomedical journals possess an increasing trend of the number of citable documents and of the impact factor (see 10, 18, 20, 35, 39 etc.);
-
from the total number of journals subject to analysis, those exhibiting a decreasing trend with respect to citation performances is relatively small;
The data presented in Table 4 show:
-
with few exceptions, the analysed journals present an increase of the cited articles percent in time (as the number of cited papers does not represent the number of citations).
-
a dependence of the impact factor on the cited articles percent cannot be noticed in all the cases; nevertheless, in some situations the increase of the number of cited papers leads to an impact factor increase (for instance positions 8, 10, 11), but there are also cases when an increase of the number of cited papers is accompanied by a decrease of the impact factor.
-
for the cases analysed, although the journals are open access, a significant increase of the immediacy index cannot be noticed. The latter varies within a relatively small interval (±0.1−0.2).
The common elements present in the bibliometric analysis of the journals presented in Tables 3 and 4, are:
-
the variation in time of the analysed bibliometric indicators is not great, and the noticed increases, when case, are relatively small.
-
for the cases analysed, a general direct relationship between the number of cited papers and the impact factor, as well as beween the immediacy index and the impact factor, cannot be established.
-
the performances of journals are not only ascribable to open access (visibility). As has been previously shown, the authors that have analysed the open access concept, proved that there are many factors, alongside visibility, that determine journal impact.
Conclusions
The debate on the topic Open Acces versus Non Open, though initiated about 10 years ago, continues up to now, resulting in many analyses and controversies. Most of the analyses point out the advantages of both authors and journals by increasing their visibility and impact. We agree that the higher the impact, in terms of the number of citations, the higher the visibility, and that open access ensures maximal visibility. Thus, there are many OA journals with rather high impact factors. At the same time, we should mention that there are also very many OA journals with a modest impact, referring to their number of citations in Google Scholar (through Publish or Perish), not to mention Web of Science citations.
Our analysis performed on several journals belonging to Hindawi Publishing Corporation and Multidisciplinary Digital Publishing Institute and following some bibliometric indicators, cannot precisely indicate whether the enhanced visibility provided by open access will lead to an increase of the citations number. We regard visibility as a necessary, but unsufficient condition for impact increase. Many researchers demonstrated that, apart from visibility, other elements influence the journals’ impact factors, and in this respect, it can be considered that the quality of the presented research is decisive.
The two analysed publishers present an important number of journals that entered in Emerging Sources Citation Index (16 % for Hindawi group and 32 % for MDPI, calculated from the total number of journals), that proves that visibility can result, to a certain extent, in bibliometric performances.
As well, the open access to scientific literature constitutes a real advantage for students—especially those found in the stages of performing researches and elaborating works—therefore it contributes to their training.
In our view, the greatest disadvantage is represented by publication fees, called by some journals Article Processing Charges, which become higher and higher. Consequently, it is necessary to establish a balance among publication fees, acceptance percentage and JRC impact factor, together with the diminishing of publication fees.
At the same time, we propose that those authors who attract a significant number of citations of the journal should benefit from a discount on Article Processing Charges. The best reason is that such authors increase the value of the journal.
Irrespective of the article value, the editors should ensure an ever greater visibility of the journals, providing a competent and exigent peer review.
A European Commission Communication identified open access as a core means to improve knowledge circulation and thus innovation in Europe, and established a set of measures to be applied at European level (Communication 2012).
As Máire Geoghegan-Quinn, former European Commissioner for Research, Innovation and Science, stated “Putting research results in the public sphere makes science better and strengthens our knowledge-based economy.”
References
Abrizah, A., Zainab, A. N., et al. (2013). Library and information science journals scientific impact and subject categorization: a comparison between Web of Science and Scopus. Scientometrics, 94(2), 721–740.
Annual Reviews rankings in Thomson Reuters JCR. (2015). http://www.annualreviews.org/page/about/isi-rankings. Accessed 11 May 2016.
Archambault, E., Campbell, D., et al. (2009). Comparing bibliometric statistics obtained from the Web of Science and Scopus. Journal of the American Society for Information Science and Technology, 60(7), 1320–1326.
Ardelean, A., Dobrescu, E. M., & Pisoschi, A. (2014). Prolegomena to science assessment-a scientometric approach, Chapter 7 (p. 160). Germany: Lambert Academic Publishing, Saarbrücken. ISBN 978-3-659-55086-7.
Ardent, J. (2010). Are article influence scores comparable across scientific fields? issues in science and technology librarianship. Winter,. doi:10.5062/F4FQ9TJW.
Bar-Illan, J., Levene, M., & Lin, A. (2007). Some measures for comparing citation databases. Journal of Infometrics, 1(1), 26–34.
Bazrafshan, A., Haghdoost, A. A., & Zare, M. (2015). A comparison of downloads, readership and citations data for the Journal of Medical Hypotheses and Ideas. Journal of Medical Hypotheses and Ideas, 9, 1–4.
Bergstrom, C. (2007). Eigenfactor—Measuring the value and prestige of scholarly journals. College and Research Libraries News, 68(5), 314–317.
Bergstrom, C. T., West, J. D., & Wiseman, M. A. (2008). The eigenfactor metrics. Journal of Neuroscience, 28, 11433–11434.
Björk, B-Ch., & Solomon, D. (2012). Open access versus subscription journals: A comparison of scientific impact. BioMed Central Medicine, 10, 73. doi:10.1186/1741-7015-10-73.
Bradford, S. C. (1934). Sources of information on specific subjects. Engineering, 137(3550), 85–86.
Braun, T. (2007). The impact factor of scientific and scholarly journals: its use and misuse in research evaluation: a selection of papers reprinted mainly from the Journal Scientometrics. Budapest: Akadémiai Kiadó.
Braun, T., Glänzel, W., & Schubert, A. (2006). A Hirsch-type index for journals. Scientometrics, 69(1), 169–173.
Budapest Open Access Initiative. (2002). http://www.budapestopenaccessinitiative.org/read. Accessed 20 February 2014.
Creative Commons. (2015). http://creativecommons.org/. Accessed 10 April 2016.
Communication from the Commission to the European Parliament, the Council, the European Economic and Social Committee and the Committee of the Regions. (2012). Towards better access to scientific information: Boosting the benefits of public investments in research, Brussels, 17.7.2012 COM(2012) 401 final, http://ec.europa.eu/research/science-society/document_library/pdf_06/era-communication-towards-better-access-to-scientific-information_en.pdf. Accessed 6 February 2015.
Craig, I. D., Plume, A. M., Mc Veight, M. E., et al. (2007). Do open access articles have greater citation impact? A critical review of the literature. Journal of Informetrics, 1, 239–248.
Davis, P. M. (2009). Author-choice open-access publishing in the biological and medical literature: A citation analysis. Journal of the American Society for Information Science and Technolog, 60(1), 3–8.
Davis, P. M. (2011). Open access, readership, citations: A randomized controlled trial of scientific journal publishing. The Journal of Federation of American Societies for Experimental Biology (FASEB Journal), 25, 1–6. doi:10.1096/fj.11-183988.
Davis, P. M., Lewenstein, B. V., Simon, D. H., et al. (2008). Open access publishing, article downloads, and citations: Randomised controlled trial. The British Medical Journal, 337(7665), 343–345.
De Solla Price, D. J. (1964). The science of science, chicago. In M. Goldsmith & A. Mackay (Eds.), The science of science. New York: Souvenir Press.
Egghe, L. (2006). Theory and practise of the g-index. Scientometrics, 69(1), 131–152.
Eigenfactor.org. (2016). www.eigenFACTOR.org. Accessed 22 February 2016.
Ennas, G., & Di Guardo, M. C. (2015). Features of top-rated gold open access journals: An analysisof the scopus database. Journal of Informetrics, 9, 79–89.
ESCI. (2015). http://ip-science.thomsonreuters.com/cgi-bin/jrnlst/jloptions.cgi?PC=EX. Accesed 07 May 2016
Eysenbach, G. (2006). Citation advantage of open access articles. Public Library of Science Biology, 4, e157. doi:10.1371/journal.pbio.0040157.
Falagas, M. E., Pitsouni, E. I., Malietzis, G. A., et al. (2008). Comparison of PubMed, Scopus, Web of Science, and Google Scholar: Strengths and weaknesses. The Journal of Federation of American Societies for Experimental Biology (FASEB Journal), 22, 338–342.
Garfield, E. (1955). Citation indexes for science: A new dimension in documentation through association of ideas. Science, 122(3159), 108–111.
Garfield, E. (1972). Citation analysis as a tool in journal evaluation. Science, 178(4060), 471–479.
Garfield, E. (1979). Citation indexing: Its theory and application in science, technology, and humanities. New York: Wiley.
Garfield, E. (1990). How ISI selects journals for coverage: Quantitative and qualitative considerations. Current Contents, 22, 5–13.
Glänzel, W., & Thijs, B. (2004). The influence of author self-citations on bibliometric macro indicators. Scientometrics, 59(3), 281–310.
Gunasekaran, S., & Arunachalam, S. (2014). The impact factors of open access and subscription journals across fields. Current Science, 107(3), 380–388.
Haensly, P. J., Hodges, P. E., & Davenport, S. A. (2009). Acceptance rates and journal quality: An analysis of journals in economics and finance. Journal of Business and Finance Librarianship, 14(1), 2–31.
Harnad, S., & Brody, T. (2004). Comparing the impact of Open Access (OA) vs. Non-OA articles in the same journals. D-Lib Magazine, June 2004, 10(6) http://www.dlib.org/dlib/june04/harnad/06harnad.html. Accessed 07 February 2015
Harzing, A.W. (2007). Publish or Perish. http://www.harzing.com/pop.htm. Accessed 10 February 2015.
Hirsch, J. E. (2005). An index to quantify an individual’s scietific research output. Proceedings of the National Academy of Sciences, 102(46), 16569–16572.
Institut de France, Académie des Sciences. (2015). http://www.academie-sciences.fr/activite/cr.htm. Accessed 07 November 2015.
Journal Database. (2016). http://www.journal-database.com/journal/. Accessed 03 May 2016
Journal des Savants. (2015). http://www.aibl.fr/publications/collections/journal-des-savants/?lang=fr. Accessed 07 November 2015.
Khabsa, M., & Giles, C. L. (2014). The number of scholarly documents on the public web. Public Library of Science One, 9(5), e93949. doi:10.1371/journal.pone.0093949.
Koler-Povh, T., Juznic, P., & Turk, G. (2014). Impact of open access on citation of scholarly publications in the field of civil engineering. Scientometrics, 98, 1033–1045.
Kulkarni, A. V., Aziz, B., Shams, I., et al. (2009). Comparisons of citations in Web of Science, Scopus, and Google Scholar for articles published in general medical journals. The Journal of the American Medical Association, 302(10), 1092–1096.
Kurata, K., Morioka, T., Yokoi, K., et al. (2013). Remarkable growth of Open Access in the biomedical field: analysis of PubMed articles from 2006 to 2010. Public Library of Science One, 8(5), e60925. doi:10.1371/journal.pone.0060925.
Lawrence, S. (2001). Free online availability substantially increases a paper’s impact. Nature, 31 May, http://www.nature.com/nature/debates/e-access/Articles/lawrence.html. Accessed 06 February 2016.
Lotka, A. J. (1926). The frecvency distribution of scientific productivity. Journal of the Washington Academy of Sciences, 16(12), 317–323.
Manista, F. C. (2012). Open don’t mean free: A reflection on the potential advantages and disadvantages of publishing research via Open Access. Journal of Librarianship and Scholarly Communication, 1(2), eP1049. doi:10.7710/2162-3309.1049.
Moed, H. F. (2005). Citation analysis in research evaluation. The Netherlands: Springer.
Moed, H. F. (2010). Measuring contextual citation impact of scientific journals. Journal of Infometrics, 4(3), 265–277.
Morrison, H. (2012). About 30% of peer-reviewed scholarly journals are now open access. The Imaginary Journal of Poetic Economics. http://poeticeconomics.blogspot.ro/2012/05/about-30-of-peer-reviewed-scholarly.html. Accessed 27 February 2015
Okubo, Y. (1997). Indicateurs bibliometriques et analyse des systemes de recherche. Documents de travail de la Direction de la science, de la technologie et de l’industrie O.C.D.E., Paris.
Philosophical Transactions. (2015). http://rstl.royalsocietypublishing.org/. Accessed 30 November 2015
Rousseau, R. (2002). Journal evaluation: technical and practical issues. Library Trends, 50(3), 418–439.
Schlögl, C., Gorraiz, J., Gumpenberger, C., et al. (2014). Comparison of downloads, citations and readership data for two information systems journals. Scientometrics, 101, 1113–1128.
SCImago Journal & Country Rank. (2016). http://www.scimagojr.com/. Accessed 12 June 2016
Statistics Canada. (2015). Table 477-0059-Financial information of universities and degree-granting colleges, expenditures by type of fund and function, annual (dollars), CANSIM (database). http://www5.statcan.gc.ca/cansim/a26?lang=eng&retrLang=eng&id=4770059&&pattern=&stByVal=1&p1=1&p2=-1&tabMode=dataTable&csid=. Accessed 22 February 2016.
Vishwakarma, P., & Mukherjee, B. (2014). Developing qualitative indicators for journal evaluation: Case study of Library Science Journals of SAARC countries. DESIDOC Journal of Library and Information Technology, 34(2), 152–161.
Walters, W. H. (2008). Journal prices, book acquisitions, and sustainable college library collections. College and Research Libraries, 2008, 576–585.
Wray, K. B. (2016). No new evidence for a citation benefit for Author-Pay Open Access Publications in the social sciences and humanities. Scientometrics, 106, 1031–1035.
Zhang, Y. (2006). The effect of open access on citation impact: a comparison study based on Web citation analysis. Libri, 56(3), 145–156.
Zhang, Ch-T. (2009). The e-Index, complementing the h-Index for excess citations. Public Library Of Science One, 4(5), e5429. doi:10.1371/journal.pone.0005429.
Zwemer, R. L. (1970). Identification of journal characteristics useful in improving input and output of a retrieval system. Federation Proceedings, 29, 1595–1604.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Pisoschi, A.M., Pisoschi, C.G. Is open access the solution to increase the impact of scientific journals?. Scientometrics 109, 1075–1095 (2016). https://doi.org/10.1007/s11192-016-2088-x
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11192-016-2088-x