Impact factor

The impact factor (IF) or journal impact factor (JIF) of an academic journal is a measure reflecting the yearly average number of citations to recent articles published in that journal. It is frequently used as a proxy for the relative importance of a journal within its field; journals with higher impact factors are often deemed to be more important than those with lower ones. The impact factor was devised by Eugene Garfield, the founder of the Institute for Scientific Information. Impact factors are calculated yearly starting from 1975 for journals listed in the Journal Citation Reports.

Calculation

In any given year, the impact factor of a journal is the number of citations, received in that year, of articles published in that journal during the two preceding years, divided by the total number of articles published in that journal during the two preceding years:[1]

For example, Nature had an impact factor of 41.456 in 2014:[2]

This means that, on average, its papers published in 2012 and 2013 received roughly 41 citations each in 2014. Note that 2014 impact factors are reported in 2015; they cannot be calculated until all of the 2014 publications have been processed by the indexing agency.

New journals, which are indexed from their first published issue, will receive an impact factor after two years of indexing; in this case, the citations to the year prior to Volume 1, and the number of articles published in the year prior to Volume 1, are known zero values. Journals that are indexed starting with a volume other than the first volume will not get an impact factor until they have been indexed for three years. Occasionally, Journal Citation Reports assigns an impact factor to new journals with less than two years of indexing, based on partial citation data.[3][4] The calculation always uses two complete and known years of item counts, but for new titles one of the known counts is zero. Annuals and other irregular publications sometimes publish no items in a particular year, affecting the count. The impact factor relates to a specific time period; it is possible to calculate it for any desired period. For example, the Journal Citation Reports (JCR) also includes a five-year impact factor, which is calculated by dividing the number of citations to the journal in a given year by the number of articles published in that journal in the previous five years.[5][6]

Use

The impact factor is used to compare different journals within a certain field. The Web of Science indexes more than 11,500 science and social science journals.[7]

Journal impact factors are often used to evaluate the merit of individual articles and individual researchers. This particular use of impact factors was summarised by Hoeffel:[8]

Impact Factor is not a perfect tool to measure the quality of articles but there is nothing better and it has the advantage of already being in existence and is, therefore, a good technique for scientific evaluation. Experience has shown that in each specialty the best journals are those in which it is most difficult to have an article accepted, and these are the journals that have a high impact factor. Most of these journals existed long before the impact factor was devised. The use of impact factor as a measure of quality is widespread because it fits well with the opinion we have in each field of the best journals in our specialty.

...

In conclusion, prestigious journals publish papers of high level. Therefore, their impact factor is high, and not the contrary.

As impact factors is a journal-level metric, rather than an article or individual level metric, this use is controversial. Garfield agrees with Hoeffel,[9] but warns about the "misuse in evaluating individuals" because there is "a wide variation [of citations] from article to article within a single journal".[10]

Some companies are producing false impact factors. According to an article published in the United States National Library of Medicine, these include Global Impact Factor (GIF), Citefactor, and Universal Impact Factor (UIF).[11]

Criticisms

Numerous criticisms have been made regarding the use of impact factors.[12][13][14] For one thing, the impact factor might not be consistently reproduced in an independent audit.[15] There is also a more general debate on the validity of the impact factor as a measure of journal importance and the effect of policies that editors may adopt to boost their impact factor (perhaps to the detriment of readers and writers). Other criticism focuses on the effect of the impact factor on behavior of scholars, editors and other stakeholders.[16][17] Others have made more general criticisms, arguing that emphasis on impact factor results from negative influence of neoliberal policies on academia claiming that what is needed is not just replacement of the impact factor with more sophisticated metrics for science publications but also discussion on the social value of research assessment and the growing precariousness of scientific careers in higher education.[18][19][20]

Validity as a measure of importance

It has been stated that impact factors and citation analysis in general are affected by field-dependent factors[21] which may invalidate comparisons not only across disciplines but even within different fields of research of one discipline.[22] The percentage of total citations occurring in the first two years after publication also varies highly among disciplines from 1–3% in the mathematical and physical sciences to 5–8% in the biological sciences.[23] Thus impact factors cannot be used to compare journals across disciplines.

Because citation counts have highly skewed distributions,[24] the mean number of citations is potentially misleading if used to gauge the typical impact of articles in the journal rather than the overall impact of the journal itself.[25] For example, about 90% of Nature's 2004 impact factor was based on only a quarter of its publications, and thus the actual number of citations for a single article in the journal is in most cases much lower than the mean number of citations across articles.[26] Furthermore, the strength of the relationship between impact factors of journals and the citation rates of the papers therein has been steadily decreasing since articles began to be available digitally.[27]

Indeed, impact factors are sometimes used to evaluate not only the journals but the papers therein, thereby devaluing papers in certain subjects.[28] The Higher Education Funding Council for England was urged by the House of Commons Science and Technology Select Committee to remind Research Assessment Exercise panels that they are obliged to assess the quality of the content of individual articles, not the reputation of the journal in which they are published.[29] The effect of outliers can be seen in the case of the article "A short history of SHELX", which included this sentence: "This paper could serve as a general literature citation when one or more of the open-source SHELX programs (and the Bruker AXS version SHELXTL) are employed in the course of a crystal-structure determination". This article received more than 6,600 citations. As a consequence, the impact factor of the journal Acta Crystallographica Section A rose from 2.051 in 2008 to 49.926 in 2009, more than Nature (at 31.434) and Science (at 28.103).[30] The second-most cited article in Acta Crystallographica Section A in 2008 only had 28 citations.[31] Also, impact factor is a journal metric and should not be used to assess individual researchers or institutions.[32][33]

Journal rankings constructed based solely on impact factors only moderately correlate with those compiled from the results of expert surveys.[34]

A.E. Cawkell, sometime Director of Research at the Institute for Scientific Information remarked that the Science Citation Index (SCI), on which the impact factor is based, "would work perfectly if every author meticulously cited only the earlier work related to his theme; if it covered every scientific journal published anywhere in the world; and if it were free from economic constraints."[35]

Editorial policies that affect the impact factor

A journal can adopt editorial policies to increase its impact factor.[36][37] For example, journals may publish a larger percentage of review articles which generally are cited more than research reports.[38] Thus review articles can raise the impact factor of the journal and review journals will therefore often have the highest impact factors in their respective fields.[17] Some journal editors set their submissions policy to "by invitation only" to invite exclusively senior scientists to publish "citable" papers to increase the journal impact factor.[17]

Journals may also attempt to limit the number of "citable items"—i.e., the denominator of the impact factor equation—either by declining to publish articles that are unlikely to be cited (such as case reports in medical journals) or by altering articles (e.g., by not allowing an abstract or bibliography in hopes that Journal Citation Reports will not deem it a "citable item"). As a result of negotiations over whether items are "citable", impact factor variations of more than 300% have been observed.[39] Items considered to be uncitable—and thus are not incorporated in impact factor calculations—can, if cited, still enter into the numerator part of the equation despite the ease with which such citations could be excluded. This effect is hard to evaluate, for the distinction between editorial comment and short original articles is not always obvious. For example, letters to the editor may refer to either class.

Another less insidious tactic journals employ is to publish a large portion of its papers, or at least the papers expected to be highly cited, early in the calendar year. This gives those papers more time to gather citations. Several methods, not necessarily with nefarious intent, exist for a journal to cite articles in the same journal which will increase the journal's impact factor.[40][41]

Beyond editorial policies that may skew the impact factor, journals can take overt steps to game the system. For example, in 2007, the specialist journal Folia Phoniatrica et Logopaedica, with an impact factor of 0.66, published an editorial that cited all its articles from 2005 to 2006 in a protest against the "absurd scientific situation in some countries" related to use of the impact factor.[42] The large number of citations meant that the impact factor for that journal increased to 1.44. As a result of the increase, the journal was not included in the 2008 and 2009 Journal Citation Reports.[43]

Coercive citation is a practice in which an editor forces an author to add extraneous citations to an article before the journal will agree to publish it, in order to inflate the journal's impact factor. A survey published in 2012 indicates that coercive citation has been experienced by one in five researchers working in economics, sociology, psychology, and multiple business disciplines, and it is more common in business and in journals with a lower impact factor.[44] However, cases of coercive citation have occasionally been reported for other disciplines.[45]

Responses

Because "the impact factor is not always a reliable instrument", in November 2007 the European Association of Science Editors (EASE) issued an official statement recommending "that journal impact factors are used only—and cautiously—for measuring and comparing the influence of entire journals, but not for the assessment of single papers, and certainly not for the assessment of researchers or research programmes".[13]

In July 2008, the International Council for Science (ICSU) Committee on Freedom and Responsibility in the Conduct of Science (CFRS) issued a "statement on publication practices and indices and the role of peer review in research assessment", suggesting many possible solutions—e.g., considering a limit number of publications per year to be taken into consideration for each scientist, or even penalising scientists for an excessive number of publications per year—e.g., more than 20.[46]

In February 2010, the Deutsche Forschungsgemeinschaft (German Research Foundation) published new guidelines to evaluate only articles and no bibliometric information on candidates to be evaluated in all decisions concerning "performance-based funding allocations, postdoctoral qualifications, appointments, or reviewing funding proposals, [where] increasing importance has been given to numerical indicators such as the h-index and the impact factor".[47] This decision follows similar ones of the National Science Foundation (US) and the Research Assessment Exercise (UK).

In response to growing concerns over the inappropriate use of journal impact factors in evaluating scientific outputs and scientists themselves, the American Society for Cell Biology together with a group of editors and publishers of scholarly journals created the San Francisco Declaration on Research Assessment (DORA). Released in May 2013, DORA has garnered support from thousands of individuals and hundreds of institutions,[20] including in March 2015 the League of European Research Universities (a consortium of 21 of the most renowned research universities in Europe),[48] who have endorsed the document on the DORA website.

Université de Montréal, Imperial College London, PLOS, eLife, EMBO Journal, The Royal Society, Nature and Science proposed citation distributions metrics as alternative to impact factors.[49][50][51]

Closely related indices

Some related values, also calculated and published by the same organization, include:

  • Cited half-life: the median age of the articles that were cited in Journal Citation Reports each year. For example, if a journal's half-life in 2005 is 5, that means the citations from 2001-2005 are half of all the citations from that journal in 2005, and the other half of the citations precede 2001.[52]
  • Aggregate impact factor for a subject category: it is calculated taking into account the number of citations to all journals in the subject category and the number of articles from all the journals in the subject category.
  • Immediacy index: the number of citations the articles in a journal receive in a given year divided by the number of articles published.

As with the impact factor, there are some nuances to this: for example, ISI excludes certain article types (such as news items, correspondence, and errata) from the denominator.[53][54][55]

Other measures of impact

Additional journal-level metrics are available from other organizations. For example, CiteScore: is a metric for serial titles in Scopus launched in December 2016 by Elsevier.[56][57] While these metrics apply only to journals, there are also author-level metrics, such as the H-index, that apply to individual researchers. In addition, article-level metrics measure impact at an article level instead of journal level. Other more general alternative metrics, or "altmetrics", may include article views, downloads, or mentions in social media.

Counterfeit

Fake impact factors are produced by companies not affiliated with Journal Citation Reports.[58] These are often used by predatory publishers;[59] Jeffrey Beall maintained a list of such misleading metrics.[60] Consulting Journal Citation Reports' master journal list can confirm if a publication is indexed by Journal Citation Reports, which is a necessary (but not sufficient) condition for obtaining an IF.[61] Use of fake impact metrics is considered a "red flag".[62]

See also

References

  1. ^ "Journal Citation Reports: Impact Factor". Retrieved 2016-09-12.
  2. ^ "Nature". 2014 Journal Citation Reports. Web of Science (Science ed.). Thomson Reuters. 2015.
  3. ^ "RSC Advances receives its first partial impact factor". RSC Advances Blog. 24 June 2013. Retrieved 16 July 2018.
  4. ^ "Our first (partial) impact factor and our continuing (full) story". news.cell.com. 30 July 2014. Archived from the original on 7 March 2016. Retrieved 21 May 2015.
  5. ^ "JCR with Eigenfactor". Archived from the original on 2010-01-02. Retrieved 2009-08-26.
  6. ^ "ISI 5-Year Impact Factor". APA. Retrieved 2017-11-12.
  7. ^ "Every journal has a story to tell". Journal Citation Reports. Clarivate Analytics. Retrieved 2019-03-15.
  8. ^ Hoeffel, C. (1998). "Journal impact factors". Allergy. 53 (12): 1225. doi:10.1111/j.1398-9995.1998.tb03848.x. ISSN 0105-4538. PMID 9930604.
  9. ^ Garfield, Eugene (2006-01-04). "The History and Meaning of the Journal Impact Factor". JAMA. 295 (1): 90–3. doi:10.1001/jama.295.1.90. ISSN 0098-7484. PMID 16391221.
  10. ^ Eugene Garfield (June 1998). "The Impact Factor and Using It Correctly". Der Unfallchirurg. 101 (6): 413–414. PMID 9677838.
  11. ^ Jalalian, M (2015). "The story of fake impact factor companies and how we detected them". Electronic Physician. 7 (2): 1069–72. doi:10.14661/2015.1069-1072. PMC 4477767. PMID 26120416.
  12. ^ "Time to remodel the journal impact factor". Nature. 535 (466): 466. 2016. Bibcode:2016Natur.535..466.. doi:10.1038/535466a. PMID 27466089.
  13. ^ a b "European Association of Science Editors (EASE) Statement on Inappropriate Use of Impact Factors". Retrieved 2012-07-23.
  14. ^ Callaway, Ewen (2016-07-14). "Beat it, impact factor! Publishing elite turns against controversial metric". Nature. 535 (7611): 210–211. Bibcode:2016Natur.535..210C. doi:10.1038/nature.2016.20224. PMID 27411614.
  15. ^ Rossner, M.; Van Epps, H.; Hill, E. (17 December 2007). "Show me the data". Journal of Cell Biology. 179 (6): 1091–2. doi:10.1083/jcb.200711140. PMC 2140038. PMID 18086910.
  16. ^ Wesel, M. van (2016). "Evaluation by Citation: Trends in Publication Behavior, Evaluation Criteria, and the Strive for High Impact Publications". Science and Engineering Ethics. 22 (1): 199–225. doi:10.1007/s11948-015-9638-0. PMC 4750571. PMID 25742806.
  17. ^ a b c Moustafa, Khaled (2015). "The disaster of the impact factor". Science and Engineering Ethics. 21 (1): 139–142. doi:10.1007/s11948-014-9517-0. PMID 24469472.
  18. ^ Brembs, B.,; Button, K.; Munafò, M. (2013). "Deep impact: Unintended consequences of journal rank". Frontiers in Human Neuroscience. 7 (291): 1–12. doi:10.3389/fnhum.2013.00291. PMC 3690355. PMID 23805088.
  19. ^ Kansa, Eric (27 January 2014). "It's the Neoliberalism, Stupid: Why instrumentalist arguments for Open Access, Open Data, and Open Science are not enough". LSE Impact Blog. Retrieved 16 July 2018.
  20. ^ a b Cabello, F.; Rascón, M.T. (2015). "The Index and the Moon. Mortgaging Scientific Evaluation". International Journal of Communication. 9: 1880–1887.
  21. ^ Bornmann, L.; Daniel, H. D. (2008). "What do citation counts measure? A review of studies on citing behavior". Journal of Documentation. 64 (1): 45–80. doi:10.1108/00220410810844150.
  22. ^ Anauati, Maria Victoria; Galiani, Sebastian; Gálvez, Ramiro H. (November 11, 2014). "Quantifying the Life Cycle of Scholarly Articles Across Fields of Economic Research". SSRN 2523078.
  23. ^ Erjen van Nierop (2009). "Why Do Statistics Journals Have Low Impact Factors?". Statistica Neerlandica. 63 (1): 52–62. doi:10.1111/j.1467-9574.2008.00408.x.
  24. ^ Callaway, Ewen (14 July 2016). "Beat it, impact factor! Publishing elite turns against controversial metric". Nature. 535 (7611): 210–211. Bibcode:2016Natur.535..210C. doi:10.1038/nature.2016.20224. PMID 27411614. Retrieved 10 December 2016.
  25. ^ Joint Committee on Quantitative Assessment of Research (12 June 2008). "Citation Statistics" (PDF). International Mathematical Union.
  26. ^ "Not-so-deep impact". Nature. 435 (7045): 1003–1004. 23 June 2005. doi:10.1038/4351003b. PMID 15973362.
  27. ^ Lozano, George A.; Larivière, Vincent; Gingras, Yves (2012). "The weakening relationship between the impact factor and papers' citations in the digital age". Journal of the American Society for Information Science and Technology. 63 (11): 2140–2145. arXiv:1205.4328. doi:10.1002/asi.22731.
  28. ^ John Bohannon (2016). "Hate journal impact factors? New study gives you one more reason". Science. doi:10.1126/science.aag0643.
  29. ^ "House of Commons – Science and Technology – Tenth Report". 2004-07-07. Retrieved 2008-07-28.
  30. ^ Grant, Bob (21 June 2010). "New impact factors yield surprises". The Scientist. Retrieved 31 March 2011.
  31. ^ mmcveigh (17 June 2010). "What does it mean to be #2 in Impact?". Retrieved 2018-07-16.
  32. ^ Seglen, P. O. (1997). "Why the impact factor of journals should not be used for evaluating research". BMJ. 314 (7079): 498–502. doi:10.1136/bmj.314.7079.497. PMC 2126010. PMID 9056804.
  33. ^ "EASE Statement on Inappropriate Use of Impact Factors". European Association of Science Editors. November 2007. Retrieved 2013-04-13.
  34. ^ Serenko, A.; Dohan, M. (2011). "Comparing the expert survey and citation impact journal ranking methods: Example from the field of Artificial Intelligence" (PDF). Journal of Informetrics. 5 (4): 629–648. doi:10.1016/j.joi.2011.06.002.
  35. ^ Cawkell, Anthony E. (1977). "Science perceived through the Science Citation Index". Endeavour. 1 (2): 57–62. doi:10.1016/0160-9327(77)90107-7. PMID 71986.
  36. ^ Monastersky, Richard (14 October 2005). "The Number That's Devouring Science". The Chronicle of Higher Education.
  37. ^ Arnold, Douglas N.; Fowler, Kristine K. (2011). "Nefarious Numbers". Notices of the American Mathematical Society. 58 (3): 434–437. arXiv:1010.0278. Bibcode:2010arXiv1010.0278A.
  38. ^ Garfield, Eugene (20 June 1994). "The Thomson Reuters Impact Factor". Thomson Reuters.
  39. ^ PLoS Medicine Editors (6 June 2006). "The Impact Factor Game". PLoS Medicine. 3 (6): e291. doi:10.1371/journal.pmed.0030291. PMC 1475651. PMID 16749869.CS1 maint: Extra text: authors list (link)
  40. ^ Agrawal, A. (2005). "Corruption of Journal Impact Factors" (PDF). Trends in Ecology and Evolution. 20 (4): 157. doi:10.1016/j.tree.2005.02.002. PMID 16701362.
  41. ^ Fassoulaki, A.; Papilas, K.; Paraskeva, A.; Patris, K. (2002). "Impact factor bias and proposed adjustments for its determination". Acta Anaesthesiologica Scandinavica. 46 (7): 902–5. doi:10.1034/j.1399-6576.2002.460723.x. PMID 12139549.
  42. ^ Schuttea, H. K.; Svec, J. G. (2007). "Reaction of Folia Phoniatrica et Logopaedica on the Current Trend of Impact Factor Measures". Folia Phoniatrica et Logopaedica. 59 (6): 281–285. doi:10.1159/000108334. PMID 17965570.
  43. ^ "Journal Citation Reports – Notices". Archived from the original on 2010-05-15. Retrieved 2009-09-24.
  44. ^ Wilhite, A. W.; Fong, E. A. (2012). "Coercive Citation in Academic Publishing". Science. 335 (6068): 542–3. Bibcode:2012Sci...335..542W. doi:10.1126/science.1212540. PMID 22301307.
  45. ^ Smith, Richard (1997). "Journal accused of manipulating impact factor". BMJ. 314 (7079): 463. doi:10.1136/bmj.314.7079.461d. PMC 2125988. PMID 9056791.
  46. ^ "International Council for Science statement". Icsu.org. 2014-05-02. Retrieved 2014-05-18.
  47. ^ "Quality not Quantity: DFG Adopts Rules to Counter the Flood of Publications in Research". Deutsche Forschungsgemeinschaft. 23 February 2010. Retrieved 2018-07-16.
  48. ^ "Not everything that can be counted counts …". League of European Research Universities. 16 March 2015. Archived from the original on 2017-12-01.
  49. ^ Veronique Kiermer (2016). "Measuring Up: Impact Factors Do Not Reflect Article Citation Rates". PLOS.
  50. ^ "Ditching Impact Factors for Deeper Data". The Scientist. Retrieved 2016-07-29.
  51. ^ Corneliussen, Steven T. (2016). "Bad summer for the journal impact factor". Physics Today. doi:10.1063/PT.5.8183.
  52. ^ "Impact Factor, Immediacy Index, Cited Half-life". Swedish University of Agricultural Sciences. Archived from the original on 23 May 2008. Retrieved 30 October 2016.
  53. ^ "Bibliometrics (journal measures)". Elsevier. Archived from the original on 2012-08-18. Retrieved 2012-07-09. a measure of the speed at which content in a particular journal is picked up and referred to
  54. ^ "Glossary of Thomson Scientific Terminology". Thomson Reuters. Retrieved 2012-07-09.
  55. ^ "Journal Citation Reports Contents -- Immediacy Index" ((online)). Clarivate Analytics. Retrieved 2012-07-09. The Immediacy Index is the average number of times an article is cited in the year it is published. The journal Immediacy Index indicates how quickly articles in a journal are cited. The aggregate Immediacy Index indicates how quickly articles in a subject category are cited.
  56. ^ Elsevier. "Metrics - Features - Scopus - Solutions | Elsevier". www.elsevier.com. Retrieved 2016-12-09.
  57. ^ Van Noorden, Richard (2016). "Controversial impact factor gets a heavyweight rival". Nature. 540 (7633): 325–326. Bibcode:2016Natur.540..325V. doi:10.1038/nature.2016.21131. PMID 27974784.
  58. ^ Jalalian M (2015). "The story of fake impact factor companies and how we detected them". Electronic Physician. 7 (2): 1069–72. doi:10.14661/2015.1069-1072. PMC 4477767. PMID 26120416.
  59. ^ Jeffrey Beall. "Scholarly Open-Access - Fake impact factors". Archived from the original on 2016-03-21.
  60. ^ Misleading Metrics Archived 2017-01-11 at the Wayback Machine
  61. ^ "Thomson Reuters Intellectual Property & Science Master Journal List".
  62. ^ Ebrahimzadeh, Mohammad H. (April 2016). "Validated Measures of Publication Quality: Guide for Novice Researchers to Choose an Appropriate Journal for Paper Submission". Archives of Bone and Joint Surgery. 4 (2): 94–96. PMC 4852052. PMID 27200383.

Further reading

External links

American Journal of Botany

The American Journal of Botany is a monthly peer-reviewed scientific journal which covers all aspects of plant biology. It has been published by the Botanical Society of America since 1914. According to the Journal Citation Reports, the journal has a 2012 impact factor of 2.586. Access is available through JSTOR with a moving wall of 5 years.

American Journal of Mathematics

The American Journal of Mathematics is a bimonthly mathematics journal published by the Johns Hopkins University Press.

Animal Cognition

Animal Cognition is a peer-reviewed scientific journal published by Springer Science+Business Media. It covers research in ethology, behavioral ecology, animal behavior, cognitive sciences, and all aspects of human and animal cognition. According to the Journal Citation Reports, the journal has a 2017 impact factor of 2.805.

Australian Systematic Botany

Australian Systematic Botany is an international peer-reviewed scientific journal published by CSIRO Publishing. It is devoted to publishing original research, and sometimes review articles, on topics related to systematic botany, such as biogeography, taxonomy and evolution. The journal is broad in scope, covering all plant, algal and fungal groups, including fossils.

First published in 1978 as Brunonia, the journal adopted its current name in 1988.

The current Editor-in-Chief is Daniel Murphy (Royal Botanic Gardens Melbourne).

Cell (journal)

Cell is a peer-reviewed scientific journal publishing research papers across a broad range of disciplines within the life sciences. Areas covered include molecular biology, cell biology, systems biology, stem cells, developmental biology, genetics and genomics, proteomics, cancer research, immunology, neuroscience, structural biology, microbiology, virology, physiology, biophysics, and computational biology. The journal was established in 1974 by Benjamin Lewin and is published twice monthly by Cell Press, an imprint of Elsevier.

Gene (journal)

Gene is a peer-reviewed scientific journal in genetics and molecular biology, focusing on the cloning, structure, function, as well as the biomedical and biotechnological importance of genes. It was established in 1976 and is published by Elsevier.According to the Journal Citation Reports, the journal has a 2015 impact factor of 2.319.. The 2017 impact factor of the journal is 2.498.

Beyond Gene, there are several sub-specialty journals linked to Gene including Meta Gene, Plant Gene, Agri Gene and Gene Reports.

Geology (journal)

Geology is a peer-reviewed publication of the Geological Society of America (GSA). GSA claims that it is the most widely read scientific journal in the field of earth science. It is published monthly, with each issue containing 20 or more articles. In 2017, the journal's impact factor was 4.635.One of the goals of the journal is to provide a forum for shorter articles (four pages each) and less focus on purely academic research–type articles.

Journal of Biological Chemistry

The Journal of Biological Chemistry is a weekly peer-reviewed scientific journal that was established in 1905. Since 1925, it is published by the American Society for Biochemistry and Molecular Biology. It covers research in areas of biochemistry and molecular biology. The editor-in-chief is Lila Gierasch. All its articles are available free after one year of publication. In press articles are available free on its website immediately after acceptance.

Journal of Ornithology

The Journal of Ornithology (formerly Journal für Ornithologie) is a scientific journal published by Springer Science+Business Media on behalf of the Deutsche Ornithologen-Gesellschaft. It was founded by Jean Cabanis in 1853, becoming the official journal of the Deutsche Ornithologen-Gesellschaft in 1854.

The first issue was produced in January 1853 and Cabanis noted that although there were specialist journals in entomology and conchology that there was nothing to deal with ornithology in Germany. Among the first essays published in the journal was an essay by Reichenbach on the concept of species.According to the Journal Citation Reports, the journal has a 2012 impact factor of 1.632.

Journal of Vertebrate Paleontology

The Journal of Vertebrate Paleontology is a bimonthly peer-reviewed scientific journal that was established in 1980 by Jiri Zidek (University of Oklahoma). It covers all aspects of vertebrate paleontology, including vertebrate origins, evolution, functional morphology, taxonomy, biostratigraphy, paleoecology, paleobiogeography, and paleoanthropology. The journal is published by Taylor & Francis on behalf of the Society of Vertebrate Paleontology. According to the Journal Citation Reports, the journal has a 2017 impact factor of 2.190.

Malacologia

Malacologia is a peer-reviewed scientific journal in the field of malacology, the study of mollusks. The journal publishes articles in the fields of molluscan systematics, ecology, population ecology, genetics, molecular genetics, evolution, and phylogenetics.The journal specializes in publishing long papers and monographs. The journal publishes at least one, sometimes two, volumes of about 400 pages per year, which may consist of 1 or 2 issues. According to the Journal Citation Reports, its 2010 impact factor is 1.024. This ranks Malacologia 66th out of 145 listed journals in the category "Zoology". The journal started publication in 1962.

Nature Genetics

Nature Genetics is a scientific journal founded as part of the Nature family of journals in 1992. It publishes high quality research in genetics.

The journal encompasses genetic and functional genomic studies on human traits and on other model organisms, including mouse, fly, nematode and yeast. Current emphasis is on the genetic basis for common and complex diseases and on the functional mechanism, architecture and evolution of gene networks, studied by experimental perturbation.

It is an internationally recognized scientific publication with an Impact Factor of 27.959, making it the second ranked journal in the category of genetics and heredity, second to Nature Reviews Genetics, so first in research. Its sister journal is Nature Reviews Genetics.

Palaeontology (journal)

Palaeontology is one of the two scientific journals of the Palaeontological Association (the other being Papers in Palaeontology). It was established in 1957 and is published on behalf of the Association by Wiley-Blackwell. The editor-in-chief is Andrew Smith (Natural History Museum, London). Palaeontology publishes articles on a range of palaeontological topics, including taphonomy, functional morphology, systematics, palaeo-environmental reconstruction and biostratigraphy. According to the Journal Citation Reports, the journal has a 2017 impact factor of 3.730, ranking it 1st out of 55 journals in the category "Paleontology".

Physical Review Letters

Physical Review Letters (PRL), established in 1958, is a peer-reviewed, scientific journal that is published 52 times per year by the American Physical Society. As also confirmed by various measurement standards, which include the Journal Citation Reports impact factor and the journal h-index proposed by Google Scholar, many physicists and other scientists consider Physical Review Letters to be one of the most prestigious journals in the field of physics.PRL is published as a print journal, and is in electronic format, online and CD-ROM. Its focus is rapid dissemination of significant, or notable, results of fundamental research on all topics related to all fields of physics. This is accomplished by rapid publication of short reports, called "Letters". Papers are published and available electronically one article at a time. When published in such a manner, the paper is available to be cited by other work. The Lead Editor is Hugues Chaté. The Managing Editor is Reinhardt B. Schuhmann.

Reviews of Modern Physics

Reviews of Modern Physics is a quarterly peer-reviewed scientific journal published by the American Physical Society. It was established in 1929 and the current editor-in-chief is Michael Thoennessen. The journal publishes review articles, usually by established researchers, on all aspects of physics and related fields. The reviews are usually accessible to non-specialists and serve as introductory material to graduate students, which survey recent work, discuss key problems to be solved and provide perspectives toward the end. RMP is arguably one of the most, if not the most, prestigious, authoritative and highly impacting journals in the field of physics. Its most recent impact factor for 2016 is 36.917 with a 5-year impact factor of 45.547.

TESOL Quarterly

TESOL Quarterly is a quarterly peer-reviewed academic journal published by Wiley-Blackwell on behalf of TESOL International Association. It covers English language teaching and learning and standard English as a second dialect, including articles on the psychology and sociology of language learning and teaching, professional preparation, curriculum development, and testing and evaluation. The editors-in-chief are Charlene Polio and Peter De Costa, both at Michigan State University. TESOL also publishes TESOL Journal.

According to the Journal Citation Reports, the journal had a 2016 impact factor of 2.056, ranking it 14th out of 182 journals in the category "Linguistics" and 34th out of 235 journals in the category "Education & Educational Research". There has been a substantial increase in the past three years under the editorial leadership of previous editors, Brian Paltridge and Ahmar Mahboob, both of the University of Sydney: the 2015 impact factor was 1.513, and 2014 impact factor was 0.940.

Tetrahedron (journal)

Tetrahedron is a weekly peer-reviewed scientific journal covering the field of organic chemistry. According to the Journal Citation Reports, the journal has a 2014 impact factor of 2.641. Tetrahedron and Elsevier, its publisher, support an annual symposium. In 2010, complaints were raised over its high subscription cost.

Tetrahedron Letters

Tetrahedron Letters is a weekly international journal for rapid publication of full original research papers in the field of organic chemistry. According to the Journal Citation Reports, the journal has a 2014 impact factor of 2.379 and it is ranked 22nd out of 57 journals in the "Organic Chemistry" category.

Webometrics

The science of webometrics (also cybermetrics) tries to measure the World Wide Web to get knowledge about the number and types of hyperlinks, structure of the World Wide Web and usage patterns. According to Björneborn and Ingwersen (2004), the definition of webometrics is "the study of the quantitative aspects of the construction and use of information resources, structures and technologies on the Web drawing on bibliometric and informetric approaches." The term webometrics was first coined by Almind and Ingwersen (1997). A second definition of webometrics has also been introduced, "the study of web-based content with primarily quantitative methods for social science research goals using techniques that are not specific to one field of study" (Thelwall, 2009), which emphasizes the development of applied methods for use in the wider social sciences. The purpose of this alternative definition was to help publicize appropriate methods outside of the information science discipline rather than to replace the original definition within information science.

Similar scientific fields are Bibliometrics, Informetrics, Scientometrics, Virtual ethnography, and Web mining.

One relatively straightforward measure is the "Web Impact Factor" (WIF) introduced by Ingwersen (1998). The WIF measure may be defined as the number of web pages in a web site receiving links from other web sites, divided by the number of web pages published in the site that are accessible to the crawler. However the use of WIF has been disregarded due to the mathematical artifacts derived from power law distributions of these variables. Other similar indicators using size of the institution instead of number of webpages have been proved more useful.

Journals
Papers
Other types of publication
Impact and ranking
Reform
Indexes and search engines
Related topics
Lists

This page is based on a Wikipedia article written by authors (here).
Text is available under the CC BY-SA 3.0 license; additional terms may apply.
Images, videos and audio are available under their respective licenses.