Peer review

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

A reviewer at the American National Institutes of Health evaluates a grant proposal.

Peer review is the evaluation of work by one or more people with similar competences as the producers of the work (peers). It functions as a form of self-regulation by qualified members of a profession within the relevant field. Peer review methods are used to maintain quality standards, improve performance, and provide credibility. In academia, scholarly peer review is often used to determine an academic paper's suitability for publication. Peer review can be categorized by the type of activity and by the field or profession in which the activity occurs, e.g., medical peer review.

Professional[edit]

Professional peer review focuses on the performance of professionals, with a view to improving quality, upholding standards, or providing certification. In academia, peer review is used to inform in decisions related to faculty advancement and tenure.[1] Henry Oldenburg (1619–1677) was a German-born British philosopher who is seen as the 'father' of modern scientific peer review.[2][3][4]

WA prototype is a professional peer-review process originally recommended in the Ethics of the Physician written by Ishāq ibn ʻAlī al-Ruhāwī (854–931). He stated that a visiting physician had to make duplicate notes of a patient's condition on every visit. When the patient was cured or had died, the notes of the physician were examined by a local medical council of other physicians, who would decide whether the treatment had met the required standards of medical care.[5]

Professional peer review is common in the field of health care, where it is usually called clinical peer review.[6] Further, since peer review activity is commonly segmented by clinical discipline, there is also physician peer review, nursing peer review, dentistry peer review, etc.[7] Many other professional fields have some level of peer review process: accounting,[8] law,[9][10] engineering (e.g., software peer review, technical peer review), aviation, and even forest fire management.[11]

Peer review is used in education to achieve certain learning objectives, particularly as a tool to reach higher order processes in the affective and cognitive domains as defined by Bloom's taxonomy. This may take a variety of forms, including closely mimicking the scholarly peer review processes used in science and medicine.[12][13]

Scholarly[edit]

Scholarly peer review (also known as refereeing) is the process of subjecting an author's scholarly work, research, or ideas to the scrutiny of others who are experts in the same field, before a paper describing this work is published in a journal, conference proceedings or as a book. The peer review helps the publisher (that is, the editor-in-chief, the editorial board or the program committee) decide whether the work should be accepted, considered acceptable with revisions, or rejected.

Peer review requires a community of experts in a given (and often narrowly defined) field, who are qualified and able to perform reasonably impartial review. Impartial review, especially of work in less narrowly defined or inter-disciplinary fields, may be difficult to accomplish, and the significance (good or bad) of an idea may never be widely appreciated among its contemporaries. Peer review is generally considered necessary to academic quality and is used in most major scholarly journals, but it by no means prevents publication of invalid research.[14] Meta-research has identified weaknesses in common peer review practices, leading critics to argue for reform. Researchers within the fields of metascience and journalology work to produce such reform.

A number of alternative peer review systems have been proposed to address known problems in the peer review process.[15][16] Traditionally, peer reviewers have been anonymous, but there are several examples of open peer review, where the comments are visible to readers, generally with the identities of the peer reviewers disclosed as well, e.g., F1000, eLife, BMJ, and BioMed Central.

Government policy[edit]

The European Union has been using peer review in the "Open Method of Co-ordination" of policies in the fields of active labour market policy since 1999.[17] In 2004, a program of peer reviews started in social inclusion.[18] Each program sponsors about eight peer review meetings in each year, in which a "host country" lays a given policy or initiative open to examination by half a dozen other countries and the relevant European-level NGOs. These usually meet over two days and include visits to local sites where the policy can be seen in operation. The meeting is preceded by the compilation of an expert report on which participating "peer countries" submit comments. The results are published on the web.

The United Nations Economic Commission for Europe, through UNECE Environmental Performance Reviews, uses peer review, referred to as "peer learning", to evaluate progress made by its member countries in improving their environmental policies.

The State of California is the only U.S. state to mandate scientific peer review. In 1997, the Governor of California signed into law Senate Bill 1320 (Sher), Chapter 295, statutes of 1997, which mandates that, before any CalEPA Board, Department, or Office adopts a final version of a rule-making, the scientific findings, conclusions, and assumptions on which the proposed rule are based must be submitted for independent external scientific peer review. This requirement is incorporated into the California Health and Safety Code Section 57004.[19]

Medical[edit]

Medical peer review may be distinguished in 4 classifications: 1) clinical peer review; 2) peer evaluation of clinical teaching skills for both physicians and nurses;[20][21] 3) scientific peer review of journal articles; 4) a secondary round of peer review for the clinical value of articles concurrently published in medical journals.[22] Additionally, "medical peer review" has been used by the American Medical Association to refer not only to the process of improving quality and safety in health care organizations, but also to the process of rating clinical behavior or compliance with professional society membership standards.[23][24] Thus, the terminology has poor standardization and specificity, particularly as a database search term.[citation needed]

Technical[edit]

In engineering, technical peer review is a type of engineering review. Technical peer reviews are a well defined review process for finding and fixing defects, conducted by a team of peers with assigned roles. Technical peer reviews are carried out by peers representing areas of life cycle affected by material being reviewed (usually limited to 6 or fewer people). Technical peer reviews are held within development phases, between milestone reviews, on completed products or completed portions of products.[25]

Criticism[edit]

To an outsider, the anonymous, pre-publication peer review process is opaque. Certain journals are accused of not carrying out stringent peer review in order to more easily expand their customer base, particularly in journals where authors pay a fee before publication.[26] Richard Smith, MD, former editor of the British Medical Journal, has claimed that peer review is "ineffective, largely a lottery, anti-innovatory, slow, expensive, wasteful of scientific time, inefficient, easily abused, prone to bias, unable to detect fraud and irrelevant; Several studies have shown that peer review is biased against the provincial and those from low- and middle-income countries; Many journals take months and even years to publish and the process wastes researchers' time. As for the cost, the Research Information Network estimated the global cost of peer review at £1.9 billion in 2008."[27]

In addition, Australia's Innovative Research Universities group (a coalition of seven comprehensive universities committed to inclusive excellence in teaching, learning and research in Australia) has found that "peer review disadvantages researchers in their early careers, when they rely on competitive grants to cover their salaries, and when unsuccessful funding applications often mark the end of a research idea".[28]

Low-end distinctions in articles understandable to all peers[edit]

John Ioannidis argues that since the exams and other tests that people pass on their way from "layman" to "expert" focus on answering the questions in time and in accordance with a list of answers, and not on making precise distinctions (the latter of which would be unrecognizable to experts of lower cognitive precision anyway), there is as much individual variation in the ability to distinguish causation from correlation among "experts" as there is among "laymen". Ioannidis argues that as a result, scholarly peer review by many "experts" allows only articles that are understandable at a wide range of cognitive precision levels including very low ones to pass, biasing publications towards favoring articles that infer causation from correlation while mislabelling articles that make the distinction as "incompetent overestimation of one's ability" on the side of the authors because some of the reviewing "experts" are cognitively unable to distinguish the distinction from alleged rationalization of specific conclusions. It is argued by Ioannidis that this makes peer review a cause of selective publication of false research findings while stopping publication of rigorous criticism thereof, and that further postpublication review repeats the same bias by selectively retracting the few rigorous articles that may have made it through initial prepublication peer review while letting the low-end ones that confuse correlation and causation remain in print.[29]

Peer review and trust[edit]

Researchers have peer reviewed manuscripts prior to publishing them in a variety of ways since the 18th century.[30][31] The main goal of this practice is to improve the relevance and accuracy of scientific discussions. Even though experts often criticize peer review for a number of reasons, the process is still often considered the "gold standard" of science.[32] Occasionally however, peer review approves studies that are later found to be wrong and rarely deceptive or fraudulent results are discovered prior to publication.[33][34] Thus, there seems to be an element of discord between the ideology behind and the practice of peer review. By failing to effectively communicate that peer review is imperfect, the message conveyed to the wider public is that studies published in peer-reviewed journals are "true" and that peer review protects the literature from flawed science. A number of well-established criticisms exist of many elements of peer review.[35][36][37] In the following we describe cases of the wider impact inappropriate peer review can have on public understanding of scientific literature.

Multiple examples across several areas of science find that scientists elevated the importance of peer review for research that was questionable or corrupted. For example, climate change deniers have published studies in the Energy and Environment journal, attempting to undermine the body of research that shows how human activity impacts the Earth's climate. Politicians in the United States who reject the established science of climate change have then cited this journal on several occasions in speeches and reports.[note 1]

At times, peer review has been exposed as a process that was orchestrated for a preconceived outcome. The New York Times gained access to confidential peer review documents for studies sponsored by the National Football Leagues (NFL) that were cited as scientific evidence that brain injuries do not cause long-term harm to its players.[note 2] During the peer review process, the authors of the study stated that all NFL players were part of a study, a claim that the reporters found to be false by examining the database used for the research. Furthermore, The Times noted that the NFL sought to legitimize the studies" methods and conclusion by citing a "rigorous, confidential peer-review process" despite evidence that some peer reviewers seemed "desperate" to stop their publication. Recent research has also demonstrated that widespread industry funding for published medical research often goes undeclared and that such conflicts of interest are not appropriately addressed by peer review.[38][39]

Another problem that peer review fails to catch is ghostwriting, a process by which companies draft articles for academics who then publish them in journals, sometimes with little or no changes.[40] These studies can then be used for political, regulatory and marketing purposes. In 2010, the US Senate Finance Committee released a report that found this practice was widespread, that it corrupted the scientific literature and increased prescription rates.[note 3] Ghostwritten articles have appeared in dozens of journals, involving professors at several universities.[note 4]

Just as experts in a particular field have a better understanding of the value of papers published in their area, scientists are considered to have better grasp of the value of published papers than the general public and to see peer review as a human process, with human failings,[41] and that "despite its limitations, we need it. It is all we have, and it is hard to imagine how we would get along without it".[42] But these subtleties are lost on the general public, who are often misled into thinking that published in a journal with peer review is the "gold standard" and can erroneously equate published research with the truth.[41] Thus, more care must be taken over how peer review, and the results of peer reviewed research, are communicated to non-specialist audiences; particularly during a time in which a range of technical changes and a deeper appreciation of the complexities of peer review are emerging.[43][44][45][46] This will be needed as the scholarly publishing system has to confront wider issues such as retractions[34][47][48] and replication or reproducibility "crisis'.[49][50][51]

Views of peer review[edit]

Peer review is often considered integral to scientific discourse in one form or another. Its gatekeeping role is supposed to be necessary to maintain the quality of the scientific literature[52][53] and avoid a risk of unreliable results, inability to separate signal from noise, and slow scientific progress.[54][55]

Shortcomings of peer review have been met with calls for even stronger filtering and more gatekeeping. A common argument in favor of such initiatives is the belief that this filter is needed to maintain the integrity of the scientific literature.[56][57]

Calls for more oversight have at least two implications that are counterintuitive of what is known to be true scholarship.[41]

  1. The belief that scholars are incapable of evaluating the quality of work on their own, that they are in need of a gatekeeper to inform them of what is good and what is not.
  2. The belief that scholars need a "guardian" to make sure they are doing good work.

Others argue[41] that authors most of all have a vested interest in the quality of a particular piece of work. Only the authors could have, as Feynman (1974)[note 5] puts it, the "extra type of integrity that is beyond not lying, but bending over backwards to show how you're maybe wrong, that you ought to have when acting as a scientist." If anything, the current peer review process and academic system could penalize, or at least fail to incentivize, such integrity.

Instead, the credibility conferred by the "peer-reviewed" label could diminish what Feynman calls the culture of doubt necessary for science to operate a self-correcting, truth-seeking process.[58] The effects of this can be seen in the ongoing replication crisis, hoaxes, and widespread outrage over the inefficacy of the current system.[35][30] It's common to think that more oversight is the answer, as peer reviewers are not at all lacking in skepticism. But the issue is not the skepticism shared by the select few who determine whether an article passes through the filter. It is the validation, and accompanying lack of skepticism, that comes afterwards.[note 6] Here again more oversight only adds to the impression that peer review ensures quality, thereby further diminishing the culture of doubt and counteracting the spirit of scientific inquiry.[note 7]

Quality research - even some of our most fundamental scientific discoveries - dates back centuries, long before peer review took its current form.[30][59][31] Whatever peer review existed centuries ago, it took a different form than it does in modern times, without the influence of large, commercial publishing companies or a pervasive culture of publish or perish.[59] Though in its initial conception it was often a laborious and time-consuming task, researchers took peer review on nonetheless, not out of obligation but out of duty to uphold the integrity of their own scholarship. They managed to do so, for the most part, without the aid of centralised journals, editors, or any formalised or institutionalised process whatsoever. Supporters of modern technology argue[41] that it makes it possible to communicate instantaneously with scholars around the globe, make such scholarly exchanges easier, and restore peer review to a purer scholarly form, as a discourse in which researchers engage with one another to better clarify, understand, and communicate their insights.[44][60]

Such modern technology includes posting results to preprint servers, preregistration of studies, open peer review, and other open science practices.[50][61][62] In all these initiatives, the role of gatekeeping remains prominent, as if a necessary feature of all scholarly communication, but critics argue[37] that a proper, real-world implementation could test and disprove this assumption; demonstrate researchers' desire for more that traditional journals can offer; show that researchers can be entrusted to perform their own quality control independent of journal-coupled review. Jon Tennant also argues that the outcry over the inefficiencies of traditional journals centers on their inability to provide rigorous enough scrutiny, and the outsourcing of critical thinking to a concealed and poorly-understood process. Thus, the assumption that journals and peer review are required to protect scientific integrity seems to undermine the very foundations of scholarly inquiry.[41]

To test the hypothesis that filtering is indeed unnecessary to quality control, many of the traditional publication practices would need to be redesigned, editorial boards repurposed if not disbanded, and authors granted control over the peer review of their own work. Putting authors in charge of their own peer review is seen as serving a dual purpose.[41] On one hand, it removes the conferral of quality within the traditional system, thus eliminating the prestige associated with the simple act of publishing. Perhaps paradoxically, the removal of this barrier might actually result in an increase of the quality of published work, as it eliminates the cachet of publishing for its own sake. On the other hand, readers know that there is no filter so they must interpret anything they read with a healthy dose of skepticism, thereby naturally restoring the culture of doubt to scientific practice.[63][64][65]

In addition to concerns about the quality of work produced by well-meaning researchers, there are concerns that a truly open system would allow the literature to be populated with junk and propaganda by those with a vested interest in certain issues. A counterargument is that the conventional model of peer review diminishes the healthy skepticism that is a hallmark of scientific inquiry, and thus confers credibility upon subversive attempts to infiltrate the literature.[41] Allowing such "junk" to be published could make individual articles less reliable but render the overall literature more robust by fostering a "culture of doubt".[63]

One initiative experimenting in this area is Researchers.One, a non-profit peer review publication platform featuring a novel author-driven peer review process.[66] Other similar examples include the Self-Journal of Science, PRElights, and The Winnower, which do not yet seem to have greatly disrupted the traditional peer review workflow. Supporters conclude that researchers are more than responsible and competent enough to ensure their own quality control; they just need the means and the authority to do so.[41]

See also[edit]

Notes[edit]

  1. ^ "Skeptics get a journal" (PDF)., Paul Thacker, 2005.
  2. ^ "N.F.L.'s Flawed Concussion Research and Ties to Tobacco Industry"..
  3. ^ "Ghostwriting in medical literature" (PDF)..
  4. ^ "Frequently asked questions about medical ghostwriting"..
  5. ^ "Cargo cult science"., Richard Feynman.
  6. ^ "Peer Review: The Worst Way to Judge Research, Except for All the Others"., Aaron E. Carroll, New York Times.
  7. ^ "Bucking the Big Bang"., Eric Lerner, New Scientist.

References[edit]

  1. ^ Schimanski, Lesley A.; Alperin, Juan Pablo (2018). "The evaluation of scholarship in academic promotion and tenure processes: Past, present, and future". F1000Research. 7: 1605. doi:10.12688/f1000research.16493.1. ISSN 2046-1402. PMC 6325612. PMID 30647909.
  2. ^ Hatch, Robert A. (February 1998). "The Scientific Revolution: Correspondence Networks". University of Florida. Retrieved August 21, 2016.
  3. ^ Oldenburg, Henry (1665). "Epistle Dedicatory". Philosophical Transactions of the Royal Society. 1: 0. doi:10.1098/rstl.1665.0001.
  4. ^ Hall, Marie Boas (2002). Henry Oldenburg: shaping the Royal Society. Oxford: Oxford University Press. Bibcode:2002heol.book.....B. ISBN 978-0-19-851053-6.
  5. ^ Spier, Ray (2002). "The history of the peer-review process". Trends in Biotechnology. 20 (8): 357–8. doi:10.1016/S0167-7799(02)01985-6. PMID 12127284.
  6. ^ Dans, PE (1993). "Clinical peer review: burnishing a tarnished image". Annals of Internal Medicine. 118 (7): 566–8. doi:10.7326/0003-4819-118-7-199304010-00014. PMID 8442628.
  7. ^ Milgrom P, Weinstein P, Ratener P, Read WA, Morrison K; Weinstein; Ratener; Read; Morrison (1978). "Dental Examinations for Quality Control: Peer Review versus Self-Assessment". American Journal of Public Health. 68 (4): 394–401. doi:10.2105/AJPH.68.4.394. PMC 1653950. PMID 645987.CS1 maint: multiple names: authors list (link)
  8. ^ "AICPA Peer Review Program Manual". American Institute of CPAs.
  9. ^ "Peer Review". UK Legal Services Commission. July 12, 2007. Archived from the original on October 14, 2010.
  10. ^ "Martindale-Hubbell Attorney Reviews and Ratings". Martindale. Retrieved January 27, 2020.
  11. ^ "Peer Review Panels – Purpose and Process" (PDF). USDA Forest Service. February 6, 2006. Retrieved October 4, 2010.
  12. ^ Sims Gerald K. (1989). "Student Peer Review in the Classroom: A Teaching and Grading Tool" (PDF). Journal of Agronomic Education. 18: 105–108. The review process was double-blind to provide anonymity for both authors and reviewers, but was otherwise handled in a fashion similar to that used by scientific journals
  13. ^ Liu, Jianguo; Pysarchik, Dawn Thorndike; Taylor, William W. (2002). "Peer Review in the Classroom" (PDF). BioScience. 52 (9): 824–829. doi:10.1641/0006-3568(2002)052[0824:PRITC]2.0.CO;2.
  14. ^ KupferschmidtAug. 17, Kai; 2018; Am, 9:15 (August 14, 2018). "Researcher at the center of an epic fraud remains an enigma to those who exposed him". Science | AAAS. Retrieved August 11, 2019.
  15. ^ Cosgrove, Andrew; Cheifet, Barbara (November 27, 2018). "Transparent peer review trial: the results". Genome Biology. 19 (1): 206. doi:10.1186/s13059-018-1584-0. ISSN 1474-760X. PMC 6260718. PMID 30482224.
  16. ^ Patterson, Mark; Schekman, Randy (June 26, 2018). "A new twist on peer review". eLife. 7: e36545. doi:10.7554/eLife.36545. ISSN 2050-084X. PMC 6019064. PMID 29944117.
  17. ^ "Mutual Learning Programme - Employment, Social Affairs & Inclusion - European Commission". ec.europa.eu.
  18. ^ "Social Peer to Peer – Online Casino Reviews". www.peer-review-social-inclusion.eu.
  19. ^ "What is Scientific Peer Review?". ceparev.berkeley.edu. Retrieved March 30, 2017.
  20. ^ Medschool.ucsf.edu Archived August 14, 2010, at the Wayback Machine
  21. ^ Ludwick R, Dieckman BC, Herdtner S, Dugan M, Roche M (November–December 1998). "Documenting the scholarship of clinical teaching through peer review". Nurse Educator. 23 (6): 17–20. doi:10.1097/00006223-199811000-00008. PMID 9934106.
  22. ^ Haynes RB, Cotoi C, Holland J, et al. (2006). "Second-order peer review of the medical literature for clinical practitioners". JAMA. 295 (15): 1801–8. doi:10.1001/jama.295.15.1801. PMID 16622142.
  23. ^ Snelson, Elizabeth A. (2010). Physician's Guide to Medical Staff Organization Bylaws (PDF). ama-assn.org. p. 131. Archived from the original (PDF) on August 6, 2011.
  24. ^ "Medical Peer Review". Ama-assn.org. Archived from the original on March 6, 2010.
  25. ^ NASA Systems Engineering Handbook (PDF). NASA. 2007. SP-610S.
  26. ^ Couchman, John R. (November 11, 2013). "Peer Review and Reproducibility. Crisis or Time for Course Correction?". Journal of Histochemistry & Cytochemistry. 62 (1): 9–10. doi:10.1369/0022155413513462. PMC 3873808. PMID 24217925.
  27. ^ "The peer review drugs don't work". Times Higher Education (THE). May 28, 2015. Retrieved October 23, 2018.
  28. ^ "Peer review 'works against' early career researchers". Times Higher Education (THE). July 16, 2018. Retrieved October 23, 2018.
  29. ^ JPA Ioannidis (2005) "Why Most Published Research Findings Are False"
  30. ^ a b c Csiszar, Alex (2016). "Peer Review: Troubled from the Start". Nature. 532 (7599): 306–308. Bibcode:2016Natur.532..306C. doi:10.1038/532306a. PMID 27111616.
  31. ^ a b Moxham, Noah; Fyfe, Aileen (2018). "The Royal Society and the Prehistory of Peer Review, 1665–1965" (PDF). The Historical Journal. 61 (4): 863–889. doi:10.1017/S0018246X17000334.
  32. ^ Moore, John (2006). "Does Peer Review Mean the Same to the Public as It Does to Scientists?". Nature. doi:10.1038/nature05009.
  33. ^ Ferguson, Cat; Marcus, Adam; Oransky, Ivan (2014). "Publishing: The Peer-Review Scam". Nature. 515 (7528): 480–482. Bibcode:2014Natur.515..480F. doi:10.1038/515480a. PMID 25428481.
  34. ^ a b Budd, J. M.; Sievert, M.; Schultz, T. R. (1998). "Phenomena of Retraction: Reasons for Retraction and Citations to the Publications". JAMA. 280 (3): 296–7. doi:10.1001/jama.280.3.296. PMID 9676689.
  35. ^ a b Smith, Richard (2006). "Peer Review: A Flawed Process at the Heart of Science and Journals". Journal of the Royal Society of Medicine. 99 (4): 178–82. doi:10.1177/014107680609900414. PMC 1420798. PMID 16574968.
  36. ^ Ross-Hellauer, Tony (2017). "What Is Open Peer Review? A Systematic Review". F1000Research. 6: 588. doi:10.12688/f1000research.11369.2. PMC 5437951. PMID 28580134.
  37. ^ a b Tennant, Jonathan P.; Dugan, Jonathan M.; Graziotin, Daniel; Jacques, Damien C.; Waldner, François; Mietchen, Daniel; Elkhatib, Yehia; b. Collister, Lauren; Pikas, Christina K.; Crick, Tom; Masuzzo, Paola; Caravaggi, Anthony; Berg, Devin R.; Niemeyer, Kyle E.; Ross-Hellauer, Tony; Mannheimer, Sara; Rigling, Lillian; Katz, Daniel S.; Greshake Tzovaras, Bastian; Pacheco-Mendoza, Josmel; Fatima, Nazeefa; Poblet, Marta; Isaakidis, Marios; Irawan, Dasapta Erwin; Renaut, Sébastien; Madan, Christopher R.; Matthias, Lisa; Nørgaard Kjær, Jesper; O'Donnell, Daniel Paul; et al. (2017). "A Multi-Disciplinary Perspective on Emergent and Future Innovations in Peer Review". F1000Research. 6: 1151. doi:10.12688/f1000research.12037.3. PMC 5686505. PMID 29188015.
  38. ^ Wong, Victoria S. S.; Avalos, Lauro Nathaniel; Callaham, Michael L. (2019). "Industry Payments to Physician Journal Editors". PLOS One. 14 (2): e0211495. Bibcode:2019PLoSO..1411495W. doi:10.1371/journal.pone.0211495. PMC 6366761. PMID 30730904.
  39. ^ Weiss, Glen J.; Davis, Roger B. (2019). "Discordant Financial Conflicts of Interest Disclosures between Clinical Trial Conference Abstract and Subsequent Publication". PeerJ. 7: e6423. doi:10.7717/peerj.6423. PMC 6375255. PMID 30775185.
  40. ^ Flaherty, D. K. (2013). "Ghost- and Guest-Authored Pharmaceutical Industry–Sponsored Studies: Abuse of Academic Integrity, the Peer Review System, and Public Trust". The Annals of Pharmacotherapy. 47 (7–8): 1081–3. doi:10.1345/aph.1R691. PMID 23585648.
  41. ^ a b c d e f g h i Vanholsbeeck, Marc; Thacker, Paul; Sattler, Susanne; Ross-Hellauer, Tony; Rivera-López, Bárbara S.; Rice, Curt; Nobes, Andy; Masuzzo, Paola; Martin, Ryan; Kramer, Bianca; Havemann, Johanna; Enkhbayar, Asura; Davila, Jacinto; Crick, Tom; Crane, Harry; Tennant, Jonathan P. (March 11, 2019). "Ten Hot Topics around Scholarly Publishing". Publications. 7 (2): 34. doi:10.3390/publications7020034.
  42. ^ Relman, A. S. (1990). "Peer Review in Scientific Journals--What Good Is It?". Western Journal of Medicine. 153 (5): 520–22. PMC 1002603. PMID 2260288.
  43. ^ Bravo, Giangiacomo; Grimaldo, Francisco; López-Iñesta, Emilia; Mehmani, Bahar; Squazzoni, Flaminio (2019). "The Effect of Publishing Peer Review Reports on Referee Behavior in Five Scholarly Journals". Nature Communications. 10 (1): 322. Bibcode:2019NatCo..10..322B. doi:10.1038/s41467-018-08250-2. PMC 6338763. PMID 30659186.
  44. ^ a b Tennant, Jonathan P. (2018). "The State of the Art in Peer Review". FEMS Microbiology Letters. 365 (19). doi:10.1093/femsle/fny204. PMC 6140953. PMID 30137294.
  45. ^ Squazzoni, Flaminio; Grimaldo, Francisco; Marušić, Ana (2017). "Publishing: Journals Could Share Peer-Review Data". Nature. 546 (7658): 352. Bibcode:2017Natur.546Q.352S. doi:10.1038/546352a. PMID 28617464.
  46. ^ Allen, Heidi; Boxer, Emma; Cury, Alexandra; Gaston, Thomas; Graf, Chris; Hogan, Ben; Loh, Stephanie; Wakley, Hannah; Willis, Michael (2018). "What Does Better Peer Review Look like? Definitions, Essential Areas, and Recommendations for Better Practice". doi:10.17605/OSF.IO/4MFK2. Cite journal requires |journal= (help)
  47. ^ Fang, Ferric C.; Casadevall, Arturo (2011). "Retracted Science and the Retraction Index". Infection and Immunity. 79 (10): 3855–3859. doi:10.1128/IAI.05661-11. PMC 3187237. PMID 21825063.
  48. ^ Moylan, Elizabeth C.; Kowalczuk, Maria K. (2016). "Why Articles Are Retracted: A Retrospective Cross-Sectional Study of Retraction Notices at BioMed Central". BMJ Open. 6 (11): e012047. doi:10.1136/bmjopen-2016-012047. PMC 5168538. PMID 27881524.
  49. ^ Open Science Collaboration (2015). "Estimating the Reproducibility of Psychological Science". Science. 349 (6251): aac4716. doi:10.1126/science.aac4716. hdl:10722/230596. PMID 26315443.
  50. ^ a b Munafò, Marcus R.; Nosek, Brian A.; Bishop, Dorothy V. M.; Button, Katherine S.; Chambers, Christopher D.; Percie Du Sert, Nathalie; Simonsohn, Uri; Wagenmakers, Eric-Jan; Ware, Jennifer J.; Ioannidis, John P. A. (2017). "A Manifesto for Reproducible Science". Nature Human Behaviour. 1. doi:10.1038/s41562-016-0021.
  51. ^ Fanelli, Daniele (2018). "Opinion: Is Science Really Facing a Reproducibility Crisis, and Do We Need It To?". Proceedings of the National Academy of Sciences. 115 (11): 2628–2631. doi:10.1073/pnas.1708272114. PMC 5856498. PMID 29531051.
  52. ^ Goodman, Steven N. (1994). "Manuscript Quality before and after Peer Review and Editing at Annals of Internal Medicine". Annals of Internal Medicine. 121 (1): 11–21. doi:10.7326/0003-4819-121-1-199407010-00003. PMID 8198342.
  53. ^ Pierson, Charon A. (2018). "Peer review and journal quality". Journal of the American Association of Nurse Practitioners. 30 (1): 1–2. doi:10.1097/JXX.0000000000000018. PMID 29757914.
  54. ^ Caputo, Richard K. (2019). "Peer Review: A Vital Gatekeeping Function and Obligation of Professional Scholarly Practice". Families in Society: The Journal of Contemporary Social Services. 100: 6–16. doi:10.1177/1044389418808155.
  55. ^ Siler, Kyle; Lee, Kirby; Bero, Lisa (2015). "Measuring the effectiveness of scientific gatekeeping". Proceedings of the National Academy of Sciences. 112 (2): 360–365. Bibcode:2015PNAS..112..360S. doi:10.1073/pnas.1418218112. PMC 4299220. PMID 25535380.
  56. ^ Resnik, David B.; Elmore, Susan A. (2016). "Ensuring the Quality, Fairness, and Integrity of Journal Peer Review: A Possible Role of Editors". Science and Engineering Ethics. 22 (1): 169–188. doi:10.1007/s11948-015-9625-5. PMID 25633924.
  57. ^ Bornmann, Lutz (2011). "Scientific Peer Review". Annual Review of Information Science and Technology. 45: 197–245. doi:10.1002/aris.2011.1440450112.
  58. ^ "Cargo Cult Science". Caltech Magazine. 1974. Archived from the original on August 24, 2019.
  59. ^ a b "Untangling Academic Publishing. A History of the Relationship between Commercial Interests, Academic Prestige and the Circulation of Research". 26.
  60. ^ Priem, Jason; Hemminger, Bradley M. (2012). "Decoupling the Scholarly Journal". Frontiers in Computational Neuroscience. 6: 19. doi:10.3389/fncom.2012.00019. PMC 3319915. PMID 22493574.
  61. ^ Bowman, Nicholas David; Keene, Justin Robert (2018). "A Layered Framework for Considering Open Science Practices". Communication Research Reports. 35 (4): 363–372. doi:10.1080/08824096.2018.1513273.
  62. ^ McKiernan, E. C.; Bourne, P. E.; Brown, C. T.; Buck, S.; Kenall, A.; Lin, J.; McDougall, D.; Nosek, B. A.; Ram, K.; Soderberg, C. K.; Spies, J. R.; Thaney, K.; Updegrove, A.; Woo, K. H.; Yarkoni, T. (2016). "Point of View: How Open Science Helps Researchers Succeed". eLife. 5. doi:10.7554/eLife.16800. PMC 4973366. PMID 27387362.
  63. ^ a b "In Peer Review We (Don't) Trust: How Peer Review's Filtering Poses a Systemic Risk to Science".
  64. ^ Brembs, Björn (2019). "Reliable Novelty: New Should Not Trump True". PLOS Biology. 17 (2): e3000117. doi:10.1371/journal.pbio.3000117. PMC 6372144. PMID 30753184.
  65. ^ Stern, Bodo M.; o'Shea, Erin K. (2019). "A Proposal for the Future of Scientific Publishing in the Life Sciences". PLOS Biology. 17 (2): e3000116. doi:10.1371/journal.pbio.3000116. PMC 6372143. PMID 30753179.
  66. ^ "The RESEARCHERS.ONE Mission".

Further reading[edit]

External links[edit]