Peer review

From HandWiki
Short description: Evaluation of work by one or more people of similar competence to the producers of the work


A reviewer at the American National Institutes of Health evaluates a grant proposal.

Peer review is the evaluation of work by one or more people with similar competencies as the producers of the work (peers). It functions as a form of self-regulation by qualified members of a profession within the relevant field. Peer review methods are used to maintain quality standards, improve performance, and provide credibility. In academia, scholarly peer review is often used to determine an academic paper's suitability for publication. Peer review can be categorized by the type of activity and by the field or profession in which the activity occurs, e.g., medical peer review.

Professional

Professional peer review focuses on the performance of professionals, with a view to improving quality, upholding standards, or providing certification. In academia, peer review is used to inform in decisions related to faculty advancement and tenure.[1] Henry Oldenburg (1619–1677) was a German-born British philosopher who is seen as the 'father' of modern scientific peer review.[2][3][4]

A prototype professional peer-review process was recommended in the Ethics of the Physician written by Ishāq ibn ʻAlī al-Ruhāwī (854–931). He stated that a visiting physician had to make duplicate notes of a patient's condition on every visit. When the patient was cured or had died, the notes of the physician were examined by a local medical council of other physicians, who would decide whether the treatment had met the required standards of medical care.[5]

Professional peer review is common in the field of health care, where it is usually called clinical peer review.[6] Further, since peer review activity is commonly segmented by clinical discipline, there is also physician peer review, nursing peer review, dentistry peer review, etc.[7] Many other professional fields have some level of peer review process: accounting,[8] law,[9][10] engineering (e.g., software peer review, technical peer review), aviation, and even forest fire management.[11]

Peer review is used in education to achieve certain learning objectives, particularly as a tool to reach higher order processes in the affective and cognitive domains as defined by Bloom's taxonomy. This may take a variety of forms, including closely mimicking the scholarly peer review processes used in science and medicine.[12][13]

Scholarly

Main page: Scholarly peer review
Short description: Subjecting an author's work or research to scrutiny


Scholarly peer review or academic peer review (also known as refereeing) is the process of having a draft version of a researcher's methods and findings reviewed (usually anonymously) by experts (or "peers") in the same field. Peer review is widely used for helping the academic publisher (that is, the editor-in-chief, the editorial board or the program committee) decide whether the work should be accepted, considered acceptable with revisions, or rejected for official publication in an academic journal, a monograph or in the proceedings of an academic conference. If the identities of authors are not revealed to each other, the procedure is called dual-anonymous peer review.

Academic peer review requires a community of experts in a given (and often narrowly defined) academic field, who are qualified and able to perform reasonably impartial review. Impartial review, especially of work in less narrowly defined or inter-disciplinary fields, may be difficult to accomplish, and the significance (good or bad) of an idea may never be widely appreciated among its contemporaries. Peer review is generally considered necessary to academic quality and is used in most major scholarly journals. However, peer review does not prevent publication of invalid research,[14] and as experimentally controlled studies of this process are difficult to arrange, direct evidence that peer review improves the quality of published papers is scarce.[15]

Scholarly peer review has been subject to several criticisms, and various proposals for reforming the system have been suggested over the years. Many studies have emphasized the problems inherent to the process of peer review.[16] Moreover, Ragone et al.,[17] have shown that there is a low correlation between peer review outcomes and the future impact measured by citations. Brezis and Birukou also show that the peer review process is not working properly. They underline that the ratings are not robust, e.g., changing reviewers can have a dramatic impact on the review results. Two main elements affect the bias in the peer process:[18]

  • The first element is that referees display homophily in their taste and perception of innovative ideas. So reviewers who are developing conventional ideas will tend to give low grades to innovative projects, while reviewers who have developed innovative ideas tend, by homophily, to give higher grades to innovative projects.
  • The second element leading to a high variance in the peer review process is that reviewers are not investing the same amount of time to analyze the projects (or equivalently are not with the same abilities). Brezis and Biruku[18] show that this heterogeneity among referees will lead to seriously affect the whole peer review process, and will lead to main arbitrariness in the results of the process.[18]

The peer process is also in use for projects acceptance. (For projects, the acceptance rates are small and are between 1% and 20%, with an average of 10%. In the European H2020 calls, the acceptance rate is 1.8%.) Peer review is more problematic when choosing the projects to be funded since innovative projects are not highly ranked in the existing peer-review process. The peer-review process leads to conformity, i.e., the selection of less controversial projects and papers. This may even influence the type of proposals scholars will propose, since scholars need to find financing for their research as discussed by Martin, 1997:[19] "A common informal view is that it is easier to obtain funds for conventional projects. Those who are eager to get funding are not likely to propose radical or unorthodox projects. Since you don't know who the referees are going to be, it is best to assume that they are middle-of-the-road. Therefore, the middle-of-the-road application is safer".[18]

Other attempts to reform the peer review process originate among others from the fields of metascience and journalology. Reformers seek to increase the reliability and efficiency of the peer review process and to provide it with a scientific foundation.[20][21][22] Alternatives to common peer review practices have been put to the test,[23][24] in particular open peer review, where the comments are visible to readers, generally with the identities of the peer reviewers disclosed as well, e.g., F1000, eLife, BMJ, and BioMed Central.[25] In the case of eLife, peer review is used not for deciding whether to publish an article, but for assessing its importance and reliability.[26] Likewise, the recognition and recruitment of peer reviewers continues to be a significant issue in the field of scholarly publishing.[27]

History

The first record of an editorial pre-publication peer-review is from 1665 by Henry Oldenburg, the founding editor of Philosophical Transactions of the Royal Society at the Royal Society of London.[28][29][30]

The first peer-reviewed publication might have been the Medical Essays and Observations published by the Royal Society of Edinburgh in 1731. The present-day peer-review system evolved from this 18th-century process,[31] began to involve external reviewers in the mid-19th-century,[32] and did not become commonplace until the mid-20th-century.[33]

Peer review became a touchstone of the scientific method, but until the end of the 19th century was often performed directly by an editor-in-chief or editorial committee.[34][35][2] Editors of scientific journals at that time made publication decisions without seeking outside input, i.e. an external panel of reviewers, giving established authors latitude in their journalistic discretion. For example, Albert Einstein's four revolutionary Annus Mirabilis papers in the 1905 issue of Annalen der Physik were peer-reviewed by the journal's editor-in-chief, Max Planck, and its co-editor, Wilhelm Wien, both future Nobel prize winners and together experts on the topics of these papers. On a much later occasion, Einstein was severely critical of the external review process, saying that he had not authorized the editor in chief to show his manuscript "to specialists before it is printed", and informing him that he would "publish the paper elsewhere" – which he did, and in fact he later had to withdraw the publication.[36]

While some medical journals started to systematically appoint external reviewers, it is only since the middle of the 20th century that this practice has spread widely and that external reviewers have been given some visibility within academic journals, including being thanked by authors and editors.[34][37] A 2003 editorial in Nature stated that, in the early 20th century, "the burden of proof was generally on the opponents rather than the proponents of new ideas."[38] Nature itself instituted formal peer review only in 1967.[39] Journals such as Science and the American Journal of Medicine increasingly relied on external reviewers in the 1950s and 1960s, in part to reduce the editorial workload.[37] In the 20th century, peer review also became common for science funding allocations. This process appears to have developed independently from that of editorial peer review.[28]:221

Gaudet[40][self-published source?] provides a social science view of the history of peer review carefully tending to what is under investigation, here peer review, and not only looking at superficial or self-evident commonalities among inquisition, censorship, and journal peer review. It builds on historical research by Gould,[41] Biagioli,[42] Spier,[2] and Rip.[43] The first Peer Review Congress met in 1989.[44] Over time, the fraction of papers devoted to peer review has steadily declined, suggesting that as a field of sociological study, it has been replaced by more systematic studies of bias and errors.[15] In parallel with "common experience" definitions based on the study of peer review as a "pre-constructed process", some social scientists have looked at peer review without considering it as pre-constructed. Hirschauer proposed that journal peer review can be understood as reciprocal accountability of judgements among peers.[45] Gaudet proposed that journal peer review could be understood as a social form of boundary judgement – determining what can be considered as scientific (or not) set against an overarching knowledge system, and following predecessor forms of inquisition and censorship.[40][self-published source?]

Pragmatically, peer review refers to the work done during the screening of submitted manuscripts. This process encourages authors to meet the accepted standards of their discipline and reduces the dissemination of irrelevant findings, unwarranted claims, unacceptable interpretations, and personal views. Publications that have not undergone peer review are likely to be regarded with suspicion by academic scholars and professionals.[citation needed] Non-peer-reviewed work does not contribute, or contributes less, to the academic credit of a scholar (such as the h-index), although this heavily depends on the field.

Justification

[P]eer review in its ideal form is both an act of altruism and an act of investment in the continuation of the scholarly enterprise by our colleagues. That is why it is so important to engage in thoughtful peer review as a scholar, and that is why it is important to do it well, acting not as a gatekeeper, but as a fellow contributor in the creation of knowledge. Collective wisdom.

Anna Marie Roos, editor-in-chief of Notes and Records: the Royal Society Journal of the History of Science, 2022[46]

It is difficult for authors and researchers, whether individually or in a team, to spot every mistake or flaw in a complicated piece of work. This is not necessarily a reflection on those concerned, but because with a new and perhaps eclectic subject, an opportunity for improvement may be more obvious to someone with special expertise or who simply looks at it with a fresh eye. Therefore, showing work to others increases the probability that weaknesses will be identified and improved. For both grant-funding and publication in a scholarly journal, it is also normally a requirement that the subject is both novel and substantial.[11][12]

The decision whether or not to publish a scholarly article, or what should be modified before publication, ultimately lies with the publisher (editor-in-chief or the editorial board) to which the manuscript has been submitted. Similarly, the decision whether or not to fund a proposed project rests with an official of the funding agency. These individuals usually refer to the opinion of one or more reviewers in making their decision. This is primarily for three reasons:[citation needed]

  • Workload. A small group of editors/assessors cannot devote sufficient time to each of the many articles submitted to many journals.
  • Miscellany of ideas. Were the editor/assessor to judge all submitted material themselves, approved material would solely reflect their opinion.
  • Limited expertise. An editor/assessor cannot be expected to be sufficiently expert in all areas covered by a single journal or funding agency to adequately judge all submitted material.

Reviewers are often anonymous and independent. However, some reviewers may choose to waive their anonymity, and in other limited circumstances, such as the examination of a formal complaint against the referee, or a court order, the reviewer's identity may have to be disclosed. Anonymity may be unilateral or reciprocal (single- or double-blinded reviewing).[citation needed]

Since reviewers are normally selected from experts in the fields discussed in the article, the process of peer review helps to keep some invalid or unsubstantiated claims out of the body of published research and knowledge. Scholars will read published articles outside their limited area of detailed expertise, and then rely, to some degree, on the peer-review process to have provided reliable and credible research that they can build upon for subsequent or related research. Significant scandal ensues when an author is found to have falsified the research included in an article, as other scholars, and the field of study itself, may have relied upon the invalid research.[citation needed]

For US universities, peer reviewing of books before publication is a requirement for full membership of the Association of American University Presses.[47]

Procedure

In the case of proposed publications, the publisher (editor-in-chief or the editorial board, often with assistance of corresponding or associate editors) sends advance copies of an author's work or ideas to researchers or scholars who are experts in the field (known as "referees" or "reviewers"). Communication is nowadays normally by e-mail or through a web-based manuscript processing system such as ScholarOne, Scholastica, or Open Journal Systems. Depending on the field of study and on the specific journal, there are usually one to three referees for a given article. For example, Springer states that there are two or three reviewers per article.[48]

The peer-review process involves three steps:[49]

Step 1: Desk evaluation

An editor evaluates the manuscript to judge whether the paper will be passed on to journal referees. At this phase many articles receive a "desk reject", that is, the editor chooses not to pass along the article. The authors may or may not receive a letter of explanation.

Desk rejection is intended to be a streamlined process so that editors may move past nonviable manuscripts quickly and provide authors with the opportunity to pursue a more suitable journal. For example, the European Accounting Review editors subject each manuscript to three questions to decide whether a manuscript moves forward to referees: 1) Is the article a fit for the journal's aims and scope, 2) is the paper content (e.g. literature review, methods, conclusions) sufficient and does the paper make a worthwhile contribution to the larger body of literature, and 3) does it follow format and technical specifications? If "no" to any of these, the manuscript receives a desk rejection.[50]

Desk rejection rates vary by journal. For example, in 2017 researchers at the World Bank compiled rejection rates of several global economics journals; the desk rejection rate ranged from 21% (Economic Lacea) to 66% (Journal of Development Economics).[51] The American Psychological Association publishes rejection rates for several major publications in the field, and although they do not specify whether the rejection is pre- or post- desk evaluation, their figures in 2016 ranged from a low of 49% to a high of 90%.[52]

Step 2: External review

If the paper is not desk rejected, the editors send the manuscript to the referees, who are chosen for their expertise and distance from the authors. At this point, referees may reject, accept without changes (rare)[53] or instruct the authors to revise and resubmit.

Reasons vary for acceptance of an article by editors, but Elsevier published an article where three editors weigh in on factors that drive article acceptance. These factors include whether the manuscript: delivers "new insight into an important issue", will be useful to practitioners, advances or proposes a new theory, raises new questions, has appropriate methods and conclusion, presents a good argument based on the literature, and tells a good story. One editor notes that he likes papers that he "wished he'd done" himself.[54]

These referees each return an evaluation of the work to the editor, noting weaknesses or problems along with suggestions for improvement. Typically, most of the referees' comments are eventually seen by the author, though a referee can also send 'for your eyes only' comments to the publisher; scientific journals observe this convention almost universally. The editor then evaluates the referees' comments, her or his own opinion of the manuscript before passing a decision back to the author(s), usually with the referees' comments.[55]

Referees' evaluations usually include an explicit recommendation of what to do with the manuscript or proposal, often chosen from options provided by the journal or funding agency. For example, Nature recommends four courses of action:[56]

  • to unconditionally accept the manuscript or the proposal,
  • to accept it in the event that its authors improve it in certain ways
  • to reject it, but encourage revision and invite re-submission
  • to reject it outright.

During this process, the role of the referees is advisory. The editor(s) is typically under no obligation to accept the opinions of the referees,[57] though he or she will most often do so. Furthermore, the referees in scientific publication do not act as a group, do not communicate with each other, and typically are not aware of each other's identities or evaluations. Proponents argue that if the reviewers of a paper are unknown to each other, the editor(s) can more easily verify the objectivity of the reviews. There is usually no requirement that the referees achieve consensus, with the decision instead often made by the editor(s) based on her best judgement of the arguments.

In situations where multiple referees disagree substantially about the quality of a work, there are a number of strategies for reaching a decision. The paper may be rejected outright, or the editor may choose which reviewer's point the authors should address.[58] When a publisher receives very positive and very negative reviews for the same manuscript, the editor will often solicit one or more additional reviews as a tie-breaker. As another strategy in the case of ties, the publisher may invite authors to reply to a referee's criticisms and permit a compelling rebuttal to break the tie. If a publisher does not feel confident to weigh the persuasiveness of a rebuttal, the publisher may solicit a response from the referee who made the original criticism. An editor may convey communications back and forth between authors and a referee, in effect allowing them to debate a point.[59]

Even in these cases, however, publishers do not allow multiple referees to confer with each other, though each reviewer may often see earlier comments submitted by other reviewers. The goal of the process is explicitly not to reach consensus or to persuade anyone to change their opinions, but instead to provide material for an informed editorial decision. One early study regarding referee disagreement found that agreement was greater than chance, if not much greater than chance, on six of seven article attributes (e.g. literature review and final recommendation to publish),[60] but this study was small and it was conducted on only one journal. At least one study has found that reviewer disagreement is not common, but this study is also small and on only one journal.[61]

Traditionally, reviewers would often remain anonymous to the authors, but this standard varies both with time and with academic field. In some academic fields, most journals offer the reviewer the option of remaining anonymous or not, or a referee may opt to sign a review, thereby relinquishing anonymity. Published papers sometimes contain, in the acknowledgments section, thanks to anonymous or named referees who helped improve the paper. For example, Nature journals provide this option.[56]

Sometimes authors may exclude certain reviewers: one study conducted on the Journal of Investigative Dermatology found that excluding reviewers doubled the chances of article acceptance.[62] Some scholars are uncomfortable with this idea, arguing that it distorts the scientific process. Others argue that it protects against referees who are biased in some manner (e.g. professional rivalry, grudges).[63] In some cases, authors can choose referees for their manuscripts. mSphere, an open-access journal in microbial science, has moved to this model. Editor-in-Chief Mike Imperiale says this process is designed to reduce the time it takes to review papers and permit the authors to choose the most appropriate reviewers.[64] But a scandal in 2015 shows how this choosing reviewers can encourage fraudulent reviews. Fake reviews were submitted to the Journal of the Renin-Angiotensin-Aldosterone System in the names of author-recommended reviewers, causing the journal to eliminate this option.[65]

Step 3: Revisions

If the manuscript has not been rejected during peer review, it returns to the authors for revisions. During this phase, the authors address the concerns raised by reviewers. William Stafford Noble offers ten rules for responding to reviewers. His rules include:[66]

  1. "Provide an overview, then quote the full set of reviews"
  2. "Be polite and respectful of all reviewers"
  3. "Accept the blame"
  4. "Make the response self-contained"
  5. "Respond to every point raised by the reviewer"
  6. "Use typography to help the reviewer navigate your response"
  7. "Whenever possible, begin your response to each comment with a direct answer to the point being raised"
  8. "When possible, do what the reviewer asks"
  9. "Be clear about what changed relative to the previous version"
  10. "If necessary, write the response twice" (i.e. write a version for "venting" but then write a version the reviewers will see)

Recruiting referees

At a journal or book publisher, the task of picking reviewers typically falls to an editor.[67] When a manuscript arrives, an editor solicits reviews from scholars or other experts who may or may not have already expressed a willingness to referee for that journal or book division. Granting agencies typically recruit a panel or committee of reviewers in advance of the arrival of applications.[68]

Referees are supposed to inform the editor of any conflict of interests that might arise. Journals or individual editors may invite a manuscript's authors to name people whom they consider qualified to referee their work. For some journals this is a requirement of submission. Authors are sometimes also given the opportunity to name natural candidates who should be disqualified, in which case they may be asked to provide justification (typically expressed in terms of conflict of interest).[citation needed]

Editors solicit author input in selecting referees because academic writing typically is very specialized. Editors often oversee many specialties, and can not be experts in all of them. But after an editor selects referees from the pool of candidates, the editor typically is obliged not to disclose the referees' identities to the authors, and in scientific journals, to each other. Policies on such matters differ among academic disciplines.[citation needed] One difficulty with respect to some manuscripts is that, there may be few scholars who truly qualify as experts, people who have themselves done work similar to that under review. This can frustrate the goals of reviewer anonymity and avoidance of conflicts of interest. Low-prestige or local journals and granting agencies that award little money are especially handicapped with regard to recruiting experts.[citation needed]

A potential hindrance in recruiting referees is that they are usually not paid, largely because doing so would itself create a conflict of interest. Also, reviewing takes time away from their main activities, such as his or her own research. To the would-be recruiter's advantage, most potential referees are authors themselves, or at least readers, who know that the publication system requires that experts donate their time. Serving as a referee can even be a condition of a grant, or professional association membership.[69] In general, because of the explosion of the electronic information and the disproportionate increase in journal number versus the steady increase in the number of scientists has created a peer review crisis. The system currently in place is not responding to modern needs and will inevitably perish, unless radical reforms are made promptly. The academic system should revolutionize and establish strict peer review activity criteria essential for promotion and tenure, based on established universal metrics. That is, reward reviewers academically as it rewards researchers, which is currently not the case. All other incentives have failed.[69]

Referees have the opportunity to prevent work that does not meet the standards of the field from being published, which is a position of some responsibility. Editors are at a special advantage in recruiting a scholar when they have overseen the publication of his or her work, or if the scholar is one who hopes to submit manuscripts to that editor's publishing entity in the future. Granting agencies, similarly, tend to seek referees among their present or former grantees.[citation needed]

Peerage of Science was an independent service and a community where reviewer recruitment happens via Open Engagement: authors submit their manuscript to the service where it is made accessible for any non-affiliated scientist, and 'validated users' choose themselves what they want to review. The motivation to participate as a peer reviewer comes from a reputation system where the quality of the reviewing work is judged and scored by other users, and contributes to user profiles. Peerage of Science does not charge any fees to scientists, and does not pay peer reviewers. Participating publishers however pay to use the service, gaining access to all ongoing processes and the opportunity to make publishing offers to the authors.[citation needed]

With independent peer review services the author usually retains the right to the work throughout the peer review process, and may choose the most appropriate journal to submit the work to.[70][71] Peer review services may also provide advice or recommendations on most suitable journals for the work. Journals may still want to perform an independent peer review, without the potential conflict of interest that financial reimbursement may cause, or the risk that an author has contracted multiple peer review services but only presents the most favorable one.[citation needed]

An alternative or complementary system of performing peer review is for the author to pay for having it performed. Example of such service provider was Rubriq (2013-2017), that for each work assigned peer reviewers who were financially compensated for their efforts.[72][73]

Different styles

Anonymous and attributed

For most scholarly publications, the identity of the reviewers is kept anonymised (also called "blind peer review"). The alternative, attributed peer review involves revealing the identities of the reviewers. Some reviewers choose to waive their right to anonymity, even when the journal's default format is blind peer review.

In anonymous peer review, reviewers are known to the journal editor or conference organiser but their names are not given to the article's author. In some cases, the author's identity can also be anonymised for the review process, with identifying information stripped from the document before review. The system is intended to reduce or eliminate bias.[2]

Some experts proposed blind review procedures for reviewing controversial research topics.[74]

In double-blind peer review, which has been fashioned by sociology journals in the 1950s[75] and remains more common in the social sciences and humanities than in the natural sciences,[citation needed] the identity of the authors is concealed from the reviewers ("blinded"), and vice versa, lest the knowledge of authorship or concern about disapprobation from the author bias their review.[76] Critics of the double-blind review process point out that, despite any editorial effort to ensure anonymity, the process often fails to do so, since certain approaches, methods, writing styles, notations, etc., point to a certain group of people in a research stream, and even to a particular person.[77][78]

In many fields of "big science", the publicly available operation schedules of major equipments, such as telescopes or synchrotrons, would make the authors' names obvious to anyone who would care to look them up. Proponents of double-blind review argue that it performs no worse than single-blind, and that it generates a perception of fairness and equality in academic funding and publishing.[79] Single-blind review is strongly dependent upon the goodwill of the participants, but no more so than double-blind review with easily identified authors.

As an alternative to single-blind and double-blind review, authors and reviewers are encouraged to declare their conflicts of interest when the names of authors and sometimes reviewers are known to the other. When conflicts are reported, the conflicting reviewer can be prohibited from reviewing and discussing the manuscript, or his or her review can instead be interpreted with the reported conflict in mind; the latter option is more often adopted when the conflict of interest is mild, such as a previous professional connection or a distant family relation. The incentive for reviewers to declare their conflicts of interest is a matter of professional ethics and individual integrity. Even when the reviews are not public, they are still a matter of record and the reviewer's credibility depends upon how they represent themselves among their peers. Some software engineering journals, such as the IEEE Transactions on Software Engineering, use non-blind reviews with reporting to editors of conflicts of interest by both authors and reviewers.[citation needed]

A more rigorous standard of accountability is known as an audit. Because reviewers are not paid, they cannot be expected to put as much time and effort into a review as an audit requires. Therefore, academic journals such as Science, organizations such as the American Geophysical Union, and agencies such as the National Institutes of Health and the National Science Foundation maintain and archive scientific data and methods in the event another researcher wishes to replicate or audit the research after publication.[80][81][82]

The traditional anonymous peer review has been criticized for its lack of accountability, the possibility of abuse by reviewers or by those who manage the peer review process (that is, journal editors),[83] its possible bias, and its inconsistency,[84] alongside other flaws.[85][86] Eugene Koonin, a senior investigator at the National Center for Biotechnology Information, asserts that the system has "well-known ills" and advocates "open peer review".[87]

Open peer review

Open peer review is the various possible modifications of the traditional scholarly peer review process. The three most common modifications to which the term is applied are:[25]

  1. Open identities: Authors and reviewers are aware of each other's identity.[88][89]
  2. Open reports: Review reports are published alongside the relevant article (rather than being kept confidential).
  3. Open participation: The wider community (and not just invited reviewers) are able to contribute to the review process.
These modifications are supposed to address various perceived shortcomings of the traditional scholarly peer review process, in particular its lack of transparency, lack of incentives, wastefulness,[25] bullying and harassment.[90]

Pre- and post-publication peer review

The process of peer review is not restricted to the publication process managed by academic journals. In particular, some forms of peer review can occur before an article is submitted to a journal and/or after it is published by the journal.

Pre-publication peer review

Manuscripts are typically reviewed by colleagues before submission, and if the manuscript is uploaded to preprint servers, such as ArXiv, BioRxiv or SSRN, researchers can read and comment on the manuscript. The practice to upload to preprint servers, and the activity of discussion heavily depend on the field,[91][92] and it allows an open pre-publication peer review. The advantage of this method is speed and transparency of the review process. Anyone can give feedback, typically in form of comments, and typically not anonymously. These comments are also public, and can be responded to, therefore author-reviewer communication is not restricted to the typical 2–4 rounds of exchanges in traditional publishing. The authors can incorporate comments from a wide range of people instead of feedback from the typically 3–4 reviewers. The disadvantage is that a far larger number of papers are presented to the community without any guarantee on quality.

Post-publication peer review

After a manuscript is published, the process of peer review continues as publications are read, known as post-publication peer review. Readers will often send letters to the editor of a journal, or correspond with the editor via an on-line journal club. In this way, all "peers" may offer review and critique of published literature. The introduction of the "epub ahead of print" practice in many journals has made possible the simultaneous publication of unsolicited letters to the editor together with the original paper in the print issue.[citation needed]

A variation on this theme is open peer commentary, in which commentaries from specialists are solicited on published articles and the authors are invited to respond. Journals using this process solicit and publish non-anonymous commentaries on the "target paper" together with the paper, and with original authors' reply as a matter of course. Open peer commentary was first implemented by the anthropologist Sol Tax, who founded the journal Current Anthropology in 1957.[93] The journal Behavioral and Brain Sciences, published by Cambridge University Press , was founded by Stevan Harnad in 1978[94] and modeled on Current Anthropology's open peer commentary feature.[95] Psycoloquy (1990–2002)[96] was based on the same feature, but this time implemented online. Since 2016 open peer commentary is also provided by the journal Animal Sentience.

In addition to journals hosting their own articles' reviews, there are also external, independent websites dedicated to post-publication peer-review, such as PubPeer which allows anonymous commenting of published literature and pushes authors to answer these comments.[97] It has been suggested that post-publication reviews from these sites should be editorially considered as well.[98] The megajournals F1000Research and ScienceOpen publish openly both the identity of the reviewers and the reviewer's report alongside the article.

Some journals use post-publication peer review as formal review method, instead of pre-publication review. This was first introduced in 2001, by Atmospheric Chemistry and Physics (ACP).[99] More recently F1000Research, Qeios, and ScienceOpen were launched as megajournals with post-publication review as formal review method.[100][101][102] At ACP, F1000Research, and Qeios peer reviewers are formally invited, much like at pre-publication review journals. Articles that pass peer review at those three journals are included in external scholarly databases.[103][104]

In 2006, a small group of UK academic psychologists launched Philica, the instant online journal Journal of Everything, to redress many of what they saw as the problems of traditional peer review. All submitted articles are published immediately and may be reviewed afterwards. Any researcher who wishes to review an article can do so and reviews are anonymous. Reviews are displayed at the end of each article, and are used to give the reader criticism or guidance about the work, rather than to decide whether it is published or not. This means that reviewers cannot suppress ideas if they disagree with them. Readers use reviews to guide their reading, and particularly popular or unpopular work is easy to identify.[citation needed]

Social media and informal peer review

Recent research has called attention to the use of social media technologies and science blogs as a means of informal, post-publication peer review, as in the case of the #arseniclife (or GFAJ-1) controversy.[7] In December 2010, an article published in Scienceexpress (the ahead-of-print version of Science) generated both excitement and skepticism, as its authors – led by NASA astrobiologist Felisa Wolfe-Simon – claimed to have discovered and cultured a certain bacteria that could replace phosphorus with arsenic in its physiological building blocks. At the time of the article's publication, NASA issued press statements suggesting that the finding would impact the search for extraterrestrial life, sparking excitement on Twitter under the hashtag #arseniclife, as well as criticism from fellow experts who voiced skepticism via their personal blogs.[105] Ultimately, the controversy surrounding the article attracted media attention,[106] and one of the most vocal scientific critics – Rosemary Redfield – formally published in July 2012[107] regarding her and her colleagues' unsuccessful attempt to replicate the NASA scientists' original findings.

Researchers following the impact of the #arseniclife case on social media discussions and peer review processes concluded the following:

Our results indicate that interactive online communication technologies can enable members in the broader scientific community to perform the role of journal reviewers to legitimize scientific information after it has advanced through formal review channels. In addition, a variety of audiences can attend to scientific controversies through these technologies and observe an informal process of post-publication peer review. (p 946)[7]

Result-blind peer review

Studies which report a positive or statistically significant result are far more likely to be published than ones which do not. A counter-measure to this positivity bias is to hide or make unavailable the results, making journal acceptance more like scientific grant agencies reviewing research proposals. Versions include:

  1. Result-blind peer review or results blind peer review, first proposed 1966: Reviewers receive an edited version of the submitted paper which omits the results and conclusion section.[108][109][110][111][112] In a two-stage version, a second round of reviews or editorial judgment is based on the full paper version, which was first proposed in 1977.[113][114]
    Conclusion-blind review, proposed by Robin Hanson in 2007 extends this further asking all authors to submit a positive and a negative version, and only after the journal has accepted the article authors reveal which is the real version.[115]
  2. Pre-accepted articles or outcome-unbiased journals or advance publication review or registered reports or prior to results submission or early acceptance[116][117][118][119][120][121][122][123][124] extends study pre-registration to the point that journals accepted or reject papers based on the version of the paper written before the results or conclusions have been made (an enlarged study protocol), but instead describes the theoretical justification, experimental design, and statistical analysis. Only once the proposed hypothesis and methodology have been accepted by reviewers, the authors would collect the data or analyze previously collected data. A limited variant of a pre-accepted article was The Lancet's study protocol review from 1997 to 2015 reviewed and published randomized trial protocols with a guarantee that the eventual paper would at least be sent out to peer review rather than immediately rejected.[125][126] For example, Nature Human Behaviour has adopted the registered report format, as it "shift[s] the emphasis from the results of research to the questions that guide the research and the methods used to answer them".[127] The European Journal of Personality defines this format: "In a registered report, authors create a study proposal that includes theoretical and empirical background, research questions/hypotheses, and pilot data (if available). Upon submission, this proposal will then be reviewed prior to data collection, and if accepted, the paper resulting from this peer-reviewed procedure will be published, regardless of the study outcomes."[128]

The following journals used result-blind peer review or pre-accepted articles:

  • The European Journal of Parapsychology, under Martin Johnson (who proposed a version of Registered Reports in 1974),[129] began accepting papers based on submitted designs and then publishing them, from 1976 to 1993, and published 25 RRs total[114]
  • The International Journal of Forecasting used opt-in result-blind peer review and pre-accepted articles from before 1986[130] through 1996/1997.[117][131]
  • The journal Applied Psychological Measurement offered an opt-in "advance publication review" process from 1989 to 1996, ending use after only 5 papers were submitted.[117][132]
  • The JAMA Internal Medicine found in a 2009 survey that 86% of its reviewers would be willing to work in a result-blind peer review process, and ran a pilot experiment with a two-stage result-blind peer review, showing the unblinded step benefited positive studies more than negatives.[133] but the journal does not currently use result-blind peer review.[citation needed]
  • The Center for Open Science encourages using "Registered Reports" (pre-accepted articles)[134][135][136][137][138][139] beginning in 2013. As of October 2017, ~80 journals offer Registered Reports in general, have had special issues of Registered Reports, or limited acceptance of Registered Reports (e.g. replications only) including AIMS Neuroscience, Cortex, Perspectives on Psychological Science, Social Psychology, & Comparative Political Studies
    • Comparative Political Studies published results of its pilot experiment of 19 submissions of which 3 were pre-accepted in 2016. the process worked well but submissions were weighted towards quantitative experimental designs, and reduced the amount of 'fishing' as submitters and reviewers focused on theoretical backing, substantive importance of results, with attention to the statistical power and implications of a null result, concluding that "we can clearly state that this form of review lead to papers that were of the highest quality. We would love to see a top journal adopt results-free review as a policy, at very least allowing results-free review as one among several standard submission options."[140]

Extended peer review

Extended peer review is the process of including people and groups with experience beyond that of working academics in the processes of assuring the quality of research. If conducted systematically, this can lead to more reliable, or applicable, results than a peer review process conducted purely by academics.[141]

Criticism

Various biomedical editors have expressed criticism of peer review.[142][143] In addition, a Cochrane review found little empirical evidence that peer review ensures quality in biomedical research,[144] while a second systematic review and meta-analysis found a need for evidence-based peer review in biomedicine given the paucity of assessment of the interventions designed to improve the process.[145]

To an outsider, the anonymous, pre-publication peer review process is opaque. Certain journals are accused of not carrying out stringent peer review in order to more easily expand their customer base, particularly in journals where authors pay a fee before publication.[146] Richard Smith, MD, former editor of the British Medical Journal, has claimed that peer review is "ineffective, largely a lottery, anti-innovatory, slow, expensive, wasteful of scientific time, inefficient, easily abused, prone to bias, unable to detect fraud and irrelevant; Several studies have shown that peer review is biased against the provincial and those from low- and middle-income countries; Many journals take months and even years to publish and the process wastes researchers' time. As for the cost, the Research Information Network estimated the global cost of peer review at £1.9 billion in 2008."[147]

In addition, Australia 's Innovative Research Universities group (a coalition of seven comprehensive universities committed to inclusive excellence in teaching, learning and research in Australia) has found that "peer review disadvantages researchers in their early careers, when they rely on competitive grants to cover their salaries, and when unsuccessful funding applications often mark the end of a research idea".[148]

There is an ongoing discussion about a peer-review crisis. In 2022 Inside Higher Ed reported a serious shortage of scholars to review submitted articles and bigger structural problems amplified by the COVID-19 pandemic.[149]

Low-end distinctions in articles understandable to all peers

Medical researcher John Ioannidis argues that since the exams and other tests that people pass on their way from "layman" to "expert" focus on answering the questions in time and in accordance with a list of answers, and not on making precise distinctions (the latter of which would be unrecognizable to experts of lower cognitive precision), there is as much individual variation in the ability to distinguish causation from correlation among "experts" as there is among "laymen". Ioannidis argues that as a result, scholarly peer review by many "experts" allows only articles that are understandable at a wide range of cognitive precision levels including very low ones to pass, biasing publications towards favoring articles that infer causation from correlation while mislabelling articles that make the distinction as "incompetent overestimation of one's ability" on the side of the authors because some of the reviewing "experts" are cognitively unable to distinguish the distinction from alleged rationalization of specific conclusions. It is argued by Ioannidis that this makes peer review a cause of selective publication of false research findings while stopping publication of rigorous criticism thereof, and that further post-publication review repeats the same bias by selectively retracting the few rigorous articles that may have made it through initial pre-publication peer review while letting the low-end ones that confuse correlation and causation remain in print.[150]

Peer review and trust

Researchers have peer-reviewed manuscripts prior to publishing them in a variety of ways since the 18th century.[151][152] The main goal of this practice is to improve the relevance and accuracy of scientific discussions. Even though experts often criticize peer review for a number of reasons, the process is still often considered the "gold standard" of science.[153] Occasionally however, peer review approves studies that are later found to be wrong and rarely deceptive or fraudulent results are discovered prior to publication.[154][155] Thus, there seems to be an element of discord between the ideology behind and the practice of peer review. By failing to effectively communicate that peer review is imperfect, the message conveyed to the wider public is that studies published in peer-reviewed journals are "true" and that peer review protects the literature from flawed science. A number of well-established criticisms exist of many elements of peer review.[156][157][158] In the following we describe cases of the wider impact inappropriate peer review can have on public understanding of scientific literature.

Multiple examples across several areas of science find that scientists elevated the importance of peer review for research that was questionable or corrupted. For example, climate change deniers have published studies in the Energy and Environment journal, attempting to undermine the body of research that shows how human activity impacts the Earth's climate. Politicians in the United States who reject the established science of climate change have then cited this journal on several occasions in speeches and reports.[159]

At times, peer review has been exposed as a process that was orchestrated for a preconceived outcome. The New York Times gained access to confidential peer review documents for studies sponsored by the National Football League (NFL) that were cited as scientific evidence that brain injuries do not cause long-term harm to its players.[160] During the peer review process, the authors of the study stated that all NFL players were part of a study, a claim that the reporters found to be false by examining the database used for the research. Furthermore, The Times noted that the NFL sought to legitimize the studies" methods and conclusion by citing a "rigorous, confidential peer-review process" despite evidence that some peer reviewers seemed "desperate" to stop their publication. Recent research has also demonstrated that widespread industry funding for published medical research often goes undeclared and that such conflicts of interest are not appropriately addressed by peer review.[161][162] Conflict of interest is less likely to be picked up in double-blinded reviews since the reviewer does not know the identity of the authors.[163]

Another problem that peer review fails to catch is ghostwriting, a process by which companies draft articles for academics who then publish them in journals, sometimes with little or no changes.[164] These studies can then be used for political, regulatory and marketing purposes. In 2010, the US Senate Finance Committee released a report that found this practice was widespread, that it corrupted the scientific literature and increased prescription rates.[165] Ghostwritten articles have appeared in dozens of journals, involving professors at several universities.[166]

Just as experts in a particular field have a better understanding of the value of papers published in their area, scientists are considered to have better grasp of the value of published papers than the general public and to see peer review as a human process, with human failings,[167] and that "despite its limitations, we need it. It is all we have, and it is hard to imagine how we would get along without it".[168] But these subtleties are lost on the general public, who are often misled into thinking that being published in a journal with peer review is the "gold standard" and can erroneously equate published research with the truth.[167] Thus, more care must be taken over how peer review, and the results of peer-reviewed research, are communicated to non-specialist audiences; particularly during a time in which a range of technical changes and a deeper appreciation of the complexities of peer review are emerging.[169][170][171][172] This will be needed as the scholarly publishing system has to confront wider issues such as retractions[155][173][174] and replication or reproducibility "crises".[175][176][177]

Views of peer review

Peer review is often considered integral to scientific discourse in one form or another. Its gatekeeping role is supposed to be necessary to maintain the quality of the scientific literature[178][179] and avoid a risk of unreliable results, inability to separate signal from noise, and slow scientific progress.[180][181]

Shortcomings of peer review have been met with calls for even stronger filtering and more gatekeeping. A common argument in favor of such initiatives is the belief that this filter is needed to maintain the integrity of the scientific literature.[182][183]

Calls for more oversight have at least two implications that are counterintuitive of what is known to be true scholarship.[167]

  1. The belief that scholars are incapable of evaluating the quality of work on their own, that they are in need of a gatekeeper to inform them of what is good and what is not.
  2. The belief that scholars need a "guardian" to make sure they are doing good work.

Others argue[167] that authors most of all have a vested interest in the quality of a particular piece of work. Only the authors could have, as Feynman (1974)[184] puts it, the "extra type of integrity that is beyond not lying, but bending over backwards to show how you're maybe wrong, that you ought to have when acting as a scientist." If anything, the current peer review process and academic system could penalize, or at least fail to incentivize, such integrity.

Instead, the credibility conferred by the "peer-reviewed" label could diminish what Feynman calls the culture of doubt necessary for science to operate a self-correcting, truth-seeking process.[185] The effects of this can be seen in the ongoing replication crisis, hoaxes, and widespread outrage over the inefficacy of the current system.[156][151] It's common to think that more oversight is the answer, as peer reviewers are not at all lacking in skepticism. But the issue is not the skepticism shared by the select few who determine whether an article passes through the filter. It is the validation, and accompanying lack of skepticism, that comes afterwards.[186] Here again more oversight only adds to the impression that peer review ensures quality, thereby further diminishing the culture of doubt and counteracting the spirit of scientific inquiry.[187]

Quality research – even some of our most fundamental scientific discoveries – dates back centuries, long before peer review took its current form.[151][188][152] Whatever peer review existed centuries ago, it took a different form than it does in modern times, without the influence of large, commercial publishing companies or a pervasive culture of publish or perish.[188] Though in its initial conception it was often a laborious and time-consuming task, researchers took peer review on nonetheless, not out of obligation but out of duty to uphold the integrity of their own scholarship. They managed to do so, for the most part, without the aid of centralised journals, editors, or any formalised or institutionalised process whatsoever. Supporters of modern technology argue[167] that it makes it possible to communicate instantaneously with scholars around the globe, make such scholarly exchanges easier, and restore peer review to a purer scholarly form, as a discourse in which researchers engage with one another to better clarify, understand, and communicate their insights.[170][189]

Such modern technology includes posting results to preprint servers, preregistration of studies, open peer review, and other open science practices.[176][190][191] In all these initiatives, the role of gatekeeping remains prominent, as if a necessary feature of all scholarly communication, but critics argue[158] that a proper, real-world implementation could test and disprove this assumption; demonstrate researchers' desire for more that traditional journals can offer; show that researchers can be entrusted to perform their own quality control independent of journal-coupled review. Jon Tennant also argues that the outcry over the inefficiencies of traditional journals centers on their inability to provide rigorous enough scrutiny, and the outsourcing of critical thinking to a concealed and poorly-understood process. Thus, the assumption that journals and peer review are required to protect scientific integrity seems to undermine the very foundations of scholarly inquiry.[167]

To test the hypothesis that filtering is indeed unnecessary to quality control, many of the traditional publication practices would need to be redesigned, editorial boards repurposed if not disbanded, and authors granted control over the peer review of their own work. Putting authors in charge of their own peer review is seen as serving a dual purpose.[167] On one hand, it removes the conferral of quality within the traditional system, thus eliminating the prestige associated with the simple act of publishing. Perhaps paradoxically, the removal of this barrier might actually result in an increase of the quality of published work, as it eliminates the cachet of publishing for its own sake. On the other hand, readers know that there is no filter so they must interpret anything they read with a healthy dose of skepticism, thereby naturally restoring the culture of doubt to scientific practice.[192][193][194]

In addition to concerns about the quality of work produced by well-meaning researchers, there are concerns that a truly open system would allow the literature to be populated with junk and propaganda by those with a vested interest in certain issues. A counterargument is that the conventional model of peer review diminishes the healthy skepticism that is a hallmark of scientific inquiry, and thus confers credibility upon subversive attempts to infiltrate the literature.[167] Allowing such "junk" to be published could make individual articles less reliable but render the overall literature more robust by fostering a "culture of doubt".[192]

Allegations of bias and suppression

The interposition of editors and reviewers between authors and readers may enable the intermediators to act as gatekeepers.[195] Some sociologists of science argue that peer review makes the ability to publish susceptible to control by elites and to personal jealousy.[196] The peer review process may sometimes impede progress and may be biased against novelty.[38][197] A linguistic analysis of review reports suggests that reviewers focus on rejecting the applications by searching for weak points, and not on finding the high-risk/high-gain groundbreaking ideas that may be in the proposal.[198] Reviewers tend to be especially critical of conclusions that contradict their own views,[199][200] and lenient towards those that match them. At the same time, established scientists are more likely than others to be sought out as referees, particularly by high-prestige journals/publishers. As a result, ideas that harmonize with the established experts' are more likely to see print and to appear in premier journals than are iconoclastic or revolutionary ones. This accords with Thomas Kuhn's well-known observations regarding scientific revolutions.[201] A theoretical model has been established whose simulations imply that peer review and over-competitive research funding foster mainstream opinion to monopoly.[202]

Criticisms of traditional anonymous peer review allege that it lacks accountability, can lead to abuse by reviewers, and may be biased and inconsistent.[86][84][203]

There have also been suggestions of gender bias in peer review, with male authors being likely to receive more favorable treatment.[204] However, a 2021 study found no evidence for such bias (and found that in some respects female authors were treated more favourably).[205][206][207]

Exploitation of free work

Most academic publishers do not financially compensate reviewers for their participation in the peer-review process, which has been criticized by the academic community.[208] Whereas some publishers have contended that it is economically not feasible to pay reviewers,[208] some journals have started to pay reviewers through platforms such as Research Square when they are unable to receive free reviews. Other publishers such as Advances.in have made paying reviewers an inherent part of their business model.[209]

Open access journals and peer review

Some critics of open access (OA) journals have argued that, compared to traditional subscription journals, open access journals might utilize substandard or less formal peer review practices, and, as a consequence, the quality of scientific work in such journals will suffer.[210][unreliable source?] In a study published in 2012, this hypothesis was tested by evaluating the relative "impact" (using citation counts) of articles published in open access and subscription journals, on the grounds that members of the scientific community would presumably be less likely to cite substandard work, and that citation counts could therefore act as one indicator of whether or not the journal format indeed impacted peer review and the quality of published scholarship.[211] This study ultimately concluded that "OA journals indexed in Web of Science and/or Scopus are approaching the same scientific impact and quality as subscription journals, particularly in biomedicine and for journals funded by article processing charges," and the authors consequently argue that "there is no reason for authors not to choose to publish in OA journals just because of the 'OA' label.

Failures

Peer review fails when a peer-reviewed article contains fundamental errors that undermine at least one of its main conclusions and that could have been identified by more careful reviewers. Many journals have no procedure to deal with peer review failures beyond publishing letters to the editor.[212] Peer review in scientific journals assumes that the article reviewed has been honestly prepared. The process occasionally detects fraud, but is not designed to do so.[213] When peer review fails and a paper is published with fraudulent or otherwise irreproducible data, the paper may be retracted. A 1998 experiment on peer review with a fictitious manuscript found that peer reviewers failed to detect some manuscript errors and the majority of reviewers may not notice that the conclusions of the paper are unsupported by its results.[214]

Fake peer review

There have been instances where peer review was claimed to be performed but in fact was not; this has been documented in some predatory open access journals (e.g., the Who's Afraid of Peer Review? affair) or in the case of sponsored Elsevier journals.[215]

In November 2014, an article in Nature exposed that some academics were submitting fake contact details for recommended reviewers to journals, so that if the publisher contacted the recommended reviewer, they were the original author reviewing their own work under a fake name.[216] The Committee on Publication Ethics issued a statement warning of the fraudulent practice.[217] In March 2015, BioMed Central retracted 43 articles[218] and Springer retracted 64 papers in 10 journals in August 2015.[219] Tumor Biology journal is another example of peer review fraud.[215]

In 2020, the Journal of Nanoparticle Research fell victim to an "organized rogue editor network", who impersonated respected academics, got a themed issue created, and got 19 substandard articles published (out of 80 submitted).[220][221] The journal was praised for dealing with the scam openly and transparently.[220]

Plagiarism

Reviewers generally lack access to raw data, but do see the full text of the manuscript, and are typically familiar with recent publications in the area. Thus, they are in a better position to detect plagiarism of prose than fraudulent data. A few cases of such textual plagiarism by historians, for instance, have been widely publicized.[222]

On the scientific side, a poll of 3,247 scientists funded by the U.S. National Institutes of Health found 0.3% admitted faking data and 1.4% admitted plagiarism.[223] Additionally, 4.7% of the same poll admitted to self-plagiarism or autoplagiarism, in which an author republishes the same material, data, or text, without citing their earlier work.[223][clarification needed] Self-plagiarisms are less likely to be detected in double-blinded peer reviews.[163]

Examples

  • "Perhaps the most widely recognized failure of peer review is its inability to ensure the identification of high-quality work. The list of important scientific papers that were rejected by some peer-reviewed journals goes back at least as far as the editor of Philosophical Transaction's 1796 rejection of Edward Jenner's report of the first vaccination against smallpox."[224]
  • The Soon and Baliunas controversy involved the publication in 2003 of a review study written by aerospace engineer Willie Soon and astronomer Sallie Baliunas in the journal Climate Research,[225] which was quickly taken up by the G.W. Bush administration as a basis for amending the first Environmental Protection Agency Report on the Environment. The paper was strongly criticized by numerous scientists for its methodology and for its misuse of data from previously published studies, prompting concerns about the peer review process of the paper. The controversy resulted in the resignation of several editors of the journal and the admission by its publisher Otto Kinne that the paper should not have been published as it was.
  • The trapezoidal rule, in which the method of Riemann sums for numerical integration was republished in a Diabetes research journal, Diabetes Care.[226] The method is almost always taught in high school calculus, and was thus considered an example of an extremely well known idea being re-branded as a new discovery.[227]
  • A conference organized by the Wessex Institute of Technology was the target of an exposé by three researchers who wrote nonsensical papers (including one that was composed of random phrases). They reported that the papers were "reviewed and provisionally accepted" and concluded that the conference was an attempt to "sell" publication possibilities to less experienced or naive researchers.[228] This may however be better described as a lack of any actual peer review, rather than peer review having failed.
  • In the humanities, one of the most infamous cases of plagiarism undetected by peer review involved Martin Stone, formerly professor of medieval and Renaissance philosophy at the Hoger Instituut voor Wijsbegeerte of the KU Leuven. Martin Stone managed to publish at least forty articles and book chapters that were almost entirely stolen from the work of others. Most of these publications appeared in highly rated peer-reviewed journals and book series.[229][230]
  • The controversial Younger Dryas impact hypothesis, which evolved directly from pseudoscience and now forms the basis for the pseudoarchaeology of Graham Hancock's Ancient Apocalypse, was first published in the peer-reviewed journal PNAS using a nonstandard review system, according to a comprehensive refutation by Holliday et al. (2023).[231] According to this 2023 review, "Claiming evidence where none exists and providing misleading citations may be accidental, but when conducted repeatedly, it becomes negligent and undermines scientific advancement as well as the credibility of science itself. Also culpable is the failure of the peer review process to prevent such errors of fact from entering the literature. The Proceedings of the National Academy of Sciences 'contributed review' system for National Academy members...is at least partially responsible. The 'pal reviews' (as some refer to them) were significantly curtailed in 2010, in part due to the YDIH controversy."

In popular culture

In 2017, the Higher School of Economics in Moscow unveiled a "Monument to an Anonymous Peer Reviewer". It takes the form of a large concrete cube, or dice, with "Accept", "Minor Changes", "Major Changes", "Revise and Resubmit" and "Reject" on its five visible sides. Sociologist Igor Chirikov, who devised the monument, said that while researchers have a love-hate relationship with peer review, peer reviewers nonetheless do valuable but mostly invisible work, and the monument is a tribute to them.[232]

See also

References

  1. Schimanski, Lesley A.; Alperin, Juan Pablo (2018). "The evaluation of scholarship in academic promotion and tenure processes: Past, present, and future". F1000Research 7: 1605. doi:10.12688/f1000research.16493.1. ISSN 2046-1402. PMID 30647909. 
  2. 2.0 2.1 2.2 2.3 Hatch, Robert A. (February 1998). "The Scientific Revolution: Correspondence Networks". University of Florida. http://www.clas.ufl.edu/users/rhatch/pages/03-Sci-Rev/SCI-REV-Home/resource-ref-read/correspond-net/08sr-crrsp.htm.  Cite error: Invalid <ref> tag; name ":0" defined multiple times with different content
  3. Oldenburg, Henry (1665). "Epistle Dedicatory". Philosophical Transactions of the Royal Society 1: 0. doi:10.1098/rstl.1665.0001. 
  4. Hall, Marie Boas (2002). Henry Oldenburg: shaping the Royal Society. Oxford: Oxford University Press. ISBN 978-0-19-851053-6. Bibcode2002heol.book.....B. 
  5. Spier, Ray (2002). "The history of the peer-review process". Trends in Biotechnology 20 (8): 357–8. doi:10.1016/S0167-7799(02)01985-6. PMID 12127284. 
  6. Dans, PE (1993). "Clinical peer review: burnishing a tarnished image". Annals of Internal Medicine 118 (7): 566–8. doi:10.7326/0003-4819-118-7-199304010-00014. PMID 8442628. http://www.annals.org/content/118/7/566.full.pdf+html. 
  7. 7.0 7.1 7.2 Milgrom P, Weinstein P, Ratener P, Read WA, Morrison K; Weinstein; Ratener; Read; Morrison (1978). "Dental Examinations for Quality Control: Peer Review versus Self-Assessment". American Journal of Public Health 68 (4): 394–401. doi:10.2105/AJPH.68.4.394. PMID 645987.  Cite error: Invalid <ref> tag; name ":1" defined multiple times with different content
  8. "AICPA Peer Review Program Manual". American Institute of CPAs. http://www.aicpa.org/INTERESTAREAS/PEERREVIEW/RESOURCES/PEERREVIEWPROGRAMMANUAL/Pages/default.aspx. 
  9. "Peer Review". UK Legal Services Commission. 12 July 2007. http://www.legalservices.gov.uk/civil/how/mq_peerreview.asp. 
  10. "Martindale-Hubbell Attorney Reviews and Ratings". Martindale. https://www.martindale.com/ratings-and-reviews/. 
  11. 11.0 11.1 "Peer Review Panels – Purpose and Process". USDA Forest Service. February 6, 2006. http://www.fs.fed.us/fire/doctrine/mgmt/briefing_papers/peer_review_panels.pdf.  Cite error: Invalid <ref> tag; name "fire" defined multiple times with different content
  12. 12.0 12.1 Sims Gerald K. (1989). "Student Peer Review in the Classroom: A Teaching and Grading Tool". Journal of Agronomic Education 18 (2): 105–108. doi:10.2134/jae1989.0105. https://www.agronomy.org/files/publications/jnrlse/pdfs/jnr018/018-02-0105.pdf. "The review process was double-blind to provide anonymity for both authors and reviewers, but was otherwise handled in a fashion similar to that used by scientific journals". 
  13. Liu, Jianguo; Pysarchik, Dawn Thorndike; Taylor, William W. (2002). "Peer Review in the Classroom". BioScience 52 (9): 824–829. doi:10.1641/0006-3568(2002)052[0824:PRITC2.0.CO;2]. http://chans-net.org/sites/chans-net.org/files/peer_review.pdf. 
  14. Kupferschmidt, Kai (2018-08-14). "Researcher at the center of an epic fraud remains an enigma to those who exposed him". Science. doi:10.1126/science.aav1079. 
  15. 15.0 15.1 "Biomedical publishing. Secretive and subjective, peer review proves resistant to study". Science 341 (6152): 1331. September 2013. doi:10.1126/science.341.6152.1331. PMID 24052283. 
  16. Squazzoni, Flaminio; Brezis, Elise; Marušić, Ana (2017-10-01). "Scientometrics of peer review" (in en). Scientometrics 113 (1): 501–502. doi:10.1007/s11192-017-2518-4. ISSN 1588-2861. PMID 29056787. PMC 5629222. https://doi.org/10.1007/s11192-017-2518-4. 
  17. Ragone, Azzurra; Mirylenka, Katsiaryna; Casati, Fabio; Marchese, Maurizio (2013-11-01). "On peer review in computer science: analysis of its effectiveness and suggestions for improvement". Scientometrics 97 (2): 317–356. doi:10.1007/s11192-013-1002-z. ISSN 0138-9130. https://doi.org/10.1007/s11192-013-1002-z. 
  18. 18.0 18.1 18.2 18.3 Brezis, Elise S.; Birukou, Aliaksandr (2020-04-01). "Arbitrariness in the peer review process" (in en). Scientometrics 123 (1): 393–411. doi:10.1007/s11192-020-03348-1. ISSN 1588-2861.  CC-BY icon.svg Text was copied from this source, which is available under a Creative Commons Attribution 4.0 International License.
  19. Martin, B.. "Chapter 5: Peer review as scholarly conformity". https://www.bmartin.cc/dissent/documents/ss/ss5.html. 
  20. Rennie, Drummond (7 July 2016). "Let's make peer review scientific". Nature 535 (7610): 31–33. doi:10.1038/535031a. PMID 27383970. Bibcode2016Natur.535...31R. 
  21. Slavov, Nikolai (2015-11-11). "Making the most of peer review". eLife 4: e12708. doi:10.7554/eLife.12708. ISSN 2050-084X. PMID 26559758. 
  22. Couzin-Frankel, Jennifer (18 September 2018). "'Journalologists' use scientific methods to study academic publishing. Is their work improving science?". Science. doi:10.1126/science.aav4758. 
  23. Cosgrove, Andrew; Cheifet, Barbara (2018-11-27). "Transparent peer review trial: the results". Genome Biology 19 (1): 206. doi:10.1186/s13059-018-1584-0. ISSN 1474-760X. PMID 30482224. 
  24. Patterson, Mark; Schekman, Randy (2018-06-26). "A new twist on peer review". eLife 7: e36545. doi:10.7554/eLife.36545. ISSN 2050-084X. PMID 29944117. 
  25. 25.0 25.1 25.2 Ross-Hellauer, Tony (2017-08-31). "What is open peer review? A systematic review". F1000Research 6: 588. doi:10.12688/f1000research.11369.2. ISSN 2046-1402. PMID 28580134. 
  26. Else, Holly (2022-11-03). "eLife won't reject papers once they are under review — what researchers think". Nature. doi:10.1038/d41586-022-03534-6. ISSN 0028-0836. PMID 36329143. 
  27. Dyke, Gareth (2023). "Interview with Dr. Neeraj Kumar Sethiya, the newest ReviewerCredits ambassador". https://www.reviewercredits.com/interview-with-dr-neeraj-kumar-sethiya-the-newest-reviewercredits-ambassador/. 
  28. 28.0 28.1 Steinzor, Rena (2006). Rescuing Science from Politics. Cambridge University Press. p. 304. ISBN 978-0521855204. https://books.google.com/books?id=VZ1axrolCpkC&pg=PA220. 
  29. Committee on Science, Engineering, and Public Policy, National Academy of Sciences, National Academy of Engineering, and Institute of Medicine On Being a Scientist: A Guide to Responsible Conduct in Research National Academies Press, Washington, D.C., 1995, 82 pages, ISBN:0309119707[page needed]
  30. "The Origin of the Scientific Journal and the Process of Peer Review", House of Commons Select Committee Report
  31. "The ups and downs of peer review". Advances in Physiology Education 31 (2): 145–152 [145]. June 2007. doi:10.1152/advan.00104.2006. PMID 17562902. "Scientific peer review has been defined as the evaluation of research findings for competence, significance, and originality by qualified experts. These peers act as sentinels on the road of scientific discovery and publication.". 
  32. Blow, Nathan S. (January 2015). "Benefits and Burdens of Peer-Review". BioTechniques 58 (1): 5. doi:10.2144/000114242. http://www.biotechniques.com/BiotechniquesJournal/2015/January/From-the-Editor/biotechniques-356227.html. 
  33. "Benefits and Burdens of Peer-Review". BioTechniques 58 (1): 5. January 2015. 
  34. 34.0 34.1 Pontille, David; Torny, Didier (2014). "From manuscript evaluation to article valuation: the changing technologies of journal peer review". Human Studies 38: 57–79. doi:10.1007/s10746-014-9335-z. https://hal.archives-ouvertes.fr/CSI-ENSMP/hal-01143310v1. 
  35. "Peer review: Troubled from the start". Nature 532 (7599): 306–308. April 2016. doi:10.1038/532306a. PMID 27111616. Bibcode2016Natur.532..306C. 
  36. Kennefick, Daniel (September 2005). "Einstein versus the Physical Review". Physics Today 58 (9): 43–48. doi:10.1063/1.2117822. Bibcode2005PhT....58i..43K. 
  37. 37.0 37.1 Baldwin, Melinda (2018-09-01). "Scientific Autonomy, Public Accountability, and the Rise of "Peer Review" in the Cold War United States". Isis 109 (3): 538–558. doi:10.1086/700070. ISSN 0021-1753. 
  38. 38.0 38.1 "Coping with peer rejection". Nature 425 (6959): 645. October 2003. doi:10.1038/425645a. PMID 14562060. Bibcode2003Natur.425..645.. 
  39. Tan, Meng H. (2018). "Chapter 7: Peer review – Past, Present and Future". in Markovac, Jasna; Kleinman, Molly; Englesbe, Michael. Medical and Scientific Publishing: Author, Editor, and Reviewer Perspectives. Academic Press. pp. 55–68. ISBN 978-0128099698. 
  40. 40.0 40.1 Gaudet, Joanne (2014-07-16). "Investigating journal peer review as scientific object of study:unabridged version – Part I". UO Research. 
  41. Gould, T.P.H. (2012). Do We Still Need Peer Review?. The Scarecrow Press. ISBN 978-0810885745. [page needed]
  42. Biagioli, M. (2002). "From book censorship to academic peer review". Emergences 12 (1): 11–45. doi:10.1080/1045722022000003435. 
  43. Rip, A. (1985). "Commentary: Peer review is alive and well in the United States". Science, Technology, & Human Values 10 (3): 82–86. doi:10.1177/016224398501000310. 
  44. "Three Decades of Peer Review Congresses". JAMA 319 (4): 350–353. January 2018. doi:10.1001/jama.2017.20606. PMID 29362775. 
  45. Hirschauer, S. (2010). "Editorial judgements: A praxeology of 'voting' in peer review". Social Studies of Science 40 (1): 71–103. doi:10.1177/0306312709335405. 
  46. Roos, A. M. (2022). "Publish and flourish, or the collective wisdom of peer review". Notes and Records 76 (3): 335–6. doi:10.1098/rsnr.2022.0022. PMID 35974900. 
  47. "AAUP Membership Benefits and Eligibility". Association of American University Presses. http://www.aaupnet.org/aaup-members/becoming-a-member/eligibility-requirements. 
  48. "Peer Review". https://www.springer.com/gp/authors-editors/editors/peer-review/32888. 
  49. Spicer, Andre. "Explainer: what is peer review?". The Conversation. https://theconversation.com/explainer-what-is-peer-review-27797. 
  50. Stolowy, Herve (2017). "Letter from the Editor: Why Are Papers Desk Rejected at European Accounting Review". European Accounting Review 26 (3): 411–418. doi:10.1080/09638180.2017.1347360. 
  51. McKenzie, David (2017-02-21). "The State of Development Journals 2017: Quality, Acceptance Rates, and Review Times". http://blogs.worldbank.org/impactevaluations/dev4peace/state-development-journals-2017-quality-acceptance-rates-and-review-times. 
  52. "Summary Report of Journal Operations, 2016". American Psychologist 72 (5): 499–500. 2017. doi:10.1037/amp0000172. PMID 28726464. 
  53. "Peer review process and editorial decision making at journals". 2013-04-11. https://www.editage.com/insights/peer-review-process-and-editorial-decision-making-at-journals. 
  54. Zwaaf, Elizabeth (15 January 2013). "8 reasons I accepted your article". Elsevier Connect. https://www.elsevier.com/connect/8-reasons-i-accepted-your-article. 
  55. "How to review a paper". Advances in Physiology Education 27 (1–4): 47–52. December 2003. doi:10.1152/advan.00057.2002. PMID 12760840. 
  56. 56.0 56.1 "Peer-review policy : authors & referees @ npg". https://www.nature.com/authors/policies/peer_review.html. 
  57. "Recommendations for the Conduct, Reporting, Editing, and Publication of Scholarly work in Medical Journals.". ICMJE. 2014-12-16. http://www.icmje.org/recommendations/. 
  58. "What should editors do when referees disagree?". Dynamic Ecology. 2014-09-02. https://dynamicecology.wordpress.com/2014/09/02/what-should-editors-do-when-referees-disagree/. 
  59. Coleman, Andrew M. (1979). "Editorial role in author-referee disagreements". Bulletin of the British Psychological Society 32: 390–391. https://www2.le.ac.uk/departments/npb/people/amc/articles-pdfs/editrole.pdf. 
  60. Scott, W. A. (1974). "Interreferee agreement on some characteristics of manuscripts submitted to the Journal of Personality and Social Psychology". American Psychologist 29 (9): 698–702. doi:10.1037/h0037631. 
  61. Pless, I. B. (August 2006). "When reviewers disagree". Injury Prevention 12 (4): 211. doi:10.1136/ip.2006.090806. PMID 16887940. 
  62. Grimm, David (2005). "Suggesting or Excluding Reviewers Can Help Get Your Paper Published". Science 309 (5743): 1974. doi:10.1126/science.309.5743.1974. PMID 16179438. 
  63. Grimm, David (2005-09-23). "Reviewers Can Help Get Your Paper Published". Science. doi:10.1126/article.61887. 
  64. "Journal's new program: Choose your own reviewers – and get a decision in days". Retraction Watch. 2016-12-12. https://retractionwatch.com/2016/12/12/journals-new-program-choose-reviewers-get-decision-days/. 
  65. "Eight retractions for fake reviews lead journal to suspend author nominations". Retraction Watch. 2015-12-24. https://retractionwatch.com/2015/12/24/eight-retractions-for-fake-reviews-leads-journal-to-suspend-author-nominations/. 
  66. "Ten simple rules for writing a response to reviewers". PLOS Computational Biology 13 (10): e1005730. October 2017. doi:10.1371/journal.pcbi.1005730. PMID 29023444. Bibcode2017PLSCB..13E5730N. 
  67. Lawrence O'Gorman (January 2008). "The (Frustrating) State of Peer Review". IAPR Newsletter 30 (1): 3–5. http://iapr.org/docs/newsletter-2008-01.pdf. 
  68. Schwartz, Samuel M.; Slater, Donald W.; Heydrick, Fred P.; Woolett, Gillian R. (September 1995). "A Report of the AIBS Peer-Review Process for the US Army's 1994 Breast Cancer Initiative". BioScience 45 (8): 558–563. doi:10.1093/bioscience/45.8.558. 
  69. 69.0 69.1 Chloros, George D.; Giannoudis, Vasileios P.; Giannoudis, Peter V. (2022-01-01). "Peer-reviewing in Surgical Journals: Revolutionize or Perish?". Annals of Surgery 275 (1): e82–e90. doi:10.1097/SLA.0000000000004756. ISSN 1528-1140. PMID 33630457. https://pubmed.ncbi.nlm.nih.gov/33630457. 
  70. Hames, Irene (2014). "The changing face of peer review". Science Editing 1: 9–12. doi:10.6087/kcse.2014.1.9. 
  71. "Journal publishing: the changing landscape". The Indian Journal of Medical Research 138 (1): 4–7. 2013. PMID 24056548. 
  72. Stemmle, Laura; Collier, Keith (2013). "RUBRIQ: Tools, services, and software to improve peer review". Learned Publishing 26 (4): 265–268. doi:10.1087/20130406. 
  73. Davis, Phil (2017-09-25). "Portable Peer Review RIP". https://scholarlykitchen.sspnet.org/2017/09/25/portable-peer-review-rip/. 
  74. Armstrong, J. Scott (1982). "Research on scientific journals: Implications for editors and authors". Journal of Forecasting 1: 83–104. doi:10.1002/for.3980010109. 
  75. Pontille, David; Torny, Didier (2014). "The Blind Shall See! The Question of Anonymity in Journal Peer Review". Ada: A Journal of Gender, New Media, and Technology 4. doi:10.7264/N3542KVW. 
  76. Cressey, Daniel (2014). "Journals weigh up double-blind peer review". Nature. doi:10.1038/nature.2014.15564. 
  77. Markus, Annette (28 Dec 2005). "Double-blind peer review?". http://blogs.nature.com/actionpotential/2005/12/doubleblind_peer_review.html. 
  78. "Working double-blind". Nature 451 (7179): 605–606. 2008. doi:10.1038/451605b. PMID 18256621. Bibcode2008Natur.451R.605.. 
  79. "Peer review--the newcomers' perspective". PLOS Biology 3 (9): e326. September 2005. doi:10.1371/journal.pbio.0030326. PMID 16149851. 
  80. "Policy on Referencing Data in and Archiving Data for AGU Publications". American Geophysical Union. 2012. http://www.agu.org/pubs/authors/policies/data_policy.shtml. "The following policy has been adopted for AGU publications in order to ensure that they can effectively and efficiently perform an expanded role in making the underlying data for articles available to researchers now and in the future." 
    • This policy was first adopted by the AGU Publications Committee in November 1993 and then revised March 1994, December 1995, October 1996.
    • See also AGU Data Policy by Bill Cook. April 4, 2012.
  81. "Data Management & Sharing Frequently Asked Questions". National Science Foundation. November 30, 2010. https://www.nsf.gov/bfa/dias/policy/dmpfaqs.jsp. 
  82. Moore, R.W.; Rajasekar, A.; Wan, M. (2005). "Data Grids, Digital Libraries, and Persistent Archives: An Integrated Approach to Sharing, Publishing, and Archiving Data". Proceedings of the IEEE 93 (3): 578–588. doi:10.1109/JPROC.2004.842761. 
  83. Bingham, Craig (2000). "Peer review and the ethics of internet publishing". in Jones, Anne Hudson; McLellan, Faith. Ethical Issues in Biomedical Publication. Baltimore: Johns Hopkins University. pp. 85–111. ISBN 978-0801863158. https://books.google.com/books?id=d_U4BjPtI7MC&pg=PA85. 
  84. 84.0 84.1 "Reproducibility of peer review in clinical neuroscience. Is agreement between reviewers any greater than would be expected by chance alone?". Brain 123 (9): 1964–1969. September 2000. doi:10.1093/brain/123.9.1964. PMID 10960059. 
  85. "The Peer Review Process". http://www.jisc.ac.uk/uploaded_documents/rowland.pdf. 
  86. 86.0 86.1 Alison McCook (February 2006). "Is Peer Review Broken?". The Scientist. http://www.the-scientist.com/?articles.view/articleNo/23672/title/Is-Peer-Review-Broken-/. 
  87. Koonin, Eugene (2006). "Reviving a culture of scientific debate". Nature. doi:10.1038/nature05005. http://www.nature.com/nature/peerreview/debate/nature05005.html. 
  88. "Open peer review: a randomised controlled trial". The British Journal of Psychiatry 176 (1): 47–51. January 2000. doi:10.1192/bjp.176.1.47. PMID 10789326. 
  89. "Effect of open peer review on quality of reviews and on reviewers' recommendations: a randomised trial". BMJ 318 (7175): 23–7. January 1999. doi:10.1136/bmj.318.7175.23. PMID 9872878. 
  90. Sanders, Jeremy K. M. (January 2020). "Editorial 2020: Changing publishing and academic culture" (in en). Royal Society Open Science 7 (1): 192197. doi:10.1098/rsos.192197. ISSN 2054-5703. PMID 32218987. Bibcode2020RSOS....792197S. 
  91. "The Stars Are Aligning for Preprints". The Scholarly Kitchen. 2017-04-18. https://scholarlykitchen.sspnet.org/2017/04/18/stars-aligning-preprints/. 
  92. "Biology preprints over time | ASAPbio". http://asapbio.org/preprint-info/biology-preprints-over-time. 
  93. "Obituary: Sol Tax, Anthropology". http://chronicle.uchicago.edu/950119/tax.shtml. 
  94. "Editorial". Behavioral and Brain Sciences 1: 1–2. 1978. doi:10.1017/S0140525X00059045. 
  95. New Scientist, 20 March 1980, p. 945
  96. Stevan Harnad (1991). "Post-Gutenberg Galaxy: The Fourth Revolution in the Means of Production of Knowledge". Public-Access Computer Systems Review 2 (1): 39–53. http://cogprints.org/1580/. Retrieved 2010-10-22. 
  97. Torny, Didier (2018). "Pubpeer: vigilante science, journal club or alarm raiser? The controversies over anonymity in post-publication peer review". https://halshs.archives-ouvertes.fr/halshs-01700198/document. 
  98. "Making the most of peer review". eLife 4. November 2015. doi:10.7554/elife.12708. PMID 26559758. 
  99. "Multi-stage open peer review: scientific evaluation integrating the strengths of traditional peer review with the virtues of transparency and self-regulation". Frontiers in Computational Neuroscience 6: 33. 2012. doi:10.3389/fncom.2012.00033. PMID 22783183. 
  100. Jeffrey Marlow (July 23, 2013). "Publish First, Ask Questions Later". Wired. https://www.wired.com/2013/07/publish-first-ask-questions-later/. Retrieved 2015-01-13. 
  101. Elizabeth Allen (September 29, 2017). "The recipe for our (not so) secret Post-Publication Peer Review sauce!". http://blog.scienceopen.com/2014/12/not-so-secret-pppr-recipe/. 
  102. Coy, Peter (2022-01-28). "Opinion | How to Disseminate Science Quickly" (in en-US). The New York Times. ISSN 0362-4331. https://www.nytimes.com/2022/01/28/opinion/peer-review-research-studies.html. 
  103. "F1000Research peer-reviewed articles now visible on PubMed and PubMed Central". STM Publishing News. December 12, 2013. http://www.stm-publishing.com/f1000research-peer-reviewed-articles-now-visible-on-pubmed-and-pubmed-central/. 
  104. Coy, Peter (2022-01-28). "Opinion | How to Disseminate Science Quickly" (in en). The New York Times. https://www.nytimes.com/2022/01/28/opinion/peer-review-research-studies.html. 
  105. Redfield, Rosemary (4 December 2010). "Arsenic-associated bacteria (NASA's claims)". RR Research blog. Retrieved 4 December 2010.
  106. Zimmer, Carl (7 December 2010). "Scientists see fatal flaws in the NASA study of arsenic-based life". Slate. Retrieved 7 December 2010.
  107. Reaves, M. L.; Sinha, S.; Rabinowitz, J. D.; Kruglyak, L.; Redfield, R. J. (July 2012). "Absence of detectable arsenate in DNA from arsenate-grown GFAJ-1 cells". Science 337 (6093): 470–473. doi:10.1126/science.1219861. PMID 22773140. Bibcode2012Sci...337..470R. 
  108. Rosenthal, Robert (1966). "Intentional Error". Experimenter Effects in Behavioral Research. p. 36. 
  109. "Towards a reduction in publication bias". British Medical Journal 295 (6599): 656–659. September 1987. doi:10.1136/bmj.295.6599.656. PMID 3117278. 
  110. Kupfersmid, Joel (1988). "Improving what is published: A model in search of an editor". American Psychologist 43 (8): 635–642. doi:10.1037/0003-066X.43.8.635. 
  111. Glymour, M. Maria; Kawachi, Ichiro (2005). "Review of publication bias in studies on publication bias: Here's a proposal for editors that may help reduce publication bias". BMJ 331 (7517): 638.2. doi:10.1136/bmj.331.7517.638-a. PMID 16166149. 
  112. Smulders, Yvo M. (2013). "A two-step manuscript submission process can reduce publication bias". Journal of Clinical Epidemiology 66 (9): 946–947. doi:10.1016/j.jclinepi.2013.03.023. PMID 23845183. 
  113. Mahoney, Michael J. (1977). "Publication prejudices: An experimental study of confirmatory bias in the peer review system". Cognitive Therapy and Research 1 (2): 161–175. doi:10.1007/BF01173636. 
  114. 114.0 114.1 Wiseman et al 2019, "Registered reports: an early example and analysis"
  115. "Conclusion-Blind Review", 16 January 2007; "Result Blind Review", 6 November 2010; "Who Wants Unbiased Journals?", 27 April 2012
  116. Walster, G. William; Cleary, T. Anne (1970). "A Proposal for a New Editorial Policy in the Social Sciences". The American Statistician 24 (2): 16–19. doi:10.1080/00031305.1970.10478884. 
  117. 117.0 117.1 117.2 Armstrong, J. Scott (1997). "Peer review for journals: Evidence on quality control, fairness, and innovation". Science and Engineering Ethics 3: 63–84. doi:10.1007/s11948-997-0017-3. 
  118. Lawlor, D. A. (2007). "Quality in epidemiological research: Should we be submitting papers before we have the results and submitting more hypothesis-generating research?". International Journal of Epidemiology 36 (5): 940–943. doi:10.1093/ije/dym168. PMID 17875575. 
  119. "Academic reforms: A four-part proposal", Brendan Nyhan, 16 April 2012
  120. "More on pre-accepted academic articles", 27 April 2012
  121. Nyhan, Brendan (2015). "Increasing the Credibility of Political Science Research: A Proposal for Journal Reforms". PS: Political Science & Politics 48: 78–83. doi:10.1017/S1049096515000463. 
  122. "A Proposal for Increasing Evaluation in CS Research Publication", David Karger, 17 February 2011
  123. "It's the incentive structure, people! Why science reform must come from the granting agencies.", Chris Said, 17 April 2012
  124. "Registered reports: a new publishing initiative at Cortex". Cortex; A Journal Devoted to the Study of the Nervous System and Behavior 49 (3): 609–610. March 2013. doi:10.1016/j.cortex.2012.12.016. PMID 23347556. http://orca.cf.ac.uk/45177/1/Chambers_Cortex_2013b_GreenOA.pdf. 
  125. "Read it, understand it, believe it, use it: Principles and proposals for a more credible research publication", Green et al 2013, citing "Protocol Review"
  126. ((The Editors of The Lancet)) (2015). "Protocol review at The Lancet: 1997–2015". The Lancet 386 (10012): 2456–2457. doi:10.1016/S0140-6736(15)01131-9. 
  127. "Promoting reproducibility with registered reports". Nature Human Behaviour 1: 0034. 2017. doi:10.1038/s41562-016-0034. 
  128. "Streamlined review and registered reports soon to be official at EJP". February 6, 2018. https://www.ejp-blog.com/blog/2017/2/3/streamlined-review-and-registered-reports-coming-soon. 
  129. EJP editorial, Johnson 1975 (European Journal of Parapsychology. 1975;1(1):1-2); "Models of control and control of bias", Johnson 1975 (European Journal of Parapsychology. 1975;1(1):36–44); "On Publication Policy Regarding Non-Significant Results", Johnson 1976 (European Journal of Parapsychology 1976;1(2):1–5)
  130. Armstrong, J. Scott; Dagum, Estella Bee; Fildes, Robert; Makridakis, Spyros (1986). "Publishing Standards for Research on Forecasting (editorial)". Marketing Papers. https://repository.upenn.edu/marketing_papers/86. 
  131. Armstrong, J.Scott (1996). "Publication of research on controversial topics: The early acceptance procedure". International Journal of Forecasting 12 (2): 299–302. doi:10.1016/0169-2070(95)00626-5. https://repository.upenn.edu/marketing_papers/106. 
  132. Weiss, David J. (1989). "An Experiment in Publication: Advance Publication Review". Applied Psychological Measurement 13: 1–7. doi:10.1177/014662168901300101. 
  133. Sridharan, Lakshmi; Greenland, Philip (2009). "Editorial Policies and Publication Bias". Archives of Internal Medicine 169 (11): 1022–1023. doi:10.1001/archinternmed.2009.100. PMID 19506169. 
  134. "Registered Reports", OSF
  135. "Registered Reports: A step change in scientific publishing; Professor Chris Chambers, Registered Reports Editor of the Elsevier journal Cortex and one of the concept's founders, on how the initiative combats publication bias", Chambers, 13 November 2014
  136. d. Chambers, Christopher; Feredoes, Eva; d. Muthukumaraswamy, Suresh; j. Etchells, Peter (2014). "Instead of "playing the game" it is time to change the rules: Registered Reports at AIMS Neuroscience and beyond". AIMS Neuroscience 1: 4–17. doi:10.3934/Neuroscience.2014.1.4. 
  137. Nosek, Brian A.; Lakens, Daniël (2014). "Registered Reports". Social Psychology 45 (3): 137–141. doi:10.1027/1864-9335/a000192. https://zenodo.org/record/235813. 
  138. Pain, Elisabeth (15 December 2015). "Register your study as a new publication option". Science. doi:10.1126/science.caredit.a1500282. 
  139. "Psychology's 'registration revolution': Moves to uphold transparency are not only making psychology more scientific – they are harnessing our knowledge of the mind to strengthen science", Guardian, 20 May 2014
  140. Findley, Michael G.; Jensen, Nathan M.; Malesky, Edmund J.; Pepinsky, Thomas B. (2016). "Can Results-Free Review Reduce Publication Bias? The Results and Implications of a Pilot Study". Comparative Political Studies 49 (13): 1667–1703. doi:10.1177/0010414016655539. 
  141. Funtowicz, S (6 December 2001). "Peer review and quality control". in Smelser, Neil J; Baltes, Paul B. International Encyclopaedia of the Social and Behavioural Sciences. pp. 11179–11183. 
  142. Rennie, Drummond; Flanagin, Annette; Smith, Richard; Smith, Jane (2003). "Fifth International Congress on Peer Review and Biomedical Publication". JAMA 289 (11): 1438. doi:10.1001/jama.289.11.1438. 
  143. "Genetically modified food: consternation, confusion, and crack-up". The Medical Journal of Australia 172 (4): 148–149. February 2000. doi:10.5694/j.1326-5377.2000.tb125533.x. PMID 10772580. http://www.mja.com.au/public/issues/172_04_210200/horton/horton.html. 
  144. Jefferson, Tom; Rudin, Melanie; Brodney Folse, Suzanne; Davidoff, Frank (2007-04-18). Cochrane Methodology Review Group. ed. "Editorial peer review for improving the quality of reports of biomedical studies" (in en). Cochrane Database of Systematic Reviews 2010 (2): MR000016. doi:10.1002/14651858.MR000016.pub3. PMID 17443635. 
  145. Bruce, Rachel; Chauvin, Anthony; Trinquart, Ludovic; Ravaud, Philippe; Boutron, Isabelle (December 2016). "Impact of interventions to improve the quality of peer review of biomedical journals: a systematic review and meta-analysis" (in en). BMC Medicine 14 (1): 85. doi:10.1186/s12916-016-0631-5. ISSN 1741-7015. PMID 27287500. 
  146. Couchman, John R. (November 11, 2013). "Peer Review and Reproducibility. Crisis or Time for Course Correction?". Journal of Histochemistry & Cytochemistry 62 (1): 9–10. doi:10.1369/0022155413513462. PMID 24217925. 
  147. "The peer review drugs don't work" (in en). Times Higher Education. 2015-05-28. https://www.timeshighereducation.com/content/the-peer-review-drugs-dont-work. 
  148. "Peer review 'works against' early career researchers" (in en). Times Higher Education (THE). 2018-07-16. https://www.timeshighereducation.com/news/peer-review-works-against-early-career-researchers. 
  149. "Peer-review crisis creates problems for journals and scholars" (in en). https://www.insidehighered.com/news/2022/06/13/peer-review-crisis-creates-problems-journals-and-scholars. 
  150. Ioannidis, John P. A. (2005). "Why Most Published Research Findings Are False". PLOS Medicine 2 (8): e124. doi:10.1371/journal.pmed.0020124. PMID 16060722. 
  151. 151.0 151.1 151.2 Csiszar, Alex (2016). "Peer Review: Troubled from the Start". Nature 532 (7599): 306–308. doi:10.1038/532306a. PMID 27111616. Bibcode2016Natur.532..306C. 
  152. 152.0 152.1 Moxham, Noah; Fyfe, Aileen (2018). "The Royal Society and the Prehistory of Peer Review, 1665–1965". The Historical Journal 61 (4): 863–889. doi:10.1017/S0018246X17000334. https://kar.kent.ac.uk/65042/3/Peer%20review%20v30%20AAM%20SUBMTD.pdf. 
  153. Moore, John (2006). "Does Peer Review Mean the Same to the Public as It Does to Scientists?". peertopeer Peer review blog. https://blogs.nature.com/peer-to-peer/2006/06/perspective_does_peer_review_m.html. 
  154. Ferguson, Cat; Marcus, Adam; Oransky, Ivan (2014). "Publishing: The Peer-Review Scam". Nature 515 (7528): 480–482. doi:10.1038/515480a. PMID 25428481. Bibcode2014Natur.515..480F. 
  155. 155.0 155.1 Budd, J. M.; Sievert, M.; Schultz, T. R. (1998). "Phenomena of Retraction: Reasons for Retraction and Citations to the Publications". JAMA 280 (3): 296–297. doi:10.1001/jama.280.3.296. PMID 9676689. 
  156. 156.0 156.1 Smith, Richard (2006). "Peer Review: A Flawed Process at the Heart of Science and Journals". Journal of the Royal Society of Medicine 99 (4): 178–182. doi:10.1177/014107680609900414. PMID 16574968. 
  157. Ross-Hellauer, Tony (2017). "What Is Open Peer Review? A Systematic Review". F1000Research 6: 588. doi:10.12688/f1000research.11369.2. PMID 28580134. 
  158. 158.0 158.1 Tennant, Jonathan P.; Dugan, Jonathan M.; Graziotin, Daniel; Jacques, Damien C.; Waldner, François; Mietchen, Daniel; Elkhatib, Yehia; b. Collister, Lauren et al. (2017). "A Multi-Disciplinary Perspective on Emergent and Future Innovations in Peer Review". F1000Research 6: 1151. doi:10.12688/f1000research.12037.3. PMID 29188015. 
  159. "Skeptics get a journal". http://www.realclimate.org/docs/thacker/skeptics.pdf. , Paul Thacker, 2005.
  160. Schwarz, Alan; Bogdanich, Walt; Williams, Jacqueline (March 24, 2016). "N.F.L.'s Flawed Concussion Research and Ties to Tobacco Industry". The New York Times. https://www.nytimes.com/2016/03/25/sports/football/nfl-concussion-research-tobacco.html. .
  161. Wong, Victoria S. S.; Avalos, Lauro Nathaniel; Callaham, Michael L. (2019). "Industry Payments to Physician Journal Editors". PLOS ONE 14 (2): e0211495. doi:10.1371/journal.pone.0211495. PMID 30730904. Bibcode2019PLoSO..1411495W. 
  162. Weiss, Glen J.; Davis, Roger B. (2019). "Discordant Financial Conflicts of Interest Disclosures between Clinical Trial Conference Abstract and Subsequent Publication". PeerJ 7: e6423. doi:10.7717/peerj.6423. PMID 30775185. 
  163. 163.0 163.1 Jarvis, Suzanne (September 2017). "The perils of peer review". Veterinary Record 181 (12): 303–304. doi:10.1136/vr.j4362. 
  164. Flaherty, D. K. (2013). "Ghost- and Guest-Authored Pharmaceutical Industry–Sponsored Studies: Abuse of Academic Integrity, the Peer Review System, and Public Trust". The Annals of Pharmacotherapy 47 (7–8): 1081–1083. doi:10.1345/aph.1R691. PMID 23585648. 
  165. "Ghostwriting in medical literature". https://www.grassley.senate.gov/sites/default/files/about/upload/Senator-Grassley-Report.pdf. .
  166. "Frequently asked questions about medical ghostwriting". June 28, 2011. https://www.pogo.org/analysis/2011/06/frequently-asked-questions-about-medical-ghostwriting/. .
  167. 167.0 167.1 167.2 167.3 167.4 167.5 167.6 167.7 Vanholsbeeck, Marc; Thacker, Paul; Sattler, Susanne; Ross-Hellauer, Tony; Rivera-López, Bárbara S.; Rice, Curt; Nobes, Andy; Masuzzo, Paola et al. (2019-03-11). "Ten Hot Topics around Scholarly Publishing". Publications 7 (2): 34. doi:10.3390/publications7020034. 
  168. Relman, A. S. (1990). "Peer Review in Scientific Journals--What Good Is It?". Western Journal of Medicine 153 (5): 520–522. PMID 2260288. 
  169. Bravo, Giangiacomo; Grimaldo, Francisco; López-Iñesta, Emilia; Mehmani, Bahar; Squazzoni, Flaminio (2019). "The Effect of Publishing Peer Review Reports on Referee Behavior in Five Scholarly Journals". Nature Communications 10 (1): 322. doi:10.1038/s41467-018-08250-2. PMID 30659186. Bibcode2019NatCo..10..322B. 
  170. 170.0 170.1 Tennant, Jonathan P. (2018). "The State of the Art in Peer Review". FEMS Microbiology Letters 365 (19). doi:10.1093/femsle/fny204. PMID 30137294. 
  171. Squazzoni, Flaminio; Grimaldo, Francisco; Marušić, Ana (2017). "Publishing: Journals Could Share Peer-Review Data". Nature 546 (7658): 352. doi:10.1038/546352a. PMID 28617464. Bibcode2017Natur.546Q.352S. 
  172. Allen, Heidi; Boxer, Emma; Cury, Alexandra; Gaston, Thomas; Graf, Chris; Hogan, Ben; Loh, Stephanie; Wakley, Hannah et al. (2018), What Does Better Peer Review Look like? Definitions, Essential Areas, and Recommendations for Better Practice, doi:10.17605/OSF.IO/4MFK2 
  173. Fang, Ferric C.; Casadevall, Arturo (2011). "Retracted Science and the Retraction Index". Infection and Immunity 79 (10): 3855–3859. doi:10.1128/IAI.05661-11. PMID 21825063. 
  174. Moylan, Elizabeth C.; Kowalczuk, Maria K. (2016). "Why Articles Are Retracted: A Retrospective Cross-Sectional Study of Retraction Notices at BioMed Central". BMJ Open 6 (11): e012047. doi:10.1136/bmjopen-2016-012047. PMID 27881524. 
  175. Open Science Collaboration (2015). "Estimating the Reproducibility of Psychological Science". Science 349 (6251): aac4716. doi:10.1126/science.aac4716. PMID 26315443. https://ink.library.smu.edu.sg/lkcsb_research/5257. 
  176. 176.0 176.1 Munafò, Marcus R.; Nosek, Brian A.; Bishop, Dorothy V. M.; Button, Katherine S.; Chambers, Christopher D.; Percie Du Sert, Nathalie; Simonsohn, Uri; Wagenmakers, Eric-Jan et al. (2017). "A Manifesto for Reproducible Science". Nature Human Behaviour 1: 0021. doi:10.1038/s41562-016-0021. PMID 33954258. PMC 7610724. http://dare.uva.nl/personal/pure/en/publications/a-manifesto-for-reproducible-science(3534b98f-a374-496b-9ad1-e61539477d66).html. 
  177. Fanelli, Daniele (2018). "Opinion: Is Science Really Facing a Reproducibility Crisis, and Do We Need It To?". Proceedings of the National Academy of Sciences 115 (11): 2628–2631. doi:10.1073/pnas.1708272114. PMID 29531051. 
  178. Goodman, Steven N. (1994). "Manuscript Quality before and after Peer Review and Editing at Annals of Internal Medicine". Annals of Internal Medicine 121 (1): 11–21. doi:10.7326/0003-4819-121-1-199407010-00003. PMID 8198342. 
  179. Pierson, Charon A. (2018). "Peer review and journal quality". Journal of the American Association of Nurse Practitioners 30 (1): 1–2. doi:10.1097/JXX.0000000000000018. PMID 29757914. 
  180. Caputo, Richard K. (2019). "Peer Review: A Vital Gatekeeping Function and Obligation of Professional Scholarly Practice". Families in Society: The Journal of Contemporary Social Services 100: 6–16. doi:10.1177/1044389418808155. 
  181. Siler, Kyle; Lee, Kirby; Bero, Lisa (2015). "Measuring the effectiveness of scientific gatekeeping". Proceedings of the National Academy of Sciences 112 (2): 360–365. doi:10.1073/pnas.1418218112. PMID 25535380. Bibcode2015PNAS..112..360S. 
  182. Resnik, David B.; Elmore, Susan A. (2016). "Ensuring the Quality, Fairness, and Integrity of Journal Peer Review: A Possible Role of Editors". Science and Engineering Ethics 22 (1): 169–188. doi:10.1007/s11948-015-9625-5. PMID 25633924. 
  183. Bornmann, Lutz (2011). "Scientific Peer Review". Annual Review of Information Science and Technology 45: 197–245. doi:10.1002/aris.2011.1440450112. 
  184. "Cargo cult science". http://calteches.library.caltech.edu/51/2/CargoCult.htm. , Richard Feynman.
  185. "Cargo Cult Science". Caltech Magazine. 1974. http://Calteches.library.caltech.edu/51/2/CargoCult.HTM. 
  186. Carroll, Aaron E. (November 5, 2018). "Peer Review: The Worst Way to Judge Research, Except for All the Others". The New York Times. https://www.nytimes.com/2018/11/05/upshot/peer-review-the-worst-way-to-judge-research-except-for-all-the-others.html. , Aaron E. Carroll, New York Times.
  187. "Bucking the Big Bang". https://www.newscientist.com/article/mg18224482-900-bucking-the-big-bang/. , Eric Lerner, New Scientist.
  188. 188.0 188.1 "Untangling Academic Publishing. A History of the Relationship between Commercial Interests, Academic Prestige and the Circulation of Research.". 26. 
  189. Priem, Jason; Hemminger, Bradley M. (2012). "Decoupling the Scholarly Journal". Frontiers in Computational Neuroscience 6: 19. doi:10.3389/fncom.2012.00019. PMID 22493574. 
  190. Bowman, Nicholas David; Keene, Justin Robert (2018). "A Layered Framework for Considering Open Science Practices". Communication Research Reports 35 (4): 363–372. doi:10.1080/08824096.2018.1513273. 
  191. McKiernan, E. C.; Bourne, P. E.; Brown, C. T.; Buck, S.; Kenall, A.; Lin, J.; McDougall, D.; Nosek, B. A. et al. (2016). "Point of View: How Open Science Helps Researchers Succeed". eLife 5. doi:10.7554/eLife.16800. PMID 27387362. 
  192. 192.0 192.1 "In peer review we (don't) trust: How peer review's filtering poses a systemic risk to science". https://researchers.one/article/2018-09-17. 
  193. Brembs, Björn (2019). "Reliable Novelty: New Should Not Trump True". PLOS Biology 17 (2): e3000117. doi:10.1371/journal.pbio.3000117. PMID 30753184. 
  194. Stern, Bodo M.; o'Shea, Erin K. (2019). "A Proposal for the Future of Scientific Publishing in the Life Sciences". PLOS Biology 17 (2): e3000116. doi:10.1371/journal.pbio.3000116. PMID 30753179. 
  195. Bradley, James V. (1981). "Pernicious Publication Practices". Bulletin of the Psychonomic Society 18: 31–34. doi:10.3758/bf03333562. 
  196. "British scientists exclude 'maverick' colleagues, says report" (2004) EurekAlert Public release date: August 16, 2004
  197. Boudreau, K. J.; Guinan, E.C.; Lakhani, K.R.; Riedl, C. (October 2016). "Looking Across and Looking Beyond the Knowledge Frontier: Intellectual Distance, Novelty, and Resource Allocation in Science". Management Science 62 (10): 2765–2783. doi:10.1287/mnsc.2015.2285. PMID 27746512. 
  198. van den Besselaar, Peter; Sandström, Ulf; Schiffbaenker, Hélène (October 2018). "Studying grant decision-making: a linguistic analysis of review reports". Scientometrics 117 (1): 313–329. doi:10.1007/s11192-018-2848-x. ISSN 0138-9130. PMID 30220747. 
  199. "Malice's wonderland: research funding and peer review". Journal of Neurobiology 14 (2): 95–112. March 1983. doi:10.1002/neu.480140202. PMID 6842193. http://post.queensu.ca/~forsdyke/peerrev1.htm. Retrieved June 24, 2015. "... they may strongly resist a rival's hypothesis that challenges their own". 
  200. Grimaldo, Francisco; Paolucci, Mario (14 March 2013). "A simulation of disagreement for control of rational cheating in peer review". Advances in Complex Systems 16 (7): 1350004. doi:10.1142/S0219525913500045. Bibcode2005AdCS....8...15L. 
  201. Petit-Zeman, Sophie (January 16, 2003). "Trial by peers comes up short". The Guardian. https://www.theguardian.com/print/0,3858,4583809-111019,00.html. 
  202. Fang, H. (2011). "Peer review and over-competitive research funding fostering mainstream opinion to monopoly". Scientometrics 87 (2): 293–301. doi:10.1007/s11192-010-0323-4. 
  203. Rowland, Fytton (2002). "The peer-review process". Learned Publishing 15 (4): 247–258. doi:10.1087/095315102760319206. 
  204. Budden, A. E.; Tregenza, T.; Aarssen, L. W.; Koricheva, J.; Leimu, R.; Lortie, C. J. (January 2008). "Double-blind review favours increased representation of female authors". Trends in Ecology & Evolution 23 (1): 4–6. doi:10.1016/j.tree.2007.07.008. PMID 17963996. 
  205. Squazzoni, Flaminio; Bravo, Giangiacomo; Farjam, Mike; Marusic, Ana; Mehmani, Bahar; Willis, Michael; Birukou, Aliaksandr; Dondio, Pierpaolo et al. (2021-01-06). "Peer review and gender bias: A study on 145 scholarly journals". Science Advances 7 (2): eabd0299. doi:10.1126/sciadv.abd0299. ISSN 2375-2548. PMID 33523967. Bibcode2021SciA....7..299S. 
  206. Akst, Jef (2020-01-06). "No Gender Bias in Peer Review: Study" (in en). https://www.the-scientist.com/news-opinion/no-gender-bias-in-peer-review-study-68320. 
  207. Yirka, Bob (2020-01-07). "Analysis: Peer review process unlikely to be primary cause of gender publishing inequalities in scholarly journals" (in en). https://phys.org/news/2021-01-analysis-peer-primary-gender-publishing.html. 
  208. 208.0 208.1 Brainard, Jeffrey (1 March 2021). "The $450 question: Should journals pay peer reviewers?". Science. doi:10.1126/science.abh3175. 
  209. "Advances.in". October 9, 2022. http://advances.in. 
  210. Suber, Peter (September 2, 2007). "Will open access undermine peer review?". SPARC Open Access Newsletter. https://dash.harvard.edu/bitstream/handle/1/4322578/suber_peer.html. 
  211. Björk, Bo-Christer; Solomon, David (2012). "Open access versus subscription journals: A comparison of scientific impact". BMC Medicine 10: 73. doi:10.1186/1741-7015-10-73. PMID 22805105. 
  212. Afifi, M. "Reviewing the "Letter-to-editor" section in the Bulletin of the World Health Organization, 2000–2004". Bulletin of the World Health Organization. https://www.who.int/bulletin/bulletin_board/84/letters/en/index.html. 
  213. Lee, Kirby (2006). "Increasing accountability". Nature. doi:10.1038/nature05007. 
  214. Baxt, W. G.; Waeckerle, J. F.; Berlin, J. A.; Callaham, M. L. (September 1998). "Who reviews the reviewers? Feasibility of using a fictitious manuscript to evaluate peer reviewer performance". Annals of Emergency Medicine 32 (3 Pt 1): 310–317. doi:10.1016/S0196-0644(98)70006-X. PMID 9737492. 
  215. 215.0 215.1 A case of peer review fraud in Tumor Biology papers (Retrieved April 25, 2017)
  216. Ferguson, C.; Marcus, A.; Oransky, I. (November 2014). "Publishing: The peer-review scam". Nature 515 (7528): 480–482. doi:10.1038/515480a. PMID 25428481. Bibcode2014Natur.515..480F. 
  217. "COPE statement on inappropriate manipulation of peer review processes". publicationethics.org. December 19, 2014. http://publicationethics.org/news/cope-statement-inappropriate-manipulation-peer-review-processes. 
  218. "Inappropriate manipulation of peer review". BioMed Central blog. 2015-03-26. http://blogs.biomedcentral.com/bmcblog/2015/03/26/manipulation-peer-review/. 
  219. Callaway, Ewen (2015). "Faked peer reviews prompt 64 retractions". Nature. doi:10.1038/nature.2015.18202. 
  220. 220.0 220.1 "Imposters hijack journal's peer review process to publish substandard papers". 2021-01-18. https://www.chemistryworld.com/news/imposters-hijack-journals-peer-review-process-to-publish-substandard-papers/4013050.article. 
  221. Pinna, Nicola; Clavel, Guylhaine; Roco, Mihail C. (2020). "The Journal of Nanoparticle Research victim of an organized rogue editor network!". Journal of Nanoparticle Research 22 (12): 376. doi:10.1007/s11051-020-05094-0. ISSN 1388-0764. Bibcode2020JNR....22..376P. 
  222. "Historians on the Hot Seat". April 23, 2010. https://historynewsnetwork.org/article/1081. 
  223. 223.0 223.1 Weiss, Rick (June 9, 2005). "Many scientists admit to misconduct: Degrees of deception vary in poll". Washington Post. https://www.washingtonpost.com/wp-dyn/content/article/2005/06/08/AR2005060802385.html. 
  224. Michaels, David (2006). "Politicizing Peer Review: Scientific Perspective". in Wagner, Wendy; Steinzor, Rena. Rescuing Science from Politics: Regulation and the Distortion of Scientific Research. Cambridge University Press. p. 224. ISBN 978-0521855204. https://books.google.com/books?id=VZ1axrolCpkC&pg=PA224. 
  225. Soon, W.; Baliunas, S. (2003). "Proxy climatic and environmental changes of the past 1000 years". Climate Research 23: 89–110. doi:10.3354/cr023089. Bibcode2003ClRes..23...89S. 
  226. "A mathematical model for the determination of total area under glucose tolerance and other metabolic curves". Diabetes Care 17 (2): 152–154. February 1994. doi:10.2337/diacare.17.2.152. PMID 8137688. 
  227. Knapp, Alex (2011). "Apparently, Calculus Was Invented In 1994". Forbes. https://www.forbes.com/sites/alexknapp/2011/11/10/apparently-calculus-was-invented-in-1994/. 
  228. Purgathofer, Werner. "Beware of VIDEA!". Technical University of Vienna. http://www.cg.tuwien.ac.at/~wp/videa.html. 
  229. Dougherty, M. V.; Harsting, P.; Friedman, R. (2009). "40 Cases of Plagiarism". Bulletin de Philosophie Médiévale 51: 350–391. http://www.nnrh.dk/NNRH-hp/40.Cases.of.Plagiarism.pdf. 
  230. Dougherty, M. V. (2017). "Correcting the Scholarly Record in the Aftermath of Plagiarism: A Snapshot of Current-Day Publishing Practices in Philosophy". Metaphilosophy 48 (3): 258–283. doi:10.1111/meta.12241. 
  231. Holliday, Vance T.; Daulton, Tyrone L.; Bartlein, Patrick J.; Boslough, Mark B.; Breslawski, Ryan P.; Fisher, Abigail E.; Jorgeson, Ian A.; Scott, Andrew C. et al. (2023-07-26). "Comprehensive refutation of the Younger Dryas Impact Hypothesis (YDIH)" (in en). Earth-Science Reviews 247: 104502. doi:10.1016/j.earscirev.2023.104502. https://linkinghub.elsevier.com/retrieve/pii/S0012825223001915. 
  232. Schiermeier, Quirin (2017). "Monument to peer review unveiled in Moscow". Nature. doi:10.1038/nature.2017.22060. 

Further reading



Government policy

The European Union has been using peer review in the "Open Method of Co-ordination" of policies in the fields of active labour market policy since 1999.[1] In 2004, a program of peer reviews started in social inclusion.[2] Each program sponsors about eight peer review meetings in each year, in which a "host country" lays a given policy or initiative open to examination by half a dozen other countries and the relevant European-level NGOs. These usually meet over two days and include visits to local sites where the policy can be seen in operation. The meeting is preceded by the compilation of an expert report on which participating "peer countries" submit comments. The results are published on the web.

The United Nations Economic Commission for Europe, through UNECE Environmental Performance Reviews, uses peer review, referred to as "peer learning", to evaluate progress made by its member countries in improving their environmental policies.

The State of California is the only U.S. state to mandate scientific peer review. In 1997, the Governor of California signed into law Senate Bill 1320 (Sher), Chapter 295, statutes of 1997, which mandates that, before any CalEPA Board, Department, or Office adopts a final version of a rule-making, the scientific findings, conclusions, and assumptions on which the proposed rule are based must be submitted for independent external scientific peer review. This requirement is incorporated into the California Health and Safety Code Section 57004.[3]

Medical

Main page: Medicine:Clinical peer review

Medical peer review may be distinguished in 4 classifications:[4]

1) Clinical peer review; Clinical peer review is a procedure for assessing a patient's involvement with experiences of care. It is a piece of progressing proficient practice assessment and centered proficient practice assessment—significant supporters of supplier credentialing and privileging.[5]

2) Peer evaluation of clinical teaching skills for both physicians and nurses;[6][7]

3) Scientific peer review of journal articles;

4) A secondary round of peer review for the clinical value of articles concurrently published in medical journals.[8]

Additionally, "medical peer review" has been used by the American Medical Association to refer not only to the process of improving quality and safety in health care organizations, but also to the process of rating clinical behavior or compliance with professional society membership standards.[9][10] The clinical network believes it to be the most ideal method of guaranteeing that distributed exploration is dependable and that any clinical medicines that it advocates are protected and viable for individuals. Thus, the terminology has poor standardization and specificity, particularly as a database search term.[11]

Technical

Main page: Engineering:Technical peer review

In engineering, technical peer review is a type of engineering review. Technical peer reviews are a well defined review process for finding and fixing defects, conducted by a team of peers with assigned roles. Technical peer reviews are carried out by peers representing areas of life cycle affected by material being reviewed (usually limited to 6 or fewer people). Technical peer reviews are held within development phases, between milestone reviews, on completed products or completed portions of products.[12]

Criticism

To an outsider, the anonymous, pre-publication peer review process is opaque. Certain journals are accused of not carrying out stringent peer review in order to more easily expand their customer base, particularly in journals where authors pay a fee before publication.[13] Richard Smith, MD, former editor of the British Medical Journal, has claimed that peer review is "ineffective, largely a lottery, anti-innovatory, slow, expensive, wasteful of scientific time, inefficient, easily abused, prone to bias, unable to detect fraud and irrelevant; Several studies have shown that peer review is biased against the provincial and those from low- and middle-income countries; Many journals take months and even years to publish and the process wastes researchers' time. As for the cost, the Research Information Network estimated the global cost of peer review at £1.9 billion in 2008."[14]

In addition, Australia's Innovative Research Universities group (a coalition of seven comprehensive universities committed to inclusive excellence in teaching, learning and research in Australia) has found that "peer review disadvantages researchers in their early careers, when they rely on competitive grants to cover their salaries, and when unsuccessful funding applications often mark the end of a research idea".[15]

Low-end distinctions in articles understandable to all peers

John Ioannidis argues that since the exams and other tests that people pass on their way from "layman" to "expert" focus on answering the questions in time and in accordance with a list of answers, and not on making precise distinctions (the latter of which would be unrecognizable to experts of lower cognitive precision), there is as much individual variation in the ability to distinguish causation from correlation among "experts" as there is among "laymen". Ioannidis argues that as a result, scholarly peer review by many "experts" allows only articles that are understandable at a wide range of cognitive precision levels including very low ones to pass, biasing publications towards favoring articles that infer causation from correlation while mislabelling articles that make the distinction as "incompetent overestimation of one's ability" on the side of the authors because some of the reviewing "experts" are cognitively unable to distinguish the distinction from alleged rationalization of specific conclusions. It is argued by Ioannidis that this makes peer review a cause of selective publication of false research findings while stopping publication of rigorous criticism thereof, and that further post-publication review repeats the same bias by selectively retracting the few rigorous articles that may have made it through initial pre-publication peer review while letting the low-end ones that confuse correlation and causation remain in print.[16]

Peer review and trust

Researchers have peer reviewed manuscripts prior to publishing them in a variety of ways since the 18th century.[17][18] The main goal of this practice is to improve the relevance and accuracy of scientific discussions. Even though experts often criticize peer review for a number of reasons, the process is still often considered the "gold standard" of science.[19] Occasionally however, peer review approves studies that are later found to be wrong and rarely deceptive or fraudulent results are discovered prior to publication.[20][21] Thus, there seems to be an element of discord between the ideology behind and the practice of peer review. By failing to effectively communicate that peer review is imperfect, the message conveyed to the wider public is that studies published in peer-reviewed journals are "true" and that peer review protects the literature from flawed science. A number of well-established criticisms exist of many elements of peer review.[22][23][24] In the following we describe cases of the wider impact inappropriate peer review can have on public understanding of scientific literature.

Multiple examples across several areas of science find that scientists elevated the importance of peer review for research that was questionable or corrupted. For example, climate change deniers have published studies in the Energy and Environment journal, attempting to undermine the body of research that shows how human activity impacts the Earth's climate. Politicians in the United States who reject the established science of climate change have then cited this journal on several occasions in speeches and reports.[note 1]

At times, peer review has been exposed as a process that was orchestrated for a preconceived outcome. The New York Times gained access to confidential peer review documents for studies sponsored by the National Football League (NFL) that were cited as scientific evidence that brain injuries do not cause long-term harm to its players.[note 2] During the peer review process, the authors of the study stated that all NFL players were part of a study, a claim that the reporters found to be false by examining the database used for the research. Furthermore, The Times noted that the NFL sought to legitimize the studies" methods and conclusion by citing a "rigorous, confidential peer-review process" despite evidence that some peer reviewers seemed "desperate" to stop their publication. Recent research has also demonstrated that widespread industry funding for published medical research often goes undeclared and that such conflicts of interest are not appropriately addressed by peer review.[25][26]

Another problem that peer review fails to catch is ghostwriting, a process by which companies draft articles for academics who then publish them in journals, sometimes with little or no changes.[27] These studies can then be used for political, regulatory and marketing purposes. In 2010, the US Senate Finance Committee released a report that found this practice was widespread, that it corrupted the scientific literature and increased prescription rates.[note 3] Ghostwritten articles have appeared in dozens of journals, involving professors at several universities.[note 4]

Just as experts in a particular field have a better understanding of the value of papers published in their area, scientists are considered to have better grasp of the value of published papers than the general public and to see peer review as a human process, with human failings,[28] and that "despite its limitations, we need it. It is all we have, and it is hard to imagine how we would get along without it".[29] But these subtleties are lost on the general public, who are often misled into thinking that published in a journal with peer review is the "gold standard" and can erroneously equate published research with the truth.[28] Thus, more care must be taken over how peer review, and the results of peer reviewed research, are communicated to non-specialist audiences; particularly during a time in which a range of technical changes and a deeper appreciation of the complexities of peer review are emerging.[30][31][32][33] This will be needed as the scholarly publishing system has to confront wider issues such as retractions[21][34][35] and replication or reproducibility "crisis'.[36][37][38]

Views of peer review

Peer review is often considered integral to scientific discourse in one form or another. Its gatekeeping role is supposed to be necessary to maintain the quality of the scientific literature[39][40] and avoid a risk of unreliable results, inability to separate signal from noise, and slow scientific progress.[41][42]

Shortcomings of peer review have been met with calls for even stronger filtering and more gatekeeping. A common argument in favor of such initiatives is the belief that this filter is needed to maintain the integrity of the scientific literature.[43][44]

Calls for more oversight have at least two implications that are counterintuitive of what is known to be true scholarship.[28]

  1. The belief that scholars are incapable of evaluating the quality of work on their own, that they are in need of a gatekeeper to inform them of what is good and what is not.
  2. The belief that scholars need a "guardian" to make sure they are doing good work.

Others argue[28] that authors most of all have a vested interest in the quality of a particular piece of work. Only the authors could have, as Feynman (1974)[note 5] puts it, the "extra type of integrity that is beyond not lying, but bending over backwards to show how you're maybe wrong, that you ought to have when acting as a scientist." If anything, the current peer review process and academic system could penalize, or at least fail to incentivize, such integrity.

Instead, the credibility conferred by the "peer-reviewed" label could diminish what Feynman calls the culture of doubt necessary for science to operate a self-correcting, truth-seeking process.[45] The effects of this can be seen in the ongoing replication crisis, hoaxes, and widespread outrage over the inefficacy of the current system.[22][17] It's common to think that more oversight is the answer, as peer reviewers are not at all lacking in skepticism. But the issue is not the skepticism shared by the select few who determine whether an article passes through the filter. It is the validation, and accompanying lack of skepticism, that comes afterwards.[note 6] Here again more oversight only adds to the impression that peer review ensures quality, thereby further diminishing the culture of doubt and counteracting the spirit of scientific inquiry.[note 7]

Quality research - even some of our most fundamental scientific discoveries - dates back centuries, long before peer review took its current form.[17][46][18] Whatever peer review existed centuries ago, it took a different form than it does in modern times, without the influence of large, commercial publishing companies or a pervasive culture of publish or perish.[46] Though in its initial conception it was often a laborious and time-consuming task, researchers took peer review on nonetheless, not out of obligation but out of duty to uphold the integrity of their own scholarship. They managed to do so, for the most part, without the aid of centralised journals, editors, or any formalised or institutionalised process whatsoever. Supporters of modern technology argue[28] that it makes it possible to communicate instantaneously with scholars around the globe, make such scholarly exchanges easier, and restore peer review to a purer scholarly form, as a discourse in which researchers engage with one another to better clarify, understand, and communicate their insights.[31][47]

Such modern technology includes posting results to preprint servers, preregistration of studies, open peer review, and other open science practices.[37][48][49] In all these initiatives, the role of gatekeeping remains prominent, as if a necessary feature of all scholarly communication, but critics argue[24] that a proper, real-world implementation could test and disprove this assumption; demonstrate researchers' desire for more that traditional journals can offer; show that researchers can be entrusted to perform their own quality control independent of journal-coupled review. Jon Tennant also argues that the outcry over the inefficiencies of traditional journals centers on their inability to provide rigorous enough scrutiny, and the outsourcing of critical thinking to a concealed and poorly-understood process. Thus, the assumption that journals and peer review are required to protect scientific integrity seems to undermine the very foundations of scholarly inquiry.[28]

To test the hypothesis that filtering is indeed unnecessary to quality control, many of the traditional publication practices would need to be redesigned, editorial boards repurposed if not disbanded, and authors granted control over the peer review of their own work. Putting authors in charge of their own peer review is seen as serving a dual purpose.[28] On one hand, it removes the conferral of quality within the traditional system, thus eliminating the prestige associated with the simple act of publishing. Perhaps paradoxically, the removal of this barrier might actually result in an increase of the quality of published work, as it eliminates the cachet of publishing for its own sake. On the other hand, readers know that there is no filter so they must interpret anything they read with a healthy dose of skepticism, thereby naturally restoring the culture of doubt to scientific practice.[50][51][52]

In addition to concerns about the quality of work produced by well-meaning researchers, there are concerns that a truly open system would allow the literature to be populated with junk and propaganda by those with a vested interest in certain issues. A counterargument is that the conventional model of peer review diminishes the healthy skepticism that is a hallmark of scientific inquiry, and thus confers credibility upon subversive attempts to infiltrate the literature.[28] Allowing such "junk" to be published could make individual articles less reliable but render the overall literature more robust by fostering a "culture of doubt".[50]

See also

Notes

References

  1. "Mutual Learning Programme - Employment, Social Affairs & Inclusion - European Commission". http://ec.europa.eu/social/mlp. 
  2. "Social Peer to Peer – Online Casino Reviews". http://www.peer-review-social-inclusion.eu. 
  3. "What is Scientific Peer Review?" (in en-US). https://ceparev.berkeley.edu/what-is-peer-review/. 
  4. "REVIEW BY PEERS". A Guide for Professional, Clinical and Administrative Processes. https://www.safetyandquality.gov.au/sites/default/files/migrated/37358-Review-by-Peers1.pdf. 
  5. Deyo-Svendsen, Mark E.; Phillips, Michael R.; Albright, Jill K.; Schilling, Keith A.; Palmer, Karl B. (October/December 2016). "A Systematic Approach to Clinical Peer Review in a Critical Access Hospital" (in en-US). Quality Management in Healthcare 25 (4): 213–218. doi:10.1097/QMH.0000000000000113. ISSN 1063-8628. PMID 27749718. 
  6. Medschool.ucsf.edu
  7. "Documenting the scholarship of clinical teaching through peer review". Nurse Educator 23 (6): 17–20. November–December 1998. doi:10.1097/00006223-199811000-00008. PMID 9934106. 
  8. "Second-order peer review of the medical literature for clinical practitioners". JAMA 295 (15): 1801–8. 2006. doi:10.1001/jama.295.15.1801. PMID 16622142. 
  9. Snelson, Elizabeth A. (2010). Physician's Guide to Medical Staff Organization Bylaws. p. 131. http://www.ama-assn.org/ama1/pub/upload/mm/21/omss-bylaws-4thedition-final.pdf. 
  10. "Medical Peer Review". http://www.ama-assn.org/ama/pub/physician-resources/legal-topics/medical-peer-review.shtml. 
  11. "Peer review: What is it and why do we do it?" (in en). 2019-03-29. https://www.medicalnewstoday.com/articles/281528. 
  12. NASA Systems Engineering Handbook. NASA. 2007. SP-610S. https://ntrs.nasa.gov/archive/nasa/casi.ntrs.nasa.gov/20080008301_2008008500.pdf. 
  13. Couchman, John R. (November 11, 2013). "Peer Review and Reproducibility. Crisis or Time for Course Correction?". Journal of Histochemistry & Cytochemistry 62 (1): 9–10. doi:10.1369/0022155413513462. PMID 24217925. 
  14. "The peer review drugs don't work" (in en). Times Higher Education (THE). 2015-05-28. https://www.timeshighereducation.com/content/the-peer-review-drugs-dont-work. 
  15. "Peer review 'works against' early career researchers" (in en). Times Higher Education (THE). 2018-07-16. https://www.timeshighereducation.com/news/peer-review-works-against-early-career-researchers. 
  16. JPA Ioannidis (2005) "Why Most Published Research Findings Are False"
  17. 17.0 17.1 17.2 Csiszar, Alex (2016). "Peer Review: Troubled from the Start". Nature 532 (7599): 306–308. doi:10.1038/532306a. PMID 27111616. Bibcode2016Natur.532..306C. 
  18. 18.0 18.1 Moxham, Noah; Fyfe, Aileen (2018). "The Royal Society and the Prehistory of Peer Review, 1665–1965". The Historical Journal 61 (4): 863–889. doi:10.1017/S0018246X17000334. https://kar.kent.ac.uk/65042/3/Peer%20review%20v30%20AAM%20SUBMTD.pdf. 
  19. Moore, John (2006). "Does Peer Review Mean the Same to the Public as It Does to Scientists?". Nature. doi:10.1038/nature05009. 
  20. Ferguson, Cat; Marcus, Adam; Oransky, Ivan (2014). "Publishing: The Peer-Review Scam". Nature 515 (7528): 480–482. doi:10.1038/515480a. PMID 25428481. Bibcode2014Natur.515..480F. 
  21. 21.0 21.1 Budd, J. M.; Sievert, M.; Schultz, T. R. (1998). "Phenomena of Retraction: Reasons for Retraction and Citations to the Publications". JAMA 280 (3): 296–7. doi:10.1001/jama.280.3.296. PMID 9676689. 
  22. 22.0 22.1 Smith, Richard (2006). "Peer Review: A Flawed Process at the Heart of Science and Journals". Journal of the Royal Society of Medicine 99 (4): 178–82. doi:10.1177/014107680609900414. PMID 16574968. 
  23. Ross-Hellauer, Tony (2017). "What Is Open Peer Review? A Systematic Review". F1000Research 6: 588. doi:10.12688/f1000research.11369.2. PMID 28580134. 
  24. 24.0 24.1 Tennant, Jonathan P.; Dugan, Jonathan M.; Graziotin, Daniel; Jacques, Damien C.; Waldner, François; Mietchen, Daniel; Elkhatib, Yehia; b. Collister, Lauren et al. (2017). "A Multi-Disciplinary Perspective on Emergent and Future Innovations in Peer Review". F1000Research 6: 1151. doi:10.12688/f1000research.12037.3. PMID 29188015. 
  25. Wong, Victoria S. S.; Avalos, Lauro Nathaniel; Callaham, Michael L. (2019). "Industry Payments to Physician Journal Editors". PLOS ONE 14 (2): e0211495. doi:10.1371/journal.pone.0211495. PMID 30730904. Bibcode2019PLoSO..1411495W. 
  26. Weiss, Glen J.; Davis, Roger B. (2019). "Discordant Financial Conflicts of Interest Disclosures between Clinical Trial Conference Abstract and Subsequent Publication". PeerJ 7: e6423. doi:10.7717/peerj.6423. PMID 30775185. 
  27. Flaherty, D. K. (2013). "Ghost- and Guest-Authored Pharmaceutical Industry–Sponsored Studies: Abuse of Academic Integrity, the Peer Review System, and Public Trust". The Annals of Pharmacotherapy 47 (7–8): 1081–3. doi:10.1345/aph.1R691. PMID 23585648. 
  28. 28.0 28.1 28.2 28.3 28.4 28.5 28.6 28.7 Vanholsbeeck, Marc; Thacker, Paul; Sattler, Susanne; Ross-Hellauer, Tony; Rivera-López, Bárbara S.; Rice, Curt; Nobes, Andy; Masuzzo, Paola et al. (2019-03-11). "Ten Hot Topics around Scholarly Publishing". Publications 7 (2): 34. doi:10.3390/publications7020034. 
  29. Relman, A. S. (1990). "Peer Review in Scientific Journals--What Good Is It?". Western Journal of Medicine 153 (5): 520–22. PMID 2260288. 
  30. Bravo, Giangiacomo; Grimaldo, Francisco; López-Iñesta, Emilia; Mehmani, Bahar; Squazzoni, Flaminio (2019). "The Effect of Publishing Peer Review Reports on Referee Behavior in Five Scholarly Journals". Nature Communications 10 (1): 322. doi:10.1038/s41467-018-08250-2. PMID 30659186. Bibcode2019NatCo..10..322B. 
  31. 31.0 31.1 Tennant, Jonathan P. (2018). "The State of the Art in Peer Review". FEMS Microbiology Letters 365 (19). doi:10.1093/femsle/fny204. PMID 30137294. 
  32. Squazzoni, Flaminio; Grimaldo, Francisco; Marušić, Ana (2017). "Publishing: Journals Could Share Peer-Review Data". Nature 546 (7658): 352. doi:10.1038/546352a. PMID 28617464. Bibcode2017Natur.546Q.352S. 
  33. Allen, Heidi; Boxer, Emma; Cury, Alexandra; Gaston, Thomas; Graf, Chris; Hogan, Ben; Loh, Stephanie; Wakley, Hannah et al. (2018). What Does Better Peer Review Look like? Definitions, Essential Areas, and Recommendations for Better Practice. doi:10.17605/OSF.IO/4MFK2. 
  34. Fang, Ferric C.; Casadevall, Arturo (2011). "Retracted Science and the Retraction Index". Infection and Immunity 79 (10): 3855–3859. doi:10.1128/IAI.05661-11. PMID 21825063. 
  35. Moylan, Elizabeth C.; Kowalczuk, Maria K. (2016). "Why Articles Are Retracted: A Retrospective Cross-Sectional Study of Retraction Notices at BioMed Central". BMJ Open 6 (11): e012047. doi:10.1136/bmjopen-2016-012047. PMID 27881524. 
  36. Open Science Collaboration (2015). "Estimating the Reproducibility of Psychological Science". Science 349 (6251): aac4716. doi:10.1126/science.aac4716. PMID 26315443. 
  37. 37.0 37.1 Munafò, Marcus R.; Nosek, Brian A.; Bishop, Dorothy V. M.; Button, Katherine S.; Chambers, Christopher D.; Percie Du Sert, Nathalie; Simonsohn, Uri; Wagenmakers, Eric-Jan et al. (2017). "A Manifesto for Reproducible Science". Nature Human Behaviour 1. doi:10.1038/s41562-016-0021. http://dare.uva.nl/personal/pure/en/publications/a-manifesto-for-reproducible-science(3534b98f-a374-496b-9ad1-e61539477d66).html. 
  38. Fanelli, Daniele (2018). "Opinion: Is Science Really Facing a Reproducibility Crisis, and Do We Need It To?". Proceedings of the National Academy of Sciences 115 (11): 2628–2631. doi:10.1073/pnas.1708272114. PMID 29531051. 
  39. Goodman, Steven N. (1994). "Manuscript Quality before and after Peer Review and Editing at Annals of Internal Medicine". Annals of Internal Medicine 121 (1): 11–21. doi:10.7326/0003-4819-121-1-199407010-00003. PMID 8198342. 
  40. Pierson, Charon A. (2018). "Peer review and journal quality". Journal of the American Association of Nurse Practitioners 30 (1): 1–2. doi:10.1097/JXX.0000000000000018. PMID 29757914. 
  41. Caputo, Richard K. (2019). "Peer Review: A Vital Gatekeeping Function and Obligation of Professional Scholarly Practice". Families in Society: The Journal of Contemporary Social Services 100: 6–16. doi:10.1177/1044389418808155. 
  42. Siler, Kyle; Lee, Kirby; Bero, Lisa (2015). "Measuring the effectiveness of scientific gatekeeping". Proceedings of the National Academy of Sciences 112 (2): 360–365. doi:10.1073/pnas.1418218112. PMID 25535380. Bibcode2015PNAS..112..360S. 
  43. Resnik, David B.; Elmore, Susan A. (2016). "Ensuring the Quality, Fairness, and Integrity of Journal Peer Review: A Possible Role of Editors". Science and Engineering Ethics 22 (1): 169–188. doi:10.1007/s11948-015-9625-5. PMID 25633924. 
  44. Bornmann, Lutz (2011). "Scientific Peer Review". Annual Review of Information Science and Technology 45: 197–245. doi:10.1002/aris.2011.1440450112. 
  45. "Cargo Cult Science". Caltech Magazine. 1974. http://Calteches.library.caltech.edu/51/2/CargoCult.HTM. 
  46. 46.0 46.1 "Untangling Academic Publishing. A History of the Relationship between Commercial Interests, Academic Prestige and the Circulation of Research.". 26. 
  47. Priem, Jason; Hemminger, Bradley M. (2012). "Decoupling the Scholarly Journal". Frontiers in Computational Neuroscience 6: 19. doi:10.3389/fncom.2012.00019. PMID 22493574. 
  48. Bowman, Nicholas David; Keene, Justin Robert (2018). "A Layered Framework for Considering Open Science Practices". Communication Research Reports 35 (4): 363–372. doi:10.1080/08824096.2018.1513273. 
  49. McKiernan, E. C.; Bourne, P. E.; Brown, C. T.; Buck, S.; Kenall, A.; Lin, J.; McDougall, D.; Nosek, B. A. et al. (2016). "Point of View: How Open Science Helps Researchers Succeed". eLife 5. doi:10.7554/eLife.16800. PMID 27387362. 
  50. 50.0 50.1 "In Peer Review We (Don't) Trust: How Peer Review's Filtering Poses a Systemic Risk to Science". https://www.researchers.one/article/2018-09-17. 
  51. Brembs, Björn (2019). "Reliable Novelty: New Should Not Trump True". PLOS Biology 17 (2): e3000117. doi:10.1371/journal.pbio.3000117. PMID 30753184. 
  52. Stern, Bodo M.; o'Shea, Erin K. (2019). "A Proposal for the Future of Scientific Publishing in the Life Sciences". PLOS Biology 17 (2): e3000116. doi:10.1371/journal.pbio.3000116. PMID 30753179. 

Further reading

  • Hames, Irene (2007). Peer Review and Manuscript Management in Scientific Journals : Guidelines for Good Practice. Oxford, UK: Blackwell Publishing. ISBN 978-1-4051-3159-9. 

External links