Research Culture: Co-reviewing and ghostwriting by early-career researchers in the peer review of manuscripts
Abstract
Many early-career researchers are involved in the peer review of manuscripts for scientific journals, typically under the guidance of or jointly with their advisor, but most of the evidence about this activity is anecdotal. Here we report the results of a literature review and a survey of researchers, with an emphasis on co-reviewing and 'ghostwriting'. The literature review identified 36 articles that addressed the involvement of early-career researchers in peer review, most of them about early-career researchers and their advisors co-reviewing manuscripts for the purposes of training: none of them addressed the topic of ghostwriting in detail. About three quarters of the respondents to the survey had co-reviewed a manuscript. Most respondents believe co-reviewing to be a beneficial (95%) and ethical (73%) form of training in peer review. About half of the respondents have ghostwritten a peer review report, despite 81% responding that ghostwriting is unethical and 82% agreeing that identifying co-reviewers to the journal is valuable. Peer review would benefit from changes in both journal policies and lab practices that encourage mentored co-review and discourage ghostwriting.
https://doi.org/10.7554/eLife.48425.001Introduction
The peer review of manuscripts submitted to scientific journals is widely viewed as fundamental to efforts to maintain the integrity of the scientific literature (Baldwin, 2018; Tennant, 2017). Early-career researchers (ECRs) often contribute to the peer review process. Indeed, in a recent survey that targeted ECRs in the life sciences, 92% of respondents reported that they had been involved in the peer review of at least one manuscript (Inside eLife, 2018). More than half of survey respondents, including 37% of graduate students, reported reviewing a manuscript without any assistance from their advisor. Journals may not be fully aware of the extent to which ECRs are involved in peer review (McDowell, 2018). Indeed, a recent editorial in this journal contained the following sentence: "It is common practice for busy group leaders to ask their more senior PhD students and postdoctoral fellows to help with peer review, but in too many cases these contributions go unacknowledged" (Patterson and Schekman, 2018).
We conducted a literature search and a survey to explore the involvement of ECRs in peer review more thoroughly and, in particular, to determine the prevalence of both co-reviewing (i.e., when the journal knows that the ECR contributed to the review) and 'ghostwriting' (i.e., when the journal does not know that the ECR contributed to the review). Please see Table 1 for a definition of terms used in this article.
Results
Lack of literature on ECR ghostwriting of peer review reports
We performed a comprehensive review of the peer-reviewed literature to identify any previous studies on the role that ECRs play in peer review, particularly with respect to ghostwriting. Exhaustive search terms that combined any synonyms of "early-career researcher" and “peer review” were used per evidenced-based guidelines for systematic reviews (prisma-statement.org; PRISMA Group et al., 2009; see Methods: Systematic literature review for details). Our search yielded 1952 unique articles. Collected articles underwent two rounds of screening performed independently by three of the present authors using titles and abstracts to evaluate relevance to the topic of ECR co-reviewing and ghostwriting peer review reports (see Methods: Relevance screening for details; Supplementary file 1; Figure 1—figure supplement 1).
We did not find any research articles on ghostwriting peer review reports by ECRs. One article (the eLife editorial mentioned previously; Patterson and Schekman, 2018) acknowledged the phenomenon of ECR ghostwriting, and stated that ECRs are eligible to act as peer reviewers for manuscripts submitted to eLife. 35 other articles addressed ECR involvement in manuscript peer review but did not address ghostwriting. Many of these instead investigated the value of co-reviewing as a training exercise (see Appendix 1). None discussed the issue of named credit for scholarly labor, nor did they include information on the frequency of ghostwriting in peer review or the opinions of ECRs on ghostwriting.
Surveying the rates and rationales for co-reviewing and ghostwriting
To address this gap in the literature, we designed a survey to evaluate the frequency of, and motivations for, ghostwriting and co-reviewing by ECRs. The IRB-approved, online survey garnered 498 responses over a month-long data collection period in September, 2018 (see Methods: Survey of peer review experiences and attitudes for details; the survey itself is available in Supplementary file 2). Respondents came from 214 institutions that were geographically diverse both within and beyond the United States. Most participants were from institutions in North America (n = 370), Europe (n = 87) and Asia (n = 21). 74% of all respondents were based in the US, of which 64% were citizens or permanent residents, and 36% held temporary visitor status. Institutions from 40 US states or territories were represented: the four universities with the most respondents were Washington University in St. Louis, the University of Kentucky, Rockefeller University, and the University of Chicago. The majority of survey respondents (65%) were ECRs in the life sciences (Figure 1). This was as expected given our efforts to primarily engage ECRs and our connections to biomedical postdoctoral populations (see Methods: Survey distribution, limitations, and future directions). We surmise that postdocs (63% of all respondents) are over-represented in this survey, although the number of postdoctoral researchers in the US is currently unknown (Pickett et al., 2017).
-
Figure 1—source data 1
- https://doi.org/10.7554/eLife.48425.005
Co reviewing by ECRs is the norm and motivated by training
It is a widespread practice to contribute ideas and/or text to a peer review report when one is not the invited reviewer. 73% of all respondents have co-reviewed and often at numerous times (33% on 6–20 occasions, and 4% on more than 20 occasions; Figure 2A). Co-reviewing by ECRs specifically is common, with 79% of postdocs and 57% of PhD students having co-reviewed when “the invited reviewer is the PI for whom you work” (Figure 2B). These data suggest that collaboration on peer review reports is an academic norm, especially by ECRs who are not the invited reviewer. By contrast, 55% of ECR respondents have never carried out independent peer review as the invited reviewer (Figure 2C).
A major motivation for ECRs to co-review is to gain training in peer review of manuscripts, a fundamental scholarly skill. All survey respondents were asked what training they received in peer review of manuscripts (Figure 3). Respondents report that PIs are the second most common source of peer review training, bested only by the passive form of learning “from receiving reviews on my own papers.” Training through co-review was the subject of many publications uncovered by our literature review (Appendix 1).
Ghostwriting is common despite a belief it is unethical
The frequency of ghostwriting was measured in two survey questions which revealed comparable rates. When we asked “To your knowledge, did your PI ever withhold your name from the editorial staff when you served as the reviewer or co-reviewer?,” 46% of respondents knew that their name had been withheld (Figure 4A). These data are consistent with results from a separate question: “When you were not the invited reviewer, what was the extent of your involvement in the peer review process?”. For this question, 44% of respondents reported having had the experience of ghostwriting: “I read the manuscript, wrote the report, my PI edited the report and my PI submitted report with only their name provided to the editorial staff” (Table 2). Taken together, these data suggest that approximately 1 in 2 survey respondents has engaged in ghostwriting of a peer review report on behalf of their PI, the invited reviewer. Furthermore, 70% of co-reviewers report the experience of making significant contributions to a peer review report without knowingly receiving credit (Table 2). This experience is much more common than the 22% of co-reviewers who experienced making significant contributions with known credit (Table 2). These data reveal a breakdown in communication between invited reviewers and co-reviewers.
In a more specific follow up question that asked “To your knowledge, did your PI ever submit your reviews without editing your work?”, 52% of respondents report that they were not involved in any editing process with their PI (Figure 4B). This proportion is similar to that reported in Inside eLife (2018). That survey asked “Have you reviewed before?” and then “If so, to what extent was your supervisor involved?” to which slightly more than half of the 264 respondents replied “not at all.” One interpretation of these data is that half of respondents had engaged in independent peer review as the invited reviewer. Another interpretation of these data is that half of respondents had engaged in co-review with no feedback from their supervisor, the invited reviewer. Our data support the latter interpretation that slightly more than half of respondents have written peer review reports without feedback from their PI when the PI is the invited reviewer.
The Office of Research Integrity (ORI), which oversees research funded by the US Public Health Service, states that “academic or professional ghost authorship in the sciences is ethically unacceptable” (https://ori.hhs.gov/plagiarism-34). Respondents are also of the view that ghostwriting peer review reports is unethical: 83% disagree with the statement that “The only person who should be named on a peer review report is the invited reviewer, regardless of who carried out the review”; 81% disagree with the statement that “Ghostwriting a peer-review report for your PI is an ethically sound scientific practice”; and 77% disagree with the statement that “It is ethical for the invited reviewer (e.g. PI) to submit a peer review report to an editor without providing the names of all individuals who have contributed ideas and/or text to the report” (Figure 5). Respondents were also supportive of co-reviewing providing their contributions are known about: 74% agree that “Anyone that contributes ideas and/or text to the review report should be included as a co-author on the review”; 82% agree that “It would be valuable to have my name added to a peer review report (e.g. to be recognized as a co-reviewer by the editor; or to use a service such as Publons to be assigned credit)”; 73% agree that “It is ethical for the invited reviewer (e.g. PI) to involve others (e.g. their trainees) in reviewing manuscripts”; and 95% agree that “Involving members of a research group in peer review is a beneficial training exercise.” The latter statement evoked the strongest positive sentiment of all 11 Agree/Disagree statements.
-
Figure 5—source data 5
- https://doi.org/10.7554/eLife.48425.011
There was a significant difference in the extent of agreement between ECRs and PIs for certain aspects of co-reviewing and ghostwriting (Table 3). In 3 of 11 statements, ECRs felt significantly more strongly than PIs but still shared the same valance (e.g. both groups agreed or both groups disagreed with the statement, just to a differing strength). For the remaining 8 statements, ECRs and PIs did not significantly differ in their opinions.
Motivations for ghostwriting
If 4 out of 5 survey respondents think ghostwriting is unethical, then why do half of all respondents participate? We measured the motivations for ghostwriting by: i) asking all respondents, regardless of peer review experience, to surmise why someone might withhold the name of a co-reviewer (Figure 6); ii) asking only respondents with ghostwriting experience to report the specific reasons that the invited reviewer gave for withholding their name (Table 4). In this way, we compared cultural beliefs with actual practice.
The main perceived barrier to naming co-reviewers was a lack of a physical mechanism to supply the name to the journal (e.g. a textbox for co-review names), with 73% of respondents selecting this as an option (Figure 6). Cultural expectations were the next most commonly-cited barriers, such as “A belief that reviews should only be done by the invited reviewer, and not by, or with assistance from, anyone else” (selected by 63% of respondents) and “A belief that including co-author information would demonstrate that the PI breached the confidentiality of the manuscript” (58%). These latter responses allude to journal policies prohibiting invited reviewers from sharing unpublished manuscripts without prior permission. Write-in responses echo themes about how ghostwriting is the status quo in peer review (Table 5). At the same time, respondents also wondered why including co-reviewer names is not common practice.
In contrast, when we asked “Consider cases where you contributed to a peer review report and you know your name was NOT provided to the editorial staff. When discussing this with your PI, what reason did they give to exclude you as a co-reviewer?” 73% of respondents reported that they had not discussed this with their PI (Table 4). This is consistent with the lack of communication between invited reviewers and co-reviewers documented above (Table 2; Figure 4B). Of the 27% of respondents who had ghostwritten and did discuss the matter with their PI, most were told that the reason their name was withheld was either a prohibitive journal policy and/or prevailing cultural expectations about co-authorship on peer review reports. Only 4% of those who had discussed the matter with their PI cited a practical barrier, such as the lack of a text box for co-reviewer names on the journal review submission form. Write-in responses to this question again refer to cultural expectations as the major drivers for ghostwriting (Table 6). Many of these write-in reasons articulate that it is good practice for ECRs to participate in peer review; however, they simultaneously fail to explain why this necessitates withholding the names of co-reviewers. These data suggest that there is a common conflation of ghostwriting (withholding names) with co-reviewing (involving ECRs in peer review, often for the purposes of training).
Other demographic analyses
We also performed preliminary analyses by gender, field, citizenship and race/ethnicity, but any differences we observed were small or not statistically significant, and some demographic subsets were too small for analysis. We are considering how to share these data more completely.
Discussion
Journal peer review is an important part of scholarship. As such, ECR training and authorship on peer review reports deserves thoughtful consideration to ensure the integrity of the peer review process. Our data reveal that involving ECRs in peer review is an academic norm, with about three-quarters of our survey population having contributed significantly to a peer review report when they were not the invited reviewer (co-reviewed), and about half having done so without being named to the journal editorial staff (ghostwritten). These high frequencies contrast with journal policies and cultural expectations that only the invited reviewer engages in the peer review of a manuscript. They also fly in the face of community values when about four-fifths of those surveyed agree that ghostwriting is unethical. What drives these differences between community values and experience?
Explanations for ghostwriting are conflated with explanations for co-reviewing
Co-reviewing as a training exercise and ghostwriting are separable processes: training through co-review can and does happen whether or not named credit is given to the co-reviewer, and excluding co-reviewer names from peer review reports can and does happen whether or not the co-reviewer has experienced quality training in the process. Even as we sought to collect data that would disentangle these two processes, the rationales for ghostwriting were often conflated with the rationales for co-reviewing. For example, when we asked respondents specifically for the reason(s) their PI gave them for excluding their names on a peer review report, many wrote responses such as "I was told this is how one gets to train to review papers..." (Table 6). This response – that it is a beneficial and common practice for ECRs to participate in peer review as a training exercise – does not actually explain why ghostwriting occurs. Reducing ghostwriting requires decoupling it in the zeitgeist from the beneficial training of ECRs through co-review. We therefore separate our discussion of the motivations for co-reviewing and ghostwriting in an effort to find solutions to ghostwriting that simultaneously support ECR co-reviewing as training in peer review.
Co reviewing by ECRs as valued training or delegation of scholarly labor
Survey respondents clearly find that co-reviewing by ECRs has significant benefits and is not inherently problematic when the issue of named credit is set aside. Co-review is considered an ethical (73% agree) and beneficial training exercise (95% agree), explaining why co-reviewing is the second most commonly reported source of training in peer review. These data should be weighed heavily when considering journal policies, since any policy that prevents co-reviewing by ECRs would remove a common and valuable training exercise in peer review. We support the adoption of policies that specifically embrace ECR co-review as training (e.g., eLife, 2019). The Transpose project is also compiling a crowd-sourced database of journal policies on peer review, co-reviewing and preprinting (Transpose, 2019).
“ECR training” is not a sufficient justification for co-reviewing for the half of survey respondents who have written a peer review report without any interaction with their PIs. In these cases, co-reviewing devolves into a delegation of scholarly labor that benefits the invited reviewer, often an overburdened PI in a hypercompetitive research environment (Alberts et al., 2014). This environment incentivizes the use of ECRs as cheap labor to fulfill productivity requirements, especially as the growth of the ECR population outpaces the growth of independent academic positions to employ them (Heggeness et al., 2017; Heggeness et al., 2016). These market forces provide a second explanation, beyond ECR training, for why co-reviewing is commonplace. Our survey respondents agree, sharing sentiments like “apparently this duty is part of my job description” (Table 6). The delegation of scholarly labor to ECRs is not necessarily, nor intentionally, exploitative, although it can easily become so given the power dynamics and documented lack of communication between mentors and mentees (Van Noorden, 2018). Any successful intervention to address concerns about the ethics of co-reviewing by ECRs must take into account that it is commonly an unstated expectation that ECRs carry out peer review on behalf of their PI, and that ECRs may not feel they have the freedom to decline even if they feel it is unethical to participate.
Limitations to depending on co-reviewing as training
Survey data suggest that training in peer review is determined by a small number of individual experiences outside of evidence-based training structures and community oversight. “Receiving reviews on my own papers” only gives a limited number of examples of how others review and is a passive form of learning that lacks individualized or iterative feedback from a mentor. Since it is a common complaint that reviews are overly critical (Schneiderhan, 2013), it seems counterintuitive for this to be the main example by which ECRs learn how to review. Training provided by one’s PI may benefit from a personalized teaching relationship but depends on the PI’s own training. The resulting trickle-down training is likely to be self-reinforcing and highly variable in quality and content. The small sample size of these different personal experiences may also reinforce bias, including selection bias where one’s few experiences may not be representative of the population, memory biases where negative experiences (e.g. from receiving harsh reviews) are more readily remembered and so taught (Kensinger, 2007), and gender or other demographic biases currently being studied in the peer review process (Murray et al., 2018).
These results highlight an area of opportunity to improve and standardize training in peer review of manuscripts as a critical scholarly skill (Appendix 1). Since the top two reported forms of training involve PIs either as manuscript reviewers or ECR mentors, interventions to ensure PIs have received training in peer review and in communicating this skill to ECRs may also be appropriate. A lack of “training the trainers” was cited as a main reason for why pairing experts with new peer reviewers failed to improve review quality in one of the few randomized controlled trials of this practice (Houry et al., 2012). Given their key role as the main source of training in peer review, it is important that PIs make deliberate efforts to teach their trainees this skill and to provide their trainees with feedback. Mentorship from PIs may be complemented with, but not replaced by, journal clubs where ECRs gain experience in reviewing published manuscripts or preprints (see, for example Avasthi et al., 2018 and PREreview, 2019). Since these ad hoc reviewing experiences with PIs and journal clubs may vary in their availability and quality, all ECRs should be offered standardized, evidence-based training in peer review. For example, peer review courses that are compulsory and ubiquitous in graduate schools would ensure that all PhD-holders are appropriately trained to perform constructive peer review.
A lack of communication about authorship of peer review reports
Even in the best case scenario for co-reviewing, when training is taking place and is effective, its benefits can still be confounded by ethical lapses such as ghostwriting. The most common explanation for ghostwriting was that authorship was simply not discussed (Table 4; 73% of responses from those who knew their names had been withheld). 47% of respondents have had the experience of “I read the manuscript, wrote a full report for my PI, and was no longer involved” (Table 2), a concerning breakdown in communication between invited reviewers and those actually writing the peer review report. At best, writing a peer review report without receiving feedback from one’s PI is a missed opportunity for training. At worst, a peer review report that is written by one person and submitted to the journal under the name of another person is a breach of academic integrity.
Power imbalances may prevent an ECR from feeling able or willing to initiate this conversation with their PI. 39% of respondents think that ghostwriting occurs because "some ECRs may not be comfortable asking for co-authorship" (Figure 6). As one write-in states: “you don’t want to piss off the boss.” Ghostwriting, therefore, may be a symptom of the larger problem in academia that ECRs are extremely dependent on the good will of their PI for retention in the hypercompetitive research environment (for example, for letters of recommendation throughout their career or immigration status). Another reason that authorship may not be discussed is that not naming co-reviewers is the expected status quo: “[PI] didn't think of including me" and "they forgot" (Table 6). One postdoc added: “I'd never really thought about this before. I just assumed it was part of the process. But it is very time consuming and I do believe that all reviewers should receive credit for the review.” If ghostwriting arises from PIs and ECRs simply not thinking to include co-reviewer names vs. intentional withholding of names, then building awareness should encourage more conversation about this issue and better mentoring practices will help overcome such miscommunication. PIs could be further incentivized to name ECR co-reviewers by accounting for this practice when they are evaluated for grant funding or tenure. For example, perhaps PIs who pay the salaries of their trainees from grant budgets should demonstrate in grant progress reports how those ECRs are being trained, which might include a list of which trainees were listed as co-reviewers.
An expectation that co-reviewers do not deserve credit regardless of what they contribute
Ghostwriting is also driven by a cultural expectation that co-reviewers do not deserve named credit to the journal regardless of how much they contribute. 43% of ghostwriting experiences were explained by “co-authorship is for papers, not for peer review reports” or “intellectual contribution not deemed sufficient.” 28% of respondents surmise that ghostwriting occurs due to "a belief that only the invited reviewer deserves authorship, even when others contribute ideas and/or text to the review report." These rationales for ghostwriting contradict community opinion about whether this should be the case, with 83% disagreeing that “the only person who should be named on a peer review report is the invited reviewer, regardless of who carried out the review” and 74% agreeing that “anyone that contributes ideas and/or text to the review report should be included as a co-author on the review.” Ghostwriting could therefore be reduced if cultural expectations were shifted to reflect consensus opinion that co-reviewers deserve to be named to the journal.
Prohibitive journal policy is out of alignment with current practice
Many journals have policies that prevent invited reviewers from sharing manuscripts with anyone else and/or policies that prevent ECRs from serving as reviewers or co-reviewers without prior permission from the editor (Transpose, 2019). These policies are the other most common justification for ghostwriting. 58% of respondents surmise that ghostwriting occurs because of “a belief that including co-author information would demonstrate that the PI breached the confidentiality of the manuscript” and 63% surmising that ghostwriting occurs because of “a belief that reviews should only be done by the invited reviewer, and not by, or with assistance from, anyone else.” These policies derive from guidelines developed by the Committee on Publication Ethics (COPE): “supervisors who wish to involve their students or junior researchers in peer review must request permission from the editor and abide by the editor’s decision” (COPE Council, 2017). Yet it seems that, in practice, invited reviewers often do not seek prior permissions, continue to involve ECR co-reviewers, and instead choose to withhold co-reviewer names upon submission. Indeed, 39% of ghostwriters who discussed with the invited reviewer the possibility of including their name cite “journal requires prior approval to share manuscript, which was not obtained” or “journal does not allow ECRs to review” as the reason why their names were withheld. In these cases, adding co-reviewer names to a peer review report is equivalent to admitting that journal policies were disobeyed. Given how frequently ghostwriting occurs based on survey data, and how commonly journal policies are cited as the reason for ghostwriting, it seems that current policies that require invited reviewers to gain permission prior to involving ECRs in peer review are not effective deterrents for ghostwriting. Instead, these policies may have the opposite, if unintended, consequence of preventing invited reviewers from feeling free to add co-reviewer names upon submission. Journals should acknowledge that peer review is often performed by ECR co-reviewers and remove barriers that prevent ECRs from being named to the editor (Rodríguez-Bravo et al., 2017).
The naming of co-reviewers would be facilitated by, for example, a mandatory text-box on the page that reviewers use to submit their review: this page could also contain language asking for co-reviewers be listed. 73% of respondents surmise that ghostwriting occurs because of a lack of such a mechanism. However, when asked to reflect on their own experiences, only 4% of ghostwriters gave this as a reason (though that might be due to a limitation in the design of this question, which did not include this answer in the drop-down menu and instead relied on respondents to write it in). Even with this consideration, these data reveal a difference between the cultural perception of this barrier (73%) and actual experience (4% or more), so any practical solutions (such as adding a text-box) must be accompanied by changes that make it clear that journals expect all co-reviewers to be named.
It is in the best interests of journals to provide mechanisms for ECR co-reviewers to be easily named (Mehmani, 2019). If a journal does not know who is reviewing a paper, it cannot be sure that there are no competing interests among the reviewers. Editors may also see an increase in accepted invitations to review once PIs feel free to share the burden of reviewing with their trainees without ethical concerns. Journals would also benefit from enlarged reviewer pools by including previous ECR co-reviewers and/or consulting existing lists of ECR reviewers (Burgess, 2018). Ultimately, journals are responsible for ensuring that all aspects of peer review are fair and ethical. Ghostwriting is neither fair nor ethical, so the evidence for ghostwriting reported here suggests that journals need to update their policies in this area.
Value judgments about naming co-reviewers
Two hypotheses for why ghostwriting is commonplace were refuted by our data. 82% of respondents disagree that there is no benefit in naming co-reviewers. Co-reviewers, especially ECRs, may value having their name provided to the journal for many reasons, including the ability to be “known” to scientific editors and potential colleagues; the ability to have their work acknowledged by a third party (e.g. Publons) for career advancement; and/or the ability to demonstrate eligibility for residency or visas. ECRs agree more strongly than PIs that there is value in receiving credit or being known to the journal editorial staff as a co-reviewer (Table 3), perhaps because invited reviewers are already known as experts in their field. In the words of one write-in response: “PI surprised I would be interested in being acknowledged, and seemed like too much trouble to acknowledge my contribution. There was no box in the online form to declare it.” The ambivalence of PIs towards giving due credit for co-reviewers likely derives from a position of relative privilege. When people of privilege are the only participants in decision-making (for example, on journal editorial boards), they may create policies that fail to consider differing values of less privileged members of the community, like ECRs. We support the growing effort to include more diverse voices in leadership roles in science (such as the “Who’s on board” initiative of Future of Research, and the Early-Career Advisory Group at eLife).
87% of respondents disagree that ghostwriting occurs because researchers see value in withholding co-reviewer names, perhaps because of a perception that adding co-reviewer names diminishes the review by providing evidence that someone other than the invited reviewer contributed. These data align with results of an ongoing experiment in co-reviewing at the Elsevier journals, in which editors did not rate the co-reviewed reports as low quality, and more than half were considering co-reviewers to serve as invited reviewers on future manuscripts (Mehmani, 2019). Research suggests that reviewers who are earlier in their careers may be perceived by editors as better reviewers (Black et al., 1998; Callaham and Tercier, 2007; Evans et al., 1993) and that being closer to bench research, rather than having more experience in reviewing itself, may be a key determinant of this trait (Stossel, 1985). Another reason for a perceived value in withholding co-reviewers names is to protect ECRs during a vulnerable time in their careers. Occasional write-in comments allude to this: “[being named on a peer review report] may give certain ECRs a bad reputation if they review things really harshly” but, on the whole, only 17% of respondents believe protectionism drives ghostwriting. Taken together, respondents find added value in naming co-reviewers and also see no loss of value to peer review when co-reviewer names are added. Finally, the null hypothesis that “a belief that ghostwriting does not occur” was selected least frequently by respondents, demonstrating that ignorance or denial that ghostwriting occurs is rare (Figure 6).
Conclusions
Ghostwriting undermines the integrity of peer review. It is pervasive because many see it as an obligatory feature of peer review training or a necessary delegation of labor. Some don’t think to discuss or feel able to discuss authorship on peer review reports. Others are deterred by vague journal policies that do not reflect the status quo – that involving ECRs as co-reviewers is common and considered valuable and ethical. To encourage naming co-reviewers to the editors, journals must clarify their expectations and reporting mechanisms for the participation of ECRs in peer review. These logistical changes should be coupled with an adjustment of cultural expectations for co-review as a training exercise and not exploitation. At a minimum, invited reviewers should discuss with co-reviewers how credit will be given for peer review work. Ideally, they should also ensure that co-review involves feedback so that it is effective training. Changing journal policies and cultural expectations to recognize and value the work of ECRs will benefit the peer review system and all of its constituents.
Methods
Systematic literature review
Search procedures
The following procedures were used to perform a systematic review of the peer-reviewed literature for any research on the topic of ECR participation in manuscript peer review. These procedures were developed under the guidance of a professional librarian (author SO) and were based on the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) criteria (PRISMA Group et al., 2009; Figure 1—figure supplement 1). The databases that were searched cover peer-reviewed literature across the life sciences, public policy and social sciences and were comprised of: PubMed, PsychInfo, Web of Science, and PAIS International. These databases were then searched using the following keyword search strategy: ("early career researcher" OR "graduate student" OR "postdoc" OR "fellow" OR "contingent faculty" OR "adjunct" OR "lecturer" OR "instructor" OR "technician" OR "junior scientist" OR "trainee" OR "lab member" OR "research scientist" OR "postdoctoral fellow" OR "research fellow" OR "teaching fellow" OR "junior researcher" OR "mentee") AND ("peer review" OR "refereeing" OR "invited reviewer" OR "referee" OR "reviewer" OR "co-reviewer" OR "first time reviewers" OR "reviewer training" OR "review partners" OR "contributing author" OR "co-reviewing" OR "reviewing" OR "journal reviewer policy" OR "reviewer guidelines" OR "instructions for reviewers"). These search terms were designed to be broadly inclusive so as to capture any research article with possible relevance to the topic of ECR involvement in manuscript peer review. The resulting collection of 2103 records were imported into the RefWorks 3 bibliographic management database, and duplicate articles were identified and removed using the “Legacy close match” de-duplication filter, resulting in a de-duplicated set of 1952 articles.
Relevance screening
Collected articles underwent two rounds of relevance screening. In the initial round, article titles and abstracts were screened independently by two study authors (JG and GM). Both authors used the same inclusion criteria to sort search results into “relevant,” “maybe relevant,” and “not relevant” categories. The criteria for article inclusion were: written in English, published in a peer-reviewed journal, mention of ECRs, and mention of peer reviews of manuscripts. Any article that did not meet the inclusion criteria above was excluded as well as database hits for conference proceedings and dissertations.
118 unique articles remained in the “relevant” and/or “maybe relevant” categories at this stage of screening (Supplementary file 1). Articles categorized as “relevant” by both initial screeners were selected for full text review (n = 3). Articles categorized as “maybe relevant” by both initial screeners and articles that were differentially categorized as “relevant” vs. “maybe relevant” or “not relevant” by the initial screeners (n = 51) underwent a second round of evaluation by a third, independent screener (author RL) to either confirm categorization as “relevant” or recategorize as “not relevant” to the topic of ECR participation in the peer review of manuscripts. A resulting list of 36 “relevant” articles underwent a full text reading with specific attention being paid to: research question, motivation for article, method of study including details concerning study participants, relevant results and discussions, discussion of peer review and ECRs, and possible motivations for author bias. Of the articles that were found to be “not relevant” to the topic of ECR participation in the peer review of manuscripts for publication in a journal, most discussed other forms of peer review outside the scope of publishing manuscripts (e.g. students engaging in peer review of each others written work in a classroom setting as a pedagogical exercise).
Survey of peer review experiences and attitudes
Survey design
We designed a survey to evaluate the peer review experiences of researchers with a specific focus on ghostwriting of peer review reports. The survey was verified by the Mount Holyoke Institutional Review Board as Exempt from human subjects research according to 45CFR46.101(b)(2): Anonymous Surveys - No Risk on 08/21/2018. All survey respondents provided their informed consent prior to participating in the survey. The survey comprised 16 questions presented to participants in the following fixed order:
6 demographic questions that collected data on their professional status (current institution, field of research, and career stage) and personal information (gender identity, race/ethnicity, and citizenship status in the United States);
7 questions that collected data about their experience participating in the peer review of manuscripts for publication in a journal: these questions included 2 questions about their experience with independent reviewing vs. co-reviewing; 4 questions about receiving credit for reviewing activities; and 1 question about whether and how respondents received training in peer review
3 questions that collected data about their opinions about co-reviewing and ghostwriting as practices, regardless of whether they had personal experience with these practices: these questions included 1 question about their degree of agreement on a 5-point Likert scale (Strongly Disagree; Slightly Disagree; No Opinion; Slightly Agree; Strongly Agree) with 11 statements about the ethics and value of co-reviewing and ghostwriting; 1 question asking their opinion about why ghostwriting as a practice may occur; and 1 exploratory future direction question asking if their opinions would change if the names of peer-reviewers were made publically available (“open peer review”).
Throughout the survey, there were many opportunities to provide write-in responses in addition to the multiple choice answers. The full text of the survey can be found in the Supplementary file 2.
Survey distribution, limitations, and future directions
The survey was distributed online through channels available to the nonprofit organization Future of Research including via blog posts (McDowell and Lijek, 2018), email lists, social media, and word-of-mouth through colleagues. The main survey data collection effort was from August 23 to September 23, 2018. The survey had gathered 498 responses at the time of data analysis.
We recognize that conclusions drawn from any survey data are limited by the size and sample of the population that is captured by the survey. We sought to address this limitation first by collecting as large and geographically and institutionally diverse of a population of ECRs as possible within the month-long timeframe we set for data collection. Secondly, we wished to preemptively address the concern that our survey distribution efforts were inherently biased towards those with strong opinions on the subject and/or those who self-select to receive communication from Future of Research (e.g. listservs, Twitter followers). We therefore attempted to create a “negative control” comparison group of participants who received our survey from channels independent of Future of Research. We created a separate survey form asking identical questions and personally asked 25+ PIs known to the authors, as well as organizational collectives of PIs, to distribute this survey link to their own networks (e.g. lab members, departments). Both surveys were live during the same month-long time period; however, the PI-distributed survey gathered only 12 responses and so was not sufficient to be used in the analyses presented here. Since the goal of the second, PI-distributed survey was to be independently distributed outside of our efforts, we are not able to determine whether it garnered so few responses because of a lack of genuine distribution or because the populations it reached were not motivated to participate in the survey. Therefore any conclusions drawn from this study reflect the 498 experiences and perspectives of those individuals so moved to participate in the survey distributed by Future of Research and our results should be considered in this context. One possible future direction for this study is to reopen the survey in conjunction with publication of this manuscript in an effort to broaden and diversify the sampled population, to compare subsequent rounds of responses to our initial 498 responses, and to improve clarity on survey questions (Appendix 2).
Survey data analysis
Survey data were analyzed using Microsoft Excel, Version 16 and IBM SPSS Statistics for Macintosh, Version 25 (SPSS Inc, Chicago, IL, USA). Whenever statistical analyses were used, the exact tests and p values are reported in the appropriate figure legend and/or results text. A p value of less than 0.05 was considered significant.
Appendix 1
Literature on ECR involvement in peer review as a training exercise and examples of formalized training in peer review
Many of the articles uncovered by our systematic literature review on the topic of ECR involvement in peer review of manuscripts that did not address our desired topic of ghostwriting instead discussed ECR involvement in peer review of manuscripts as a training exercise for the ECR. Here, we summarize their findings. We also give some examples of formalized training in peer review.
Peer review training programs with positive outcomes
Trends from the literature indicate several components of a successful peer review training program and where such programs are currently in process. Authors studying the peer review training process for new reviewers tend to conclude that students of peer review learn best by participating themselves in the review process while receiving feedback from more senior reviewers. Successful training programs, in which the reviewers expressed that they had benefited from the training or were evaluated and determined to have benefited from the training, tended to include participation in several rounds of review followed by feedback and revising. Additionally, successful programs often expanded feedback by directing the expert reviewers to report on the same manuscripts as the trainees. This method provided trainees with specific written feedback as well as a pertinent example report to reference. Authors of successful program studies and authors with general policy recommendations for peer review training strategies conclude that a hands-on, iterative process of peer review training with regular and specific feedback are components which positively benefit the peer review trainee (Castelló et al., 2017; Doran et al., 2014).
Peer review training programs without positive outcomes
However, one reviewer training study in the pool generated negative results: Houry, Green and Callaham found that after a period of training involving mentorship from an expert reviewer to a new reviewer, no differences in mean reviewer quality scores between the mentored and unmentored groups was found (Houry et al., 2012). The study goes on to conclude that this similarity in group quality scores is dependent on the mentoring relationship: the expert reviewer mentors were not given any training on how to offer feedback to the trainees. This aspect of the program was deliberately constructed in an attempt to model a training program in which there would be minimal stress on the expert mentors. However, it appears that this ultimately led to an inconsequential training program. Accounting for this information, Houry et al. conclude that a mentorship program should include training and guidelines for mentor-mentee communications which allow for regular feedback from expert to trainee.
Institutions where peer review training takes place
Sources mentioned two primary institutions where training in peer review may take place. Training institutions were identified, with undergraduate education as the first opportunity for training, followed by graduate programs. Journals were the other institutions identified. The majority of training programs represented in the sources take place within a journal setting. Journal programs tend to include ECRs, such as undergraduate or graduate students, joining the editorial board for a set time period or acting as a reviewer (Castelló et al., 2017; Doran et al., 2014; Harrison, 2009; Houry et al., 2012; Navalta and Lyons, 2010; Patterson and Schekman, 2018; Picciotto, 2018).
Training may also take place within undergraduate or graduate institutions. We found one study about training in peer review in an undergraduate setting. Despite the lack of associated literature, Riehle and Hensley determined that undergraduate students are interested in learning about the peer review process (Riehle and Hensley, 2017). In a training study, the Calibrated Peer Review system was employed in two undergraduate classes to facilitate students to peer review the work of their classmates while minimizing the extra workload such an exercise might otherwise entail for the professor (Prichard, 2005). The participating students were in two separate courses, an introductory neuroscience course and a more advanced neuroscience course. Students in the advanced class did not perform better on peer review exercises than the introductory students, suggesting that until that point, advanced students had not been exposed productively to peer review practice. Authors deemed this to be a successful method for exposing undergraduate students to the peer review process while requiring a realistic time commitment from the course instructor.
While no papers were found detailing the effectiveness of peer review training within graduate institutions, several sources did indicate graduate student perceptions about their program peer review training. In a study including psychology masters and PhD students, a large proportion of participants indicated that their education had lacked in providing information on the peer review and revision process as well as information about how to practice review (Doran et al., 2014). These students indicate that this was a negative aspect of their programs, saying more opportunities for peer review practice should be made available. Authors of this article do indicate that this information may not be generalizable to graduate students as a group because the participating students were found when they pursued a journal review program, something students with adequate peer review education may not be likely to do.
Self facilitated training
Merry et al. provides a list of recommendations for ECRs to facilitate their own training of peer review (Merry et al., 2017). Authors advise working with the mentoring faculty as well as contacting journals directly to seek out peer review opportunities. If mentors are able to give consistent feedback regarding the trainees peer review, it seems that this could be a positive environment in which to learn the skill.
Roadblocks to positive outcomes in training programs
Papers discussing journal training programs cite feasibility as the largest roadblock to success (Castelló et al., 2017; Houry et al., 2012). As discussed, it is recommended that peer review training programs for ECRs feature a system which provides regular, specific feedback from expert reviewers. Such programs require high levels of labor, involving organization and time commitment from program leaders and expert reviewers. This is a significant investment for a business which may be in conflict with the desire to maximize journal profit. One possible solution presented involves student-run journals hiring increased numbers of student reviewers and editors so experience may be gained in the field (Doran et al., 2014; Patterson and Schekman, 2018). However, this solution does not address the recommendation for expert reviewers to provide feedback.
Examples of formalized training in peer review
Some graduate programs, scientific societies and journals already provide materials for training in peer review. For example, the American Chemical Society provides a free course designed by editors, researchers and publication staff (acsreviewerlab.org). GENETICS, a journal published by the Genetics Society of America, provides a peer review training program with virtual training sessions for ECR members (genetics-gsa.org/careers/training_program.shtml). Nature also provides an online course on peer review (masterclasses.nature.com/courses/205). Publons (which provides a mechanism for recording and crediting peer review activity) has an online peer review academy (publons.com/community/academy/). The Journal of Young Investigators provides training at the undergraduate level (jyi.org).
We also found a number of examples of courses and classes that train graduate students in peer review. These include: a class run by Needhi Bhalla at the University of California Santa Cruz; Class 230 in the PhD Program in Biological and Biomedical Sciences at Harvard; a class on critiquing papers at University of California San Diego; a class in the graduate program in Systems Biology at Harvard Medical School; and a requirement to review a paper as a final exam recently introduced at the Graduate Field of Biochemistry, Molecular, and Cell Biology at Cornell University. The University of Texas Southwestern is experimenting with an advanced course combining peer review, literature review, debate and commentary communication and lay-audience oriented writing. A number of journal clubs also review preprints and send comments to authors (Avasthi et al., 2018).
Appendix 2
Future directions for survey questions
Opinions on open peer review reports and public naming of reviewers
The final (16th) question of the survey provided respondents with an open comment box in which to reflect on whether any of their opinions would change if either the contents of peer review reports or the names of reviewers were openly published to journal readership. This question is a stark contrast to the rest of the survey questions, all of which instead only ask about naming co-reviewers to the journal staff and editors but not openly to the public readership. We included this question as a test balloon for future surveys that might focus on open peer review vs. traditional models of closed peer review. Due to its tangential relationship to the goals of the current study on ghostwriting and due to the question’s open-ended, write-in nature, we did not perform a systematic analysis of responses as with the other questions. A qualitative summary of responses is below.
61% of respondents chose to write a response to this question, and of these approximately one third reported that no, their opinions would not change if the peer review reports nor names of reviewers were openly shared with the public readership (it should be noted that these comments mostly, but not necessarily, endorsed the specific open peer review features suggested in the question). The remaining respondents expressed a variety of concerns, mostly surrounding the loss of anonymity of reviewers rather than what appeared to be the less controversial concept of publishing the contents of peer review reports. Respondents’ hesitations about anonymity often centered on the effect that naming ECRs and URMs might have on these vulnerable populations. These responses reflect other conversations about open peer review (Polka et al., 2018; Ross-Hellauer, 2017; Ross-Hellauer et al., 2017; Tennant, 2017) and in the context of recent data about referee behavior in open peer review (Bravo et al., 2019) warrant further analysis beyond the scope of this manuscript.
Improving clarity in survey questions
One survey question that would benefit from disambiguation in future iterations of the survey is: “Agree/Disagree: When a journal invites a PI to review, that is equivalent to the journal inviting anyone in that PI's research group with relevant expertise to contribute to the review.” It was brought to our attention by respondents’ emails and write-in comments during the survey response period that there was confusion about this statement. We had hypothesized that we might find agreement with the statement; however, we found a substantial amount of disagreement which may be due in part to the various ways the question may have been interpreted. Our intention was to determine if respondents agreed that, in practice, it could be reasonable for all engaged in journal publication and peer review to expect that an invited reviewer would have various motivations to share a manuscript with the relevant expertise in their research group, particularly in cases where a postdoc is likely more familiar with the literature or experimental techniques than a Principal Investigator overseeing a number of projects. However it became apparent that this question was quite open to various interpretations as described in write-in comments received on the question, including:
Does the respondent believe this should be the case?
Does the respondent believe this is what is actually happening?
Does the respondent consider that journals have this intent when they invite reviewers?
This therefore renders interpretation of responses to this question difficult, and so we chose not to draw conclusions from this particular result. We aim to clarify this question should there be future iterations of this survey.
Another survey question that would benefit from future disambiguation draws from respondents’ ability to select multiple responses for the question “When you were not the invited reviewer, what was the extent of your involvement in the peer review process? Please select all that apply to your entire peer review experience (e.g. across multiple manuscripts).” and indeed the inability to discern whether respondents supplying only one answer were selecting that response because that comprised the totality of their experiences, or because they selected the most common experience. In comparing responses to this question with other questions, it may be that there are analyses that are affected by the assumption that it is possible to apply one response of many to a response to another question. We attempted to preemptively disambiguate responses by asking whether respondents had “ever” experienced certain things in subsequent questions.
Data availability
Literature review results are shared in supplementary materials; De-identified source data for Figures 1 and 5 have been provided in response to editorial request. Raw data from the survey are not shared to protect respondents' confidentiality.
References
-
The characteristics of peer reviewers who produce good-quality reviewsJournal of General Internal Medicine 8:422–428.https://doi.org/10.1007/BF02599618
-
Preparing for the 21st century biomedical research job market: using census data to inform policy and career decision-makingSJS.
-
Negative emotion enhances memory accuracyCurrent Directions in Psychological Science 16:213–218.https://doi.org/10.1111/j.1467-8721.2007.00506.x
-
Doing peer review: reflections from anInternational group of postdoctoral fellowsJournal of Research Practice 13:V2.
-
Student peer review decisions on submitted manuscripts are as stringent as faculty peer reviewersAdvances in Physiology Education 34:170–173.https://doi.org/10.1152/advan.00046.2010
-
New reviewer mentoring programThe Journal of Neuroscience 38:511.https://doi.org/10.1523/JNEUROSCI.3653-17.2017
-
Writing to learn: an evaluation of the calibrated peer reviewTM program in two neuroscience coursesJournal of Undergraduate Neuroscience Education 4:34–39.
-
What do undergraduate students know about scholarly communication?: a mixed methods studyPortal: Libraries and the Academy 17:145–178.https://doi.org/10.1353/pla.2017.0009
-
Peer review: the experience and views of early career researchersLearned Publishing 30:269–277.https://doi.org/10.1002/leap.1111
-
WebsitePeer reviewers: why you Gotta be so mean?Chronicle of Higher Education. Accessed October 3, 2019.
-
Reviewer status and review quality. experience of the journal of clinical investigationThe New England Journal of Medicine 312:658–659.https://doi.org/10.1056/NEJM198503073121024
-
WebsiteA database of journal policies on peer review, co-reviewing, and preprintingAccessed October 3, 2019.
Decision letter
-
Peter RodgersSenior and Reviewing Editor; eLife, United Kingdom
-
Blanca Rodríguez-BravoReviewer
-
Sarvenaz SarabipourReviewer; Johns Hopkins University, United States
-
Francisco GrimaldoReviewer; University of Valencia, Spain
In the interests of transparency, eLife includes the editorial decision letter and accompanying author responses. A lightly edited version of the letter sent to the authors after peer review is shown, indicating the most substantive concerns; minor comments are not usually included.
Thank you for submitting your article "Co-reviewing and ghostwriting by early career researchers in the peer review of manuscripts" for consideration by eLife. Your article has been reviewed by 3 peer reviewers, and the evaluation has been overseen by myself (Emma Pewsey) as Reviewing Editor and Peter Rodgers, the eLife Features Editor. The following individuals involved in review of your submission have agreed to reveal their identity: Blanca Rodríguez-Bravo (Reviewer #1); Sarvenaz Sarabipour (Reviewer #2); Francisco Grimaldo (Reviewer #3).
The reviewers have discussed the reviews with one another and I have drafted this decision to help you prepare a revised submission. We hope you will be able to submit the revised version within two months.
-----
SUMMARY
By surveying 489 researchers, the authors show that co-reviewing performed by ECRs when their PI is invited to review manuscripts is the norm and that ghostwriting (i.e. ECRs reviewing papers in the name of someone else, often their PI) is common despite being deemed unethical by the majority of survey respondents. This is a timely study, and the manuscript is solid: the objectives of the study are clear, the methodology is explained conveniently with the help of the supplementary materials that are provided, and the discussion is broad and relevant.
ESSENTIAL REVISIONS
1. Please state if the responses to any of the questions have been analysed by gender, field (life sciences etc.), citizenship and race/ethnicity/URM of the participants. If so, please discuss any differences that can be observed.
2. Figure 3: From the survey data, PIs and journal clubs are the two most important sources of reviewer training. It is therefore worth noting in the manuscript that PIs should take this career development training and mentoring more seriously. You could potentially also mention that PREReview (https://www.authorea.com/inst/14743-prereview) and preprint journal club reviews can complement PI mentorship and help train and prepare ECRs for independent and co-reviewer roles.
3. Lines 228-231: "In 3 of 11 statements, ECRs felt significantly more strongly than PIs but still shared the same valance ". Please clarify whether any significant differences were found in the options of ECRs and PIs for the other 8 statements.
4. Line 232: Please explain why you chose to include a 'no opinion' option when asking participants how strongly they agreed with the statements. Is this a necessary option for future surveys of this kind?
5. In the discussion section, please discuss whether there is a way to incentivize journals to provide mechanisms for PIs to acknowledge the contributions of ECRs to reviews.
6. The discussion and conclusions advocate for a cultural shift, among others, but it would be nice to also discuss the change of mind that affects ECRs when they become PIs in such a hypercompetitive environment. Are there new incentives that could be introduced to encourage PIs to name their ECR co-reviewers?
7. Figure 7: Please ensure that each reason given in this graph is addressed in the "Discussion" of the manuscript and offer potential practical solutions for each to the community. For instance, in response to the statement that "73% of respondents reported that they had not discussed being credited for co-reviewing with their PIs" (line 269), you could mentions that better mentoring practices help overcome such miscommunication.
OPTIONAL SUGGESTIONS
a. You may wish to consult the results of the project "Early Career Researchers the Harbingers of Change?" (http://ciber-research.eu/harbingers.html), in particular the publication focused on peer review: Blanca Rodríguez-Bravo, David Nicholas, Eti Herman, Chérifa Boukacem-Zeghmouri, Anthony Watkinson, Jie Xu, Abdullah Abrizah, Marzena Świgoń Peer review: The experience and views of early career researchers Learned Publishing, 30(4), pp.269-277.
One of the recommendations of the article is that journals should acknowledge that peer review is often performed by co-reviewers, and facilitate it for training purposes, since in this way the principal investigators will have no excuse for not giving credit to their collaborators.
b. You may wish to discuss ways of reducing the need for co-reviewing. For example, starting a database of ECR reviewers can help train potential reviewers and ensure a productive peer review process. Recommending this list in the manuscript to journals/associate editors may help to practically address ECR ghost reviewing by contacting ECRs directly and acknowledging their expertise. An example of such a list is: https://ecrlife.org/2018/03/02/boosting-ecr-involvement-in-peer-review/
https://doi.org/10.7554/eLife.48425.024Author response
We repeat the reviewers’ points here in italic, and include our replies point by point, as well as a description of the changes made, in Roman.
ESSENTIAL REVISIONS
1. Please state if the responses to any of the questions have been analysed by gender, field (life sciences etc.), citizenship and race/ethnicity/URM of the participants. If so, please discuss any differences that can be observed.
In addition to the analysis by career stage described in the paper (ECRs vs PIs), we also performed preliminary analyses by gender, field, citizenship and race/ethnicity of the participants because we were also very interested in these comparisons. However, there were no striking conclusions to be drawn from these demographic analyses: differences were not statistically significant nor of great magnitude and some demographic subsets (eg. U.S. domestic people of color in the life sciences) were underpowered with small n values. To report these analyses thoroughly without cherry picking would significantly increase the length of an already long manuscript. Early drafts of our preprint included some of these descriptions of demographic data and were flagged by colleagues as tangential and distracting. We are considering how to share these negative data more completely, e.g. in a preprint or blogpost, and believe they are beyond the scope of this manuscript. We hope the reviewers find this reasonable and acceptable.
2. Figure 3: From the survey data, PIs and journal clubs are the two most important sources of reviewer training. It is therefore worth noting in the manuscript that PIs should take this career development training and mentoring more seriously. You could potentially also mention that PREReview (www.authorea.com/inst/14743-prereview) and preprint journal club reviews can complement PI mentorship and help train and prepare ECRs for independent and co-reviewer roles.
We completely agree with the reviewers’ point that PIs must take this training responsibility more seriously - thank you! We also concur that gaining review experience through preprint (or regular) journals clubs is an important complement to PIs’ mentorship. Note that preprint journal clubs were already mentioned briefly in Supplementary File 3: Examples of Formalized Training in Peer Review, citing Avasthi et al. 2018. This section also references a template for how to integrate preprint journal clubs into the graduate classroom provided by Dr. Needhi Bhalla at the University of California Santa Cruz that is included in our preprint.
We added new language on line 297 to better highlight our recommendations for training in peer review, and in particular the role that PIs and journal clubs serve:
“Given their key role as the main source of training in peer review, it is important that PIs make deliberate efforts to teach their trainees this skill and provide feedback on ECRs’ work. Mentorship from PIs may be complemented with, but not replaced by, journal clubs where ECRs gain experience in reviewing published manuscripts or preprints (e.g. Avasthi et al. 2018 and PREReview https://www.authorea.com/inst/14743-prereview). Since these ad hoc reviewing experiences with PIs and journal clubs may vary in their availability and quality, all ECRs should be offered standardized, evidence-based training in peer review. For example, peer review courses that are compulsory and ubiquitous in graduate schools would ensure that all PhD-holders are appropriately trained to perform constructive peer review.”
3. Lines 228-231: "In 3 of 11 statements, ECRs felt significantly more strongly than PIs but still shared the same valance ". Please clarify whether any significant differences were found in the options of ECRs and PIs for the other 8 statements.
We added the following sentence immediately after the one quoted above at line 174: “For the remaining 8 statements, ECRs and PIs did not significantly differ in their opinions.”
4. Line 232: Please explain why you chose to include a 'no opinion' option when asking participants how strongly they agreed with the statements. Is this a necessary option for future surveys of this kind?
A ‘no opinion’ option is not necessary for future surveys, but it is our preference for this and future surveys. The ‘no opinion’ option with the Likert scale provides respondents with an opportunity to demonstrate their true neutral opinion, rather than forcing them to agree or disagree. This creates cleaner, more interpretable data (see references below on evidenced-based survey design). If a respondent has no opinion on a topic but is not offered a “no opinion” option in the survey, they would either have to a) choose an inauthentic opinion from the survey choices or b) leave the question blank. Choosing an opinion even when one has none (a forced choice) would introduce noise into the survey results. Leaving the question blank would create a missing data point. A missing data point would not be sufficient for us to interpret it as “no opinion” since there are other reasons why a question might be skipped (e.g. accidental oversight, or “I don’t understand”).
We included a “no opinion” option in our survey to be inclusive of the reality that respondents may feel neutrally about certain aspects of co-reviewing and ghostwriting. We wanted our dataset to be able to capture that indifference, should it exist, as much as we wanted to know whether respondents’ agreed or disagreed. For example, we hypothesized that PIs’ indifference about the value of naming co-reviewers might cause them to omit co-reviewer names. Evidence in support of that hypothesis is described at line 410 of the manuscript:
“ECRs agree more strongly than PIs that there is value in receiving credit or being known to the journal editorial staff as a co-reviewer (Table 3), perhaps because invited reviewers are already known as experts in their field. In the words of one write-in response: “PI surprised I would be interested in being acknowledged, and seemed like too much trouble to acknowledge my contribution. There was no box in the online form to declare it.” PIs’ ambivalence for due credit for co-reviewers likely derives from a position of relative privilege. When people of privilege are the only participants in decision-making (e.g. on journal editorial boards), they may create policies that fail to consider differing values of less privileged members of the community, like ECRs.”
Examples of research on the utility of the midpoint “no opinion” option:
· Chyung, S. Y., Roberts, K., Swanson, I., & Hankinson, A. Evidence-based survey design: The use of a midpoint on the Likert scale. Performance Improvement Journal. 2017; 56(10).
· Kulas JT, Stachowski AA. Middle category endorsement in odd-numbered Likert response scales: associated item characteristics, cognitive demands, and preferred meanings. J. Res. Pers. 2009;43:489–493.
· Nadler, J. T., Weston, R., & Voyles, E. C. Stuck in the middle: The use and interpretation of mid-points in items on questionnaires. The Journal of General Psychology. 2015; 142(2), 71-89.
5. In the discussion section, please discuss whether there is a way to incentivize journals to provide mechanisms for PIs to acknowledge the contributions of ECRs to reviews.
We strongly agree with the reviewers that systematic change will require thoughtful reconsideration of the current incentives for journals, PIs, etc. To this effect, we have added the following text at line 393 of the manuscript:
“It is in journals’ best interest to provide mechanisms for ECR co-reviewers to be easily named, as described recently by an Elsevier editor (Mehmani, 2019). Editors take care to assign appropriate reviewers that lack conflicts of interest, an effort that is stymied by ghostwriting. Editors may also see an increase in accepted invitations to review once PIs feel free to share the burden of reviewing with their trainees without ethical concerns. Journals would also benefit from enlarged reviewer pools by including previous ECR co-reviewers and/or consulting existing lists of ECR reviewers (https://ecrlife.org/2018/03/02/boosting-ecr-involvement-in-peer-review). Ultimately, the burden of proof for ethical behavior falls on journals who are the managers of the peer review process. The status quo where ghostwriting abounds and is widely considered unethical is a strong disincentive for maintaining current policies.”
More generally, we note that reviewers’ comments #5-7 and optional comment #2 all make requests to add language about how we might change the system. We are hesitant to include too many opinions about how to incentivize policy and behavioral changes in this manuscript in an effort to separate the unbiased reporting of data (in this manuscript) from personal opinions. We plan to share our recommendations for policy changes in a companion opinion article, currently in preparation, that draws on the data reported here. We are happy to provide these reviewers with a draft of that opinion manuscript should they wish to see it.
6. The discussion and conclusions advocate for a cultural shift, among others, but it would be nice to also discuss the change of mind that affects ECRs when they become PIs in such a hypercompetitive environment. Are there new incentives that could be introduced to encourage PIs to name their ECR co-reviewers?
We don’t interpret these data to mean that ECRs change their minds on these issues when they become PIs. Instead, it seems as though PIs simply re-create their own experiences because that is “how it is done” (see Table 5). So if someone was not named to the journal when they were an ECR co-reviewer, then they do not name their own trainees as co-reviewers once they become a PI, even if they think that is how it should be done. Junior/pre-tenure PIs in particular may hesitate to do anything differently for fear it might reflect poorly upon them during their own tenure evaluation.
In our opinion, the best way to incentivize PIs to do something is to make it a requirement of grant funding or tenure. In general, mentorship activities are not well accounted for in these evaluation processes and ought to be. For example, perhaps PIs who pay ECR salaries from grant budgets should demonstrate in grant progress reports how those ECRs are being trained, which might include a list of which trainees were listed as co-reviewers by journals or Publons.
We added this text to line 333 of the manuscript: “PIs could be further incentivized to name ECR co-reviewers by accounting for this practice in PIs’ evaluation for grant funding or tenure. For example, perhaps PIs who pay trainees’ salaries from grant budgets should demonstrate in grant progress reports how those ECRs are being trained, which might include a list of which trainees were listed as co-reviewers.”
Journals have the power to make naming co-reviewers so easy and obvious during the peer review report submission process that it may not require further incentivization for the PI. In other words, we don’t think that a carrot needs to be added to the system, but rather the sticks (barriers to good behavior) need to be removed. If journals make naming co-reviewers a clear, standard expectation of all invited reviewers, then it seems unlikely that a PI would actively avoid naming a co-reviewer. In those rare cases, perhaps journals could include a whistleblower hotline for a ghostwriter to alert the journal in cases where names were intentionally and maliciously withheld. A particularly strong disincentive to hiding ECR contributions is that this has been used as a pretext to dismiss faculty where ECRs have been shown confidential grant applications to assist with reviews (e.g. https://www.bloomberg.com/news/features/2019-06-13/the-u-s-is-purging-chinese-americans-from-top-cancer-research).
7. Figure 7: Please ensure that each reason given in this graph is addressed in the "Discussion" of the manuscript and offer potential practical solutions for each to the community.
For instance, in response to the statement that "73% of respondents reported that they had not discussed being credited for co-reviewing with their PIs" (line 269), you could mentions that better mentoring practices help overcome such miscommunication.
We have made sure that all of the possible responses in (previously Figure 7) now Figure 6 are discussed with possible solutions (noted by “→” below) on the following pages:
No practical mechanism pg 17
“A lack of a mechanism (such as a text box), with language demonstrating expectations that co-reviewers be listed) to include this information in the peer review report submission process.”
→ add textbox along with clear language upon submission of the peer review report that invited reviewers are expected to name co-reviewers or acknowledge the review was only performed by the invited reviewer
Prohibitive journal policies pg 16
“A belief that reviews should only be done by the invited reviewer, and not by, or with assistance from anyone else.”
“A belief that including co-author information would demonstrate that the PI breached the confidentiality of the manuscript.“
→ remove or reform current prior permissions/confidentiality policies
Co-reviewers don’t deserve authorship: pg. 15
“A belief that only the invited reviewer deserves authorship, even when others contribute ideas and/or text to the review report.”
“A belief that there is no strong ethical reason to add co-authors names.”
→ journals and PIs should realize that respondents think that co-reviewers do deserve authorship and adjust their policies/behaviors to reflect this by routinely naming all co-reviewers
Power dynamics pg 15
“Some ECRs may not be comfortable asking for co-authorship.”
→ PIs should initiate these discussions
Protectionism pg 19
“Some co-reviewers want to be able to write critical reviews anonymously.”
“A belief that keeping ECR names off of peer review protects ECRs during a vulnerable time in their career.”
→ Not a major driver of ghostwriting, no changes needed
Denialism and/or ignorance pg. Pg 19
“A belief that ghostwriting does not occur: PIs are the only people that contribute to peer review reports.”
“A belief that ghostwriting does not occur: everyone always provides the names of all contributing authors to the editorial office.”
→ Not a major driver of ghostwriting, no changes needed
The data referenced by the reviewers’ comment above from line 269 (in first submission, now line 204 in main manuscript without embedded figures) is from Table 4 and we have added the point about mentorship to the discussion of that table at line 332: “If ghostwriting arises from PIs and ECRs simply not thinking to include co-reviewer names vs. intentional withholding of names, then building awareness should encourage more conversation about this issue and better mentoring practices will help overcome such miscommunication.”
OPTIONAL SUGGESTIONS
a. You may wish to consult the results of the project "Early Career Researchers the Harbingers of Change?" (http://ciber-research.eu/harbingers.html), in particular the publication focused on peer review: Blanca Rodríguez-Bravo, David Nicholas, Eti Herman, Chérifa Boukacem-Zeghmouri, Anthony Watkinson, Jie Xu, Abdullah Abrizah, Marzena Świgoń Peer review: The experience and views of early career researchers Learned Publishing, 30(4), pp.269-277. One of the recommendations of the article is that journals should acknowledge that peer review is often performed by co-reviewers, and facilitate it for training purposes, since in this way the principal investigators will have no excuse for not giving credit to their collaborators.
Thank you, we have added that citation and the following text on line 377 of the manuscript: “Journals should acknowledge that peer review is often performed by ECR co-reviewers and remove barriers that prevent ECRs from being named to the editor (Rodríguez-Bravo et al., 2017).”
b. You may wish to discuss ways of reducing the need for co-reviewing. For example, starting a database of ECR reviewers can help train potential reviewers and ensure a productive peer review process. Recommending this list in the manuscript to journals/associate editors may help to practically address ECR ghost reviewing by contacting ECRs directly and acknowledging their expertise. An example of such a list is: https://ecrlife.org/2018/03/02/boosting-ecr-involvement-in-peer-review/
We have incorporated this into the response to point #5 above, now on line 397 of the manuscript. “Journals would also benefit from enlarged reviewer pools by including previous ECR co-reviewers and/or consulting existing lists of ECR reviewers (https://ecrlife.org/2018/03/02/boosting-ecr-involvement-in-peer-review).”
Additional changes made to manuscript in response to editorial request
(Added on September 22, 2019: Please see bottom of letter for list of additional changes to the manuscript that were requested by the editors. All of these changes have been made and the line numbers in the letter to reviewers’ have been updated accordingly.)
AUTHOR CONTRIBUTIONS
- author contributions text has been removed from the manuscript. There is a still a short section about financial disclosures.
BOXES
- converted all Boxes in into tables
- removed the top title bar from each
- reordered tables to reflect the order of first mention in manuscript text
FIGURES
- Combined old Figures 4&5 into new Figure 4A and 4B
- Added ratios to figure legend for new Figure 4
- Renumbered remaining figures
- Added requested text/ratios to Figure 1 legend
SUPPLEMENTAL FILES
- new Supplemental Files 1-3 created as requested
- formerly Supplemental Figure 1 (flow chart of literature review) changed to Figure 1 Supplement, as requested
SOURCE DATA
- deidentified source data provided for Figure 1 (demographics) and Figure 5 (formerly Figure 6 as requested, opinions)
https://doi.org/10.7554/eLife.48425.025Article and author information
Author details
Funding
Open Philanthropy Project (Future of Research - General Support)
- Gary S McDowell
The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.
Acknowledgements
The authors thank Dr Jessica Polka, Dr Adriana Bankston, Dr Carrie Niziolek and Dr Chris Pickett for their constructive comments on this manuscript. GSMD thanks the Open Philanthropy Project for providing support for his role at the Future of Research.
Ethics
Human subjects: The survey was verified by the Mount Holyoke College Institutional Review Board as Exempt from human subjects research according to 45CFR46.101(b)(2): Anonymous Surveys - No Risk on 08/21/2018. Prior to participating in the survey, all respondents were presented with the following statement and then selected a checkbox to provide their informed consent: "By choosing to submit answers to this survey, you thereby provide your informed consent to voluntarily share your experiences and opinions with the researchers, who intend to publish a summary of the results of the survey but not the raw data with participants' individual demographic information." See Supplementary Materials for full Statements of Disclosure, Ethics and Informed Consent.
Publication history
- Received:
- Accepted:
- Version of Record published:
Copyright
© 2019, McDowell et al.
This article is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use and redistribution provided that the original author and source are credited.
Metrics
-
- 7,266
- views
-
- 407
- downloads
-
- 39
- citations
Views, downloads and citations are aggregated across all versions of this paper published by eLife.
Download links
Downloads (link to download the article as PDF)
Open citations (links to open the citations from this article in various online reference manager services)
Cite this article (links to download the citations from this article in formats compatible with various reference manager tools)
Further reading
-
- Microbiology and Infectious Disease
The telomerase RNA component (Terc) constitutes a non-coding RNA critical for telomerase function, commonly associated with aging and pivotal in immunomodulation during inflammation. Our study unveils heightened susceptibility to pneumonia caused by Staphylococcus aureus (S. aureus) in Terc knockout (Tercko/ko) mice compared to both young and old infected counterparts. The exacerbated infection in Tercko/ko mice correlates with heightened inflammation, manifested by elevated interleukin-1β (IL-1β) levels and activation of the NLR family pyrin domain containing 3 (NLRP3) inflammasome within the lung. Employing mRNA sequencing methods alongside in vitro analysis of alveolar macrophages (AMs) and T cells, our study elucidates a compelling correlation between Tercko/ko, inflammation, and impaired T cell functionality. Terc deletion results in compromised T cell function, characterized by dysregulation of the T cell receptor and absence of CD247, potentially compromising the host’s capacity to mount an effective immune response against S. aureus. This investigation provides insights into the intricate mechanisms governing increased vulnerability to severe pneumonia in the context of Terc deficiency, which might also contribute to aging-related pathologies, while also highlighting the influence of Terc on T cell function.
-
- Microbiology and Infectious Disease
Phage-derived peptidoglycan hydrolases (i.e. lysins) are considered promising alternatives to conventional antibiotics due to their direct peptidoglycan degradation activity and low risk of resistance development. The discovery of these enzymes is often hampered by the limited availability of phage genomes. Herein, we report a new strategy to mine active peptidoglycan hydrolases from bacterial proteomes by lysin-derived antimicrobial peptide-primed screening. As a proof-of-concept, five peptidoglycan hydrolases from the Acinetobacter baumannii proteome (PHAb7-PHAb11) were identified using PlyF307 lysin-derived peptide as a template. Among them, PHAb10 and PHAb11 showed potent bactericidal activity against multiple pathogens even after treatment at 100°C for 1 hr, while the other three were thermosensitive. We solved the crystal structures of PHAb8, PHAb10, and PHAb11 and unveiled that hyper-thermostable PHAb10 underwent a unique folding-refolding thermodynamic scheme mediated by a dimer-monomer transition, while thermosensitive PHAb8 formed a monomer. Two mouse models of bacterial infection further demonstrated the safety and efficacy of PHAb10. In conclusion, our antimicrobial peptide-primed strategy provides new clues for the discovery of promising antimicrobial drugs.