Many people have the sense that the peer-review system is severely stressed. Myriad problems are pitched at peer review—there are too many requests from editors, editors cannot get responses to invitations, some reviewers shirk after they have agreed to review, and authors do not like the feedback (e.g., Borer Reference Borer1997). For instance, writing in The Monkey Cage, Gelbach (Reference Gelbach2013) suggests, “The peer-review process, if not broken, is seriously under strain. Editors are forced to make hasty decisions based on imperfect signals from referees. Referees, in turn, are overburdened with review requests” in what Wilson (Reference Wilson2011) calls the “tragedy of the reviewer commons.” But are we overburdened by reviews? To begin an informed conversation about reviewing workloads, we need to answer two simple questions: How much do political scientists review? And who bears the burden of reviewership?
Problems with the peer-review system may be compounded if political scientists do not find value in the system or in their contribution to it. But the problem may be simpler if we value peer reviewing and the contributions we make through it, but simply do not feel recognized by relevant evaluation bodies for tenure and promotion. Therefore, another valuable addition to disciplinary conversations about peer reviewing would also address our beliefs regarding the value of peer reviewing to us and to institutional decision makers.
DESIGN AND DATA
In October 2013, a survey link was sent by e-mail to 3,002 APSA members randomly sampled from the membership with a PhD. Footnote 1 After three reminders (the survey was open for a month), 823 began the survey and 607 completed the instrument for a completion rate of 22.3% (not counting the 275 e-mails that bounced). The resulting sample reflects the gender balance of APSA members with a PhD (31.7 % female in the sample compared to 33.2% in the APSA population) and slightly overrepresents those in PhD-granting institutions (46.5% in the sample compared to 43.3% in APSA), but is not reflective in other ways. APSA members are 65% white whereas the sample is 82% white; 19% of APSA members are associate professors compared to 29% in the sample; 26% of APSA membership has attained the rank of full professor compared to 33% in the sample. In the following text, I use several schemes for dealing with the response bias. For descriptive statistics, I weight the data by race before presenting the statistic by rank. Otherwise, I include relevant demographics, including race, in models (i.e., “model based weighting”).
RESULTS
Although the survey was focused on reviewing for journals, figure 1 reports the diversity of outlets for peer reviewing in the discipline by rank. About 90% of all tenured or tenure-track faculty report reviewing for a journal in the past year. The 79% of non-tenure track faculty who report reviewing is significantly lower than the tenure track faculty, but so is the 87% figure for full professors compared to assistants. All other forms of peer review attract a smaller portion of those with a PhD, and full professors lead the way.
Among those who reviewed for journals, there is substantial diversity in the amount of reviewing that varies in predictable ways. Figure 2 shows a scatterplot of faculty at PhD-granting and non-PhD-granting universities at each rank with means and standard deviations shown. Reviewership is almost universally low except for faculty of higher rank at PhD-granting institutions. The averages among non-tenure track faculty and faculty at non-PhD-granting universities are statistically indistinguishable (2.5–3 reviews per year). Whereas the reviewer loads vary among faculty ranks within PhD-granting institutions, the differences are not large and the averages are not high. Assistant professors averaged 5.5 reviews, associate professors averaged 7.0 reviews, and full professors averaged 8.3 reviews in the past year. From these data, Footnote 2 it is difficult to conclude that the average PhD holder in political science is overrun with reviews.
Predicting Reviewership
Reviewing is a function of being asked, and being asked is a function of reputation. We might hope that the standard route to expertise in a subfield is to publish and publish well. Each recent article should add to the likelihood of completing reviews. Publishing a book should also add incrementally to perceived expertise. Editors and their assistants notice who is getting published, but they also rely on the heuristic of the citation list of the piece under review. Moreover, publishing in a journal adds authors to the registry, and some editors ask for reviews before they let you out from under their decision thumb.
Publishing is not the only route to reputed expertise. Faculty who populate the university highways are more likely to be accessible compared to those on the byways. Simply being networked into the discipline should encourage reviews, for which the heuristic of being employed in a department with a PhD program is the proxy I use. Footnote 3 Rank rewards the productive scholar, especially in PhD-granting institutions, thus associate and full professors should have the highest reviewing load. As we have seen in figure 2, that link does not appear to hold for those outside PhD-granting institutions—rank is not related to reviewing rates. Of course, those who take on official roles with journals as editors or board members should expect to bear a higher burden.
Reviewing is a function of being asked, and being asked is a function of reputation.
In part because of the survey response bias, the model includes dummies for men and whites. If the discipline’s power structure is dominated by white men, then there may be a reputational bias in favor of white men, or there may be a skew induced by the (lack of) attention paid to work that women and non-whites tend to produce. Lastly, because reviewers have the choice to accept the request, I include an index of perceived reviewing efficacy, expecting that more efficacy is related to increased reviewership.
The estimates in table 1 from a negative binomial model confirm many of these expectations. Footnote 4 By far the dominant effect in the model is the role of journal publication. Each publication (in the past 3 years) adds .8 reviews, although it interacts with employment in a PhD-granting department. The relationship (see figure 3) shows the significant boost of about 3 reviews from being in a PhD program at low levels of publication. That gap decays to become indistinguishable at higher levels of exposure through publication. Publication of a book also boosts reviewing by .8, suggesting books have the reputational equivalent of an article. Footnote 5
Source: 2013 APSA Peer Review Survey.
Model Statistics: N=503, Pseudo R2=.12, LR test (α=0) p<.01
Associate and assistant professors are indistinguishable from full professors (the excluded category), all else constant. Footnote 6 It is no surprise that editors and board members review more—about 3.4 reviews more (I am assuming that editors did not include decision making work for the journal). Women review no less than men, but whites review about 1.4 more pieces, although it is not especially clear why.
Accepting Reviewer Requests
While the digital age has vastly simplified the activities of everyone involved with journals, some evidence indicates that this complicates negotiations with reviewers. In other words, it is easier to ignore an e-mail solicitation than a mailed manuscript. The evidence is scattered and infrequently reported by editors. In 1988, the earliest instance (in JSTOR) when I can find the editors reporting such statistics, the American Political Science Review (APSR) received reviews from an enviable 83.3% of those solicited (Patterson, Ripley, and Trish Reference Patterson, Ripley and Trish1988, 911). The Journal of Politics (JOP) editorial report from 2009 (Leighley and Mishler Reference Leighley and Mishler2009, 3) suggests 56% of solicitations were accepted, although only three-quarters of these resulted in submitted reviews (for a 42% “response rate”). Their 2011 report indicated 66% of solicitations were accepted and again one-quarter failed to result in a review (50% response rate overall). American Journal of Political Science (AJPS) reported a 59% initial acceptance of the reviewer task in 2007–08 (Stewart Reference Stewart2008), although a non-specified “some” failed to report. The response rate in 2009 for the AJPS was 49% (Stewart Reference Stewart2009), 56% in 2010 (Wilson Reference Wilson2011), 57.4% in 2011 (Wilson Reference Wilson2012), and 61% in 2012 (Wilson Reference Wilson2013). During the past two years at Politics & Religion, we have had a 45% response rate. It would be useful for editors to report this statistic regularly to permit comparison.
The 70% average response rate reported in these data, therefore, looks like voter turnout estimates from the American National Election Study—it is probably too high. Yet it varies in reasonable ways by rank and institution as figure 4 shows. Assistant professors in non-PhD-granting institutions have the highest “uptake rate” (87%) and full professors at PhD- granting institutions have the lowest (65%). There is no significant difference overall between those in PhD-granting institutions (71%) and not (74%). Footnote 7 Nonetheless, there is marked variance within and across ranks. Twenty-eight percent of full professors accept less than half of the requests to review. Now, of course, the lower uptake rate could be a function of heavy reviewing loads, the prestige of journals making “the ask,” and other factors.
Are higher uptake rates a function of being asked or saying yes? The evidence in table 2 suggests that the uptake rate is independent of the number of review requests received. Footnote 8 Instead, the uptake rate corresponds to institutional incentives and productivity. Those scholars who publish more articles (not books) accept reviewer requests more often—1% more for each article in the last 3 years. The untenured accept reviews at higher rates, too. There is some statistically marginal evidence that those scholars who feel efficacious about reviewing accept at higher rates.
Source: 2013 Peer Review Survey.
Model Statistics: N=465, Adj. R2=.07, RMSE=.24
This evidence is important because it suggests where the problem lies—full professors at PhD-granting institutions have published the most and are called to review the most (on average 21 times). Assistant professors accept at higher rates, but are asked less frequently (9 times a year at PhD-granting institutions and 4 times outside of PhD-granting institutions) and are more difficult to identify because they have published less. The problem appears not to be a function of being overrun with review requests, but rather a structural problem of searching beyond the usual suspects.
Beliefs about Peer Reviewing
From all of the complaints about reviewing, one might be lead to believe that it is a burden with little value. The respondents to this survey indicate otherwise. Figure 5 reports results from a number of items that bear on what value peer review (“PR” in figure 5) holds and for whom. Belief in the value of peer review is nearly unanimous (95% agree or strongly agree). Very few agree that peer review is rarely worth the time that can be better spent on other things, although a quarter are on the fence. Very large portions (80%) see peer reviewing as a way to keep up with current research. Of course, there is a potential dark side to being in the loop—the prospect of keeping research at the gate. A plurality (45%) agrees that gatekeeping is an important reason to peer review. There is suggestive evidence that those scholars in PhD programs who publish at very high rates are more likely to agree with the importance of gatekeeping, but it is not a statistically significant effect. Men are more likely to indicate the importance of gatekeeping (by about .3).
Whereas most believe that being asked to peer review is a measure of professional stature (only 10% explicitly disagree), relatively few (38%) believe that they are recognized for the effort and a third explicitly disagree that they are recognized for it. It is important to realize that the amount of reviewing is not related to recognition (r = .04, p = .30), so faculty continue to review despite the lack of recognition. Perhaps most importantly, a plurality (41%) believes that reviewing is not counted toward tenure and promotion. This sentiment is related to the amount of reviewing, but in the inverse—individuals who do more reviewing are more likely to believe it is not rewarded. It is not surprising to learn that vast numbers of respondents (79%) believe that peer reviewing loads should count as service in tenure and promotion decisions.
It is not surprising to learn that vast numbers of respondents (79%) believe that peer reviewing loads should count as service in tenure and promotion decisions.
CONCLUSION
Peer reviewing makes the publishing world go around. Most scholars believe that peer review is valuable service to the discipline and perform these duties despite not being recognized for the effort. If it is valuable, then this time consuming, important process should count for something. Reviewing for presses is easy to accept: they pay. But reviewers for journals feel insufficiently recognized, although they review anyway and at essentially the same rate regardless of the number of times asked. The year-end list at the back of the print journal is marginal compensation. Does anyone actually find the year-end reviewer list now that many individuals access journal articles online and through archives?
One solution is to bring peer review into the light. We need a system that collects data from journals (and other institutions) on who reviews and presents them in a format usable by scholars and their relevant tenure/promotional bodies. Footnote 9 Such a system would provide verifiable data that overcome our suspicions of reviewing self-reports as cheap talk. Rather than the typical career list of journals reviewed for, yearly statistics could easily be added as a meaningful indicator in annual reports. Moreover, such data would enable universities to gain increased exposure from the extent of their faculty’s demonstrated reputation for expertise through peer reviewing.
If universities are going to reward faculty for their reviewing contributions, they need a sense for what is normative or at least common in our reviewing practices. These data provide one look at this question, indicating that it is common for professors at all ranks outside of PhD-granting institutions to perform about 2.5 reviews a year. For those in PhD granting institutions, reviews vary by rank and increase by about 1.5 reviews per rank from 5.5 for assistant professors to 8.3 for full professors. Only 10% of this sample is doing 1 review a month or more. It is possible that political scientists believe even this workload is too high, but from this look, reports of reviewing fatigue are coming from a highly selective set of faculty.
ACKNOWLEDGMENTS
I would like to thank former executive director of APSA, Michael Brintnall, for agreeing to support this initiative and Jennifer Diascro for a great deal of assistance and a critical eye in shaping and fielding the survey. I would also like to thank Mike Brady, Jennifer Hochschild, Polly Karpowicz, Scott McClurg, Dave Peterson, Anand Sokhey, Kaare Strom, and Jenny Wolak for their close reading of the survey instrument and thoughtful feedback. Thanks also to Mike Brady, Ryan Burge, Andy Lewis, and Jake Neiheisel for their thoughtful feedback about this article.