- Research article
- Open Access
What do the JAMA editors say when they discuss manuscripts that they are considering for publication? Developing a schema for classifying the content of editorial discussion
BMC Medical Research Methodology volume 7, Article number: 44 (2007)
In an effort to identify previously unrecognized aspects of editorial decision-making, we explored the words and phrases that one group of editors used during their meetings.
We performed an observational study of discussions at manuscript meetings at JAMA, a major US general medical journal. One of us (KD) attended 12 editorial meetings in 2003 as a visitor and took notes recording phrases from discussion surrounding 102 manuscripts. In addition, editors attending the meetings completed a form for each manuscript considered, listing the reasons they were inclined to proceed to the next step in publication and reasons they were not (DR attended 4/12 meetings). We entered the spoken and written phrases into NVivo 2.0. We then developed a schema for classifying the editors' phrases, using an iterative approach.
Our classification schema has three main themes: science, journalism, and writing. We considered 2,463 phrases, of which 87 related mainly to the manuscript topic and were not classified (total 2,376 classified). Phrases related to science predominated (1,274 or 54%). The editors, most of whom were physicians, also placed major weight on goals important to JAMA's mission (journalism goals) such as importance to medicine, strategic emphasis for the journal, interest to the readership, and results (729 or 31% of phrases). About 16% (n = 373) of the phrases used related to writing issues, such as clarity and responses to the referees' comments.
Classification of editorial discourse provides insight into editorial decision making and concepts that need exploration in future studies.
No public knowledge is gained from scientific and biomedical research unless the study methods and results are properly written up and disseminated, typically by publication in a widely available journal. The decision to publish study findings involves many individuals and groups, including the investigators, the designated authors, the peer reviewers, the journal editors, and editor-in-chief. Some parts of the decision-making process are more easily studied than others, in part because traditionally, publishing decisions have been confidential. From substantial existing research, we know that investigators, not journal editors, are the ones mainly responsible for the decision to publish, and that this decision is related to the direction and strength of study findings (publication bias) [1–6].
There are many fewer studies of the editorial decision-making process [7–10]. Editors cannot discuss or make decisions about manuscripts they never receive. A 2002 study of editorial decision-making at the US general medical journal JAMA found that editors favor high quality science, and any bias against negative results is either small or does not exist . Studies of editorial decision-making have mainly addressed issues of bias, but not other factors related to journal mission, responsibility, and financial health. For example, one might ask whether editors are interested in publishing "newsworthy" results: those that are completely unexpected, that confirm a long held belief, or that settle a controversy once and for all. One might also be interested in whether a single set of rules could be devised to describe acceptance decisions. Editors may one week reject a report because it is similar to another one recently published, but may the next week accept a paper because it is similar to a recent contribution in their journal.
JAMA, a major, weekly, general medical journal in the United States, has a broad mission and low acceptance rate. We decided to attend journal editorial decision-making meetings to learn more about what was discussed, to allow construction of quantitative survey instruments that would capture a richer picture of factors influencing publication decisions. The discussions we listened to relate particularly to understanding the mechanisms used when editors must choose only 5% to 10% from a large pool of submitted manuscripts.
In an effort to identify previously unrecognized aspects of editorial decision-making, we explored the words and phrases that one group of editors used during their meetings. We did not use formal qualitative methods to analyze the information collected, rather our goal was to help to generate hypotheses for those doing research on editorial decision making and publication practices.
In 2003, 5064 manuscripts (not including letters) were submitted to JAMA. These were distributed to the editors, almost all physicians, who reject at least half outright without obtaining comments from peer reviewers. When manuscripts are returned from peer reviewers, typically two or three, editors can reject directly or bring the manuscript to an editorial meeting. Some editors bring most of their papers to the table for a decision and others bring only papers they feel have a very good chance of acceptance. Because manuscripts are assigned to a single editor for review, only one editor has typically read a manuscript before the meeting. In 2003, approximately 940 manuscripts were brought to the meeting for discussion. All discussions at the meetings are kept in strict confidence, by JAMA's standing rules.
One of us (KD) attended 12 twice weekly editorial meetings, as a visitor, at the JAMA offices in Chicago, Illinois, USA in January and February 2003 (some meetings were missed by KD because of scheduling conflicts), and took notes on the discussion surrounding 102 manuscripts (two related manuscripts were discussed as one, and so we counted them as a single manuscript). Her notes were not verbatim transcripts of the meeting's discussions. The meeting attendees varied somewhat from meeting to meeting, but typically comprised about eight in-house editors, including editors with content, managing, and statistical responsibilities, and the Editor-in-Chief. Other editors (including DR who attended 2 of the 12 meetings in person, and 2 by phone) attended by teleconference if a manuscript for which they were responsible was being discussed. Editors volunteered one-by-one, but in no particular order, to discuss the manuscripts for which they were responsible. Anecdotal reports from editors attending the manuscript meetings is that the meetings attended as part of this study were representative of other meetings.
The discussion of each manuscript began with a description of the paper topic and study characteristics, with details added as necessary. The presentation progressed to comments made by the peer reviewers. The vast majority of presented manuscripts had completed at least one round of peer review.
The note-taker (KD) recorded words and phrases spoken by the editors in the context of each manuscript discussed, the time each discussion took, and the comments and publication recommendation of each of the peer reviewers. In addition, at the end of discussion about each manuscript, editors attending the meeting completed a form on which they listed the reasons they were "inclined to proceed to the next step towards review and/or acceptance" and reasons they were not. The editors were asked not to record their names on these forms. Forms were not collected at meetings KD did not attend.
ES and CM extracted the phrases from KD's notes and from the completed forms and entered them into NVivo 2.0 qualitative analysis software (Qualitative Solutions and Research Pty. Ltd, Australia). Each manuscript was considered a separate document with an array of attributes such as date of discussion, number of positive and negative reviews by outside peer reviewers, time taken for discussion, its categorization by the editors as describing research or not, and final disposition of the manuscript.
We next used an iterative process to develop a classification schema for the 2,463 spoken and written phrases. We considered several possible themes, including the JAMA objectives [see Table 1], general journalistic goals, and ad hoc schema suggested by the phrases recorded.
For the first draft of the schema, CM and KD reviewed the phrases and documents from the editors' discussions and assigned them to 20 categories related to medical editorial decision-making and publication bias (as defined above). This schema was reviewed by two independent epidemiologists. Each performed an independent review followed by a group discussion with KD. We modified the schema further, categorizing phrases by whether they were related to science (eg, likelihood of bias relating to the study design), editorial beliefs or values (eg, likely interpretation by the public), or manuscript features (eg, short, well written). Using Nvivo, CM re-sorted most of the 20 categories into the new schema, retiring some categories and merging others. In a one hour meeting, CM and KD presented the revised schema to two independent social scientists, who suggested additional refinements.
Finally, we revised the concept of editorial beliefs and values to encompass what we called general journalistic goals, and developed a separate construct within the schema. We considered journalism in medicine to encompass a broad mission that includes educational, public health, and strategic goals such as timeliness, serving the readers' interests, presenting important medical issues, and addressing controversies. In the present instance, "journalism goals" were those that spoke to the mission of JAMA and other medical journals – meaning factors and values important to medical (or clinical) journals that publish new research (such as importance to medicine, strategic emphasis for the journal, and interest to the readership).
Thus, we classified each phrase as belonging to one of three mutually exclusive categories: science, journalism, and writing. Each phrase was further classified using a subcode within each category and each phrase was assigned a single code. All categories include phrases that are both favorable and unfavorable, although we describe the category using mainly positive terms [see Table 2]. For instance, phrases used to note exemplary ethical processes, as well as phrases used to note suspected conflict of interest, were classified as part of the "Ethics/Conflict of Interest" category.
We exported 2,463 coded phrases into Excel 2003 (Microsoft Corp) for counts. We excluded 87 phrases (76 of which were spoken) that were coded as describing the title or topic of the manuscript, leaving 2376 coded phrases for analysis.
Because our goal was to develop new ways to assess manuscript decision-making and publication bias, and because it could in no way influence the fate of manuscripts being discussed, our project did not involve written informed consent. Editors received written material describing the project, and the project was thoroughly discussed, at the initial manuscript meeting before note-taking and form completion began. We consulted officials at the Johns Hopkins Bloomberg School of Public Health Committee on Human Research who requested that we not include identifying information about the editors, peer reviewers, or authors.
Our results include 2,376 phases, 1,773 spoken and 603 written, which we present combined unless otherwise noted. Spoken phrases include 459 referring specifically to comments made by the peer reviewers and 1,314 that reflect the editors' own comments.
The most frequent phrases used concerned scientific issues (1,274 phrases), including research design and methods, presentation and interpretation of results, analysis methods, population studied, power and sample size, measurement variables and measures used, the need for additional data, generalizability, quality of the data, and conflict of interest/ethics [see Table 2]. Phrases associated with journalistic concerns (n= 729) included those classified as referring to some aspect of the manuscript as important to medicine, associated with a strategic emphasis for the journal, an interesting topic or results, a statement describing the study findings, relating to the author, a study's positive or negative results, novel/new content, readership concerns, filling a knowledge gap, timeliness, refuting or confirming standard practice or thinking, a topic relating to policy issues/a public message/public response, and special study characteristics. Finally, phrases we classified as concerned with writing (n = 373) were coded as either related directly to writing (eg, "not clear", "dense", "needs a rewrite"), or indirectly, when they referred to the refereeing process but not a specific referee comment that could be otherwise classified (eg, "a very thoughtful review", "authors responsive", "good revisions," "positive reviews").
When we examined the 603 written comments made by the editors following each manuscript discussion, we found that science, journalism, and writing were all important factors in the editors' opinions, regarding whether to proceed with acceptance or not [see Table 3]. Research design and methods (science) phrases were used most often, with importance to medicine (journalism), and comments related to the refereeing process (writing) next. The rankings of written comments were not meaningfully affected when we examined research articles only or accepted articles only.
Our goal was to identify new aspects of editorial decision-making that can be used to guide research on publication bias, particularly relating to the stage of the process where editors physically meet around a table and by teleconference to discuss the merits of the manuscripts presented to the group. We found that the discourse at manuscript meetings could be recorded and analyzed, using two means of word and phrase collection. One method was simple note-taking at the meeting, which, by its nature, is likely to be incomplete, for example because the note-taker did not hear or did not record all comments. The other method was a collection of a paper form on which editors were asked to list the reasons they were inclined to proceed to the next step in publication and reasons they were not. Neither of these methods uses standard qualitative research methodology, but instead combines qualitative and epidemiologic approaches. As far as we know, this methodology has not been employed before.
We developed our classification schema iteratively, with full knowledge of the phrases recorded. While our classification schema is only one of many possible perspectives, with overlapping categories, and the coding of individual phrases is subjective, our goal was to group editorial commentary in a way that would help future researchers to study the editorial process, rather than to identify "the truth".
By identifying the major topics of editorial discussion, our work contributes also to development of new variables for studying the publication process more generally. While a body of literature exists on factors involved in editorial decision making, the hypotheses tested in past studies (eg, whether publication is associated with statistical significance of results) have mainly reflected individual authors' opinions and experience regarding the decision-making process, and anecdotal rather than recorded evidence.
We did not collect any data on the manuscripts rejected before this stage, and cannot infer anything about the process leading to rejection. It is possible for example, that some reasons for rejection are not reflected in the discussion of manuscripts considered at the manuscript meetings.
Most of the discussion around the table at the 12 JAMA manuscript meetings concerned the science of research presented in the manuscripts. This corresponds indirectly to the findings of Olson and colleagues who found an association between acceptance for publication and factors positively associated with study quality, for reports of clinical trials submitted to JAMA .
Perhaps more interesting is the discussion related to what we categorized as "journalism" concerns. This included phrases conveying the concept that the topic or findings were "important," as well as related phrases such as "interesting" and those associated with the study findings. While there is nothing surprising about medical journal editors being attracted to important or interesting reports, nor about physician-editors attaching degrees of importance to studies likely to affect the care of patients, we now are able to see the context in which study results are discussed. We would expect, and indeed found, that comments in this category were less frequent than those related to science, given that if the science is shaky there would be little point in much further discussion.
A second aspect of the journalism category involved the editors' interest in maximizing strategic advantage for the journal and related readership issues. Phrases classified in the "strategic emphasis" category included "refer to Archives", "published elsewhere", "we want the whole thing, not just a salami slice", "similar paper in revision", "not for JAMA", "too long", "too specialized". Examination of the statement of JAMA's aims reveals extensive language indicating that such journalistic features of articles are valued. It is understandable that biomedical editors choose manuscripts that match the stated mission of a journal, which includes an embedded code of values and beliefs. No matter what the intention, those beliefs are a "differential inclination," and may contribute to publication bias, broadly defined. In addition, publishing is a business with profit-making goals, which can indirectly influence a journal's scope and selection of manuscripts , for example selection of articles likely to be frequently cited and thus positively influence the impact factor.
Written summaries of reasons to proceed to the next stage toward acceptance or not, completed for each manuscript by the editors, reflected the combined importance of science, journalism, and writing considerations. Peer reviewer comments ranked high in reasons given by editors to proceed toward acceptance or not. Given that the manuscripts discussed at meetings are those still under consideration after two elimination rounds, and that referees are not necessarily experts and can sometimes be wrong, the continuing influence of peer reviewers is especially important. Our meeting notes also indicate a major role for peer review in editorial decision-making, with more than 25% (459/1773) of the spoken phrases referring to comments by the reviewers.
In the future, it would be useful to examine whether accepted and rejected manuscripts differ in the discussion content and written summaries. It would also be interesting to see whether similar proportions of peer reviewer comments mentioned in the discussions were positive and negative. Finally, it would be useful to compare the comments attributed to the peer reviewers with phrases used by the editors, to see if their comments reflect similar concerns.
This project provides insight into the process at one journal, focusing specifically on comments leading to requests for manuscript revision, acceptance, or rejection. This journal was selected because it has a tradition of willingness to participate in research about the peer review process . Although we do not know whether the process at JAMA is generalizable to other journals, expansion of our thinking in this area will inform future research, which was our goal.
We developed a classification of editorial discourse that provides insight into editorial decision-making and concepts that need exploration in future studies.
Dickersin K: Publication bias: Recognizing the problem, understanding its origins and scope, and preventing harm. Publication Bias in Meta-analysis: Prevention, Assessment, and Adjustments. Edited by: Rothstein H, Sutton A, Borenstein M. 2005, New York: Wiley, 11-33.
Easterbrook PJ, Berlin JA, Gopalan R, Matthews DR: Publication bias in clinical research. Lancet. 1991, 337: 867-872. 10.1016/0140-6736(91)90201-Y.
Dickersin K, Min YI, Meinert CL: Factors influencing publication of research results. Follow-up of applications submitted to two institutional review boards. JAMA. 1992, 267: 374-378. 10.1001/jama.267.3.374.
Dickersin K, Min YI: NIH clinical trials and publication bias. Online J Curr Clin Trials. 1993, Doc. No. 50.
Stern JM, Simes RJ: Publication bias: evidence of delayed publication in a cohort study of clinical research projects. BMJ. 1997, 315: 640-645.
DeCullier E, Lhéritier V, Chapuis F: Fate of biomedical research protocols and publication bias in France: retrospective cohort study. BMJ. 2005, 331: 19-24. 10.1136/bmj.38488.385995.8F.
Olson CM, Rennie D, Cook D, Dickersin K, Flanagin A, Hogan J, Zhu Q, Reiling J, Pace B: Publication bias in editorial decision making. JAMA. 287 (21): 2825-2828. 10.1001/jama.287.21.2825. 2002 Jun 5
Mahoney MJ: Publication prejudices: an experimental study of confirmatory bias in the peer review system. Cogn Ther Res. 1977, 1: 161-175. 10.1007/BF01173636.
Peters DP, Ceci SJ: Peer-review practices of psychological journals: the fate of published articles submitted again. Behav Brain Sci. 1982, 5: 187-195.
Fisher RS, Powers LE: Peer-reviewed publication: A view from the inside. Epilepsia. 2004, 45: 889-894. 10.1111/j.0013-9580.2004.14204.x.
Tumber MB, Dickersin K: Publication of clinical trials: Accountability and accessibility. J Intern Med. 2004, 256: 271-283. 10.1111/j.1365-2796.2004.01392.x.
The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/7/44/prepub
The authors are grateful to the Cathy DeAngelis, Annette Flanagin, and other JAMA editors who generously allowed us access to their words and workplace. We also acknowledge the helpful suggestions by Sophie Alexander and Beatrice Blondel, who contributed to our classification scheme, and Terrie Wetle and Phil Brown who provided feedback on the penultimate classification schema. Christine Costantino assisted with the database.
This work was supported in part by Brown University, which contributed to KD's support during her 2003 sabbatical.
DR is Deputy Editor (West) at JAMA
KD conceived of, designed, and provided oversight of the project, and participated in the acquisition, analysis and interpretation of data. With CM, she drafted the manuscript and she coordinated subsequent edits and revisions. She serves as the project's guarantor. ES contributed to acquisition, analysis, and interpretation of the data, and to revision of the manuscript. CM contributed to the concept and design of the study, the acquisition, analysis and interpretation of the data. She also participated in drafting the manuscript and its finalization. DR contributed to the concept and design of the study, and to revision of the manuscript. All authors have read and approved the final manuscript.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Dickersin, K., Ssemanda, E., Mansell, C. et al. What do the JAMA editors say when they discuss manuscripts that they are considering for publication? Developing a schema for classifying the content of editorial discussion. BMC Med Res Methodol 7, 44 (2007). https://doi.org/10.1186/1471-2288-7-44
- Publication Bias
- Journal Editor
- General Medical Journal
- High Quality Science
- Refereeing Process