Skip to main content registration can supplement information in abstracts for systematic reviews: a comparison study



The inclusion of randomized controlled trials (RCTs) reported in conference abstracts in systematic reviews is controversial, partly because study design information and risk of bias is often not fully reported in the abstract. The Association for Research in Vision and Ophthalmology (ARVO) requires trial registration of abstracts submitted for their annual conference as of 2007. Our goal was to assess the feasibility of obtaining study design information critical to systematic reviews, but not typically included in conference abstracts, from the trial registration record.


We reviewed all conference abstracts presented at the ARVO meetings from 2007 through 2009, and identified 496 RCTs; 154 had a single matching registration record in Two individuals independently extracted information from the abstract and the record, including study design, sample size, inclusion criteria, masking, interventions, outcomes, funder, and investigator name and contact information. Discrepancies were resolved by consensus. We assessed the frequencies of reporting variables appearing in the abstract and the trial register and assessed agreement of information reported in both sources.


We found a substantial amount of study design information in the record that was unavailable in the corresponding conference abstract, including eligibility criteria associated with gender (83%; 128/154); masking or blinding of study participants (53%, 82/154), persons administering treatment (30%, 46/154), and persons measuring the outcomes (40%, 61/154)); and number of study centers (58%; 90/154). Only 34% (52/154) of abstracts explicitly described a primary outcome, but a primary outcome was included in the “Primary Outcome” field in the record for 82% (126/154) of studies. One or more study interventions were reported in each abstract, but agreed exactly with those reported in only slightly more than half the time (88/154, 56%). We found no contact information for study investigators in the abstract, but this information was available in less than one quarter of records (17%; 26/154).


RCT design information not reported in conference abstracts is often available in the corresponding registration record. Sometimes there is conflicting information reported in the two sources and further contact with the trial investigators may still be required.

Peer Review reports


Identification of relevant randomized controlled trials (RCTs) is an integral part in the conduct of systematic reviews of intervention efficacy and effectiveness. RCTs are usually identified through searching electronic databases (e.g., PubMed, EMBASE) and handsearching (manually screening biomedical journals, conference proceedings and other publications). Recently, the Institute of Medicine (IOM) published guidelines stating that results from conference abstracts should be included in systematic reviews because conference abstracts provide an important source of unpublished trials [1]. This is because only about 60% of controlled clinical trials presented as conference abstracts are subsequently published as journal articles, and trials with negative or null results are published less frequently than those with positive findings [2]. The end result is that inclusion of the grey literature in systematic reviews and meta-analyses leads to smaller treatment effect sizes overall [3]. This presents a problem for systematic reviewers, however, as reporting of study design in abstracts is poor [4] and results reported are often preliminary. Furthermore, there are concerns that abstracts rarely undergo peer review. Although adoption of CONSORT reporting guidelines for conference abstracts may lead to some improvement over time, controversy remains about whether to include conference abstracts in systematic reviews [57].

One approach to learning more about critical design elements of trials reported only in conference abstracts is to seek information from other sources. For example, a trials register such as includes key protocol items relevant to determining eligibility and performing critical appraisal of a study.

We hypothesized that information included in a trials register record could be used to supplement the sparse information on study design presented in a conference abstract. To test the hypothesis, we obtained reports of RCTs presented at the annual meeting of the Association for Research in Vision and Ophthalmology (ARVO) from 2007 to 2009; this is an international meeting of more than 10,000 attendees. The ARVO organizers require that any abstract submitted for presentation at the annual meeting that describes a concurrently controlled trial must be registered in an electronically searchable, publicly available trials register [8]. The online conference abstract submission form includes a box in which abstract authors reporting controlled clinical trials are asked to supply the name of the trials register and the trial registration number where the trial was registered. Accompanying instructions include a definition of a clinical trial and a hyperlink to a frequently asked questions page that describes trials registers, including a drop-down menu of acceptable registers. A previous study showed that when authors complied with this requirement, about 90% reported registering trials at [9].


Identification of included randomized controlled clinical trials

We reviewed all abstracts presented at the ARVO meetings from 2007 through 2009 (submitted through December 2008). Abstracts were classified as RCTs using the definition provided in the Cochrane Collaboration’s Handsearching Training Manual: “a study in which individuals (or other units) followed in the trial were definitely assigned prospectively to one of two (or more) alternative forms of health care using random allocation” [10]. One person hand searched the ARVO annual meeting abstracts online at A second person reviewed all studies classified as an RCT by the handsearcher and another person reviewed a sample of abstracts not classified as an RCT [9]. Discrepancies were resolved by consensus.

Two individuals independently extracted information about trial registration as reported for each abstract, including the name of the organization listed in the trial registration box and information in the box designated for the registration identification number. Abstracts listing more than one registration number or not listing any number were excluded from further analysis. Because 88% (276/312) of RCTs reporting a trials register listed as the trials register [9], we chose to include only records in our study. We entered the trial register number provided in the search box at and classified numbers as valid if a matching trial was identified in, and invalid if the search yielded no results. We did not attempt to identify a registration record for trials whose investigators had not included a registration number. We retrieved all records for trials as posted (i.e., including amendments incorporated in the record at the time of retrieval) from May and June, 2009.

Two persons independently reviewed the pairs for inclusion. We excluded pairs that described secondary analyses of trial data (e.g., analyses of ancillary study data), nested case–control studies from RCT data, and methodological studies associated with the RCT, because the objective of our study was to compare descriptions of the original RCT design; we also excluded pairs where the record stated that assignment to treatment was not randomized.

Abstraction of study design characteristics

We extracted information from both the abstract and the record, including type of randomized comparison (parallel, cross over, cluster), multi-center status (yes/no), number randomized, and inclusion criteria related to demographic characteristics of the study population (i.e., adults, children, included sexes, presence of a disease or condition, and/or healthy volunteers). Masking was characterized separately as yes/no for study participants, treatment administrators, and outcome assessors. The study intervention, primary outcome, secondary outcomes, and study funder(s) were extracted verbatim from the abstract and the record. We classified an outcome as the “primary” outcome in the abstract only if it was explicitly stated as such, and in the record only if it was included in the “Primary Outcome” field in the tabular view of the record. We classified all other outcomes reported as “non-primary” outcomes. The study funder was abstracted from the “Support” field of the abstract and either the “Sponsor” or “Collaborator” field in the record. We classified funders as industry or non-industry. If the study funder was unclear, we classified it as non-industry as a conservative measure. We also collected information about study status as reported in (“not yet recruiting”, “recruiting”, “active”, or “not recruiting, completed”). If more than one abstract had the same registration number, we extracted information from all relevant abstracts and reported it on a single data abstraction form. We also extracted the presence of a contact name, telephone number, and e-mail address from the record. All data were abstracted independently by two abstractors (RWS, LH, KD, AE, JT) on pre-tested paper data collection forms; we extracted all information from the abstract before we extracted information from the record. Discrepancies were resolved by consensus. All data from the paper forms were entered into an Access database.

Data analyses

We assessed frequencies of reporting variables in the abstract and the trials register. We assessed the concordance between study design characteristics that were described in both the conference abstract and the trials register record by comparing reports of variables across pairs. We determined whether a design characteristic that was not described in the abstract was present in the trials register. We also compared information reported in both sources and characterized the level of agreement as full, partial, or no agreement. If the abstract and the record agreed exactly or nearly exactly, we classified this as full agreement. A more precise definition of an outcome in the abstract (or register) was classified as partial agreement while a completely different outcome in the abstract was classified as a new outcome. We further categorized partial agreement by type of disagreement and source of additional information (i.e., the abstract or the record). We used SAS Version 9.2 (SAS Institute, Carey, NC) to perform all analyses.


Registration of clinical trials

The handsearching results have previously been reported [9]. Only 2.8% (496/17,953) of all abstracts presented at ARVO from 2007 to 2009 described results of an RCT; 276 of these reported registration in Excluding abstracts not meeting our eligibility criteria resulted in 158 abstracts. We also excluded 4 abstracts in which the record classified the study as not randomized. We linked each of the remaining abstracts with a single registration number, resulting in 154 pairs for analysis (see Figure 1).

Figure 1
figure 1

Flow chart of conference register pairs used for comparison of randomized controlled trial characteristics. RCT = randomized controlled trial; =

General study design

We observed generally good agreement on the randomized intervention comparison described in the abstract with what was reported in (80.5%, 124/154) (see Table 1). Whether a trial was single or multi-center was the same in 39 pairs, different in 13, and not reported in either source for 12 pairs. For the remaining 90 pairs, no information on multi-center status was available in the abstract, but was reported in

Table 1 Agreement between conference abstract and register on design of randomized comparison (n = 154 pairs)

Inclusion criteria

There was agreement between the abstract and record on inclusion criteria related to age, the presence of a disease or condition, and whether the trial participants were healthy volunteers (see Figure 2). The inclusion criterion of sex was rarely reported in the abstract, but when it was reported, the information usually agreed with what was reported in In contrast, the inclusion of men and/or women was always included in the record.

Figure 2
figure 2

Agreement between abstract and register on eligibility criteria. Bars show percent of pairs that agree (black) or disagree (gray) on eligibility criterion, or where information on a criterion was provided in the record but not the abstract (white). Eligibility criteria assessed are inclusion of adults, children, healthy volunteers, presence of a condition, men and/or boys, and women and/or girls and were categorized as present or not present. =


At least one experimental and one control intervention was reported in both the abstract and the record. We observed good agreement in the intervention reported in the abstract with that reported in when looking at the broad categories of treatment, prevention, diagnostic tests, and other (Table 1). However, there was less agreement when we compared more precise descriptions of the interventions. The interventions described in the abstract and the record agreed exactly for 56% (88/154) of pairs, while 37% (57/154) of pairs agreed partly and 6% (9/154) of pairs did not agree (see Table 2 for examples of disagreements). In eight pairs, partial agreement involved small differences in the description of the intervention (e.g., 100 μl versus 110 μl of a drug), and in 49 pairs, partial agreement involved additional information about the intervention (e.g., dosage, time of administration, duration, etc.) in one source that was unavailable in the other. In 12 pairs, additional information was present only in the record, and in 29 pairs additional information was present only in the abstract. For eight pairs, additional complementary information was present in both sources. Partial agreement also involved additional treatment arms: authors reported between one and four additional treatment arms in 11 records and between one and six additional treatment arms in 25 abstracts.

Table 2 All interventions reported in register and abstract that were classified as disagreements

Masking or blinding

We extracted information on masking separately for study participants, treatment administrators, and outcome assessors, finding that the record frequently provided information on masking that was not available in the abstract (Figure 3).

Figure 3
figure 3

Agreement between abstract and register on masking by study role. Bars show percent of pairs that did not provide information on masking or blinding (black), agree (dark gray), or disagree (light gray), or where information was provided in the record but not the conference abstract (white) on masking of study participants, persons administering the treatment, or persons measuring outcomes. Masking was categorized as present or not present. =

Sample size

The number of study participants or sample size was reported in both the abstract and for the majority (136/154, 88%) of studies. There were eight studies in which a sample size was reported in the trials register, but not the abstract. We compared the reported sample sizes for pairs where the abstract author stated that the number of participants represented all randomized study participants and the status on the record was either “completed” or “active, not recruiting” (see Table 3), and found good agreement.

Table 3 Comparison of number of randomized study participants reported in abstract with number reported in


Authors reported 800 outcomes in 152 abstracts; no outcomes were reported in 2 abstracts. Thirty four percent (52/154) of abstracts and 82% (126/154) of records explicitly described a primary outcome (Table 4). Of the 80 primary outcomes reported among the 40 abstract - pairs, 14 (18%) were classified as being in complete agreement and 39 (49%) as partial agreement (see Table 5). Partial agreement typically involved a more explicit description of an outcome as shown in in Table 6. Of the remaining 27 primary outcomes reported in the abstract, 13 were reported elsewhere in the record. There was poor agreement between non-primary outcomes reported in the abstract with those reported in (see Table 5).

Table 4 Agreement between conference abstract and register on reporting of one or more primary outcomes (n = 152 abstract-Clinical pairs)
Table 5 Agreement between conference abstract and register on outcomes (n = 40 pairs reporting ≥ 1 primary outcome and 152 pairs reporting a non-primary outcome)
Table 6 Examples of primary outcomes reported in register with primary outcome reported in abstract classified as “partial agreement”


One hundred forty-six unique funders were identified in either the abstract or the record. At least one funder was reported in each record, but only in 62% of abstracts (95/154). More than one funder was reported in 41 records and in 19 abstracts. The same funder(s) were reported in both sources for 10 studies. We observed partial agreement across funders for 37 pairs, in which an additional or different funder was reported in either the abstract or in the record.

Contact information

The name of a contact person was included on 83% (128/154) of records, but a phone number was found on only 25, and an email address on 26 records. No contact information was provided on any conference abstract except author name and affiliation.


A substantial amount of additional information on study design was available on the record that was not presented in a corresponding conference abstract. Information on multi-center status, eligibility criteria with respect to sex, who is masked, and primary outcome is present to varying degrees in the register record. In addition, the name of a contact was included in 83% of records, so that if desired information about the trial was not available, or conflicted with what was reported in the abstract, a systematic reviewer could contact the study author directly, although contact information was provided in a minority of records.

Thus, if a trial registration number is available for a study to be included in a systematic review, a systematic reviewer may be able to find additional information about study design in a trials register record. However, there are caveats to these findings. First, information about trial registration may not be required nor provided by conference organizers generally. Second, we encountered numerous disagreements between the information provided in the conference abstract with information contained within the record on a number of items. was originally intended to provide a source of information about the existence of a trial for patients and clinicians, to diminish redundant research effort, and to alert researchers to the possibility of publication bias. Currently it provides limited protocol information, although it would be a natural repository for full protocols. A published protocol or design and methods paper would provide more information, but until public availability of all trial protocols is achieved, trials registers serve as surrogates for study protocols.

The findings from our study relate specifically to ARVO abstracts and and may not be applicable to other conference abstracts or trials registers. Furthermore, our requirement that a valid registration number be reported on the abstract submission form could mean that studies were actually registered but the number not recorded or incorrectly recorded.

Previous studies have identified discrepancies between trials register records and associated full length publication on study design characteristics, especially related to the description of primary and secondary outcomes [1116]. In our study, we found discrepancies related to the detailed descriptions of study interventions and outcomes reported in the conference abstract compared with those reported in the record. Frequently there was more information available in the abstract than was in the record (e.g., dose or duration of treatment). Most disturbing was the appearance of additional treatment arms in the abstract that were not included in the register, suggesting that investigators are not regularly updating the trial register record as required or that abstracts represent preliminary findings, before an arm was dropped.

The discrepancies in outcomes may be an indication of error, selective outcome reporting, or simply due to investigators not updating the record in a timely manner. We obtained register records that were available shortly following presentation of abstract results at ARVO, so any additional outcomes or changes in the study design should have been incorporated into the trial register as an amendment to the protocol items. In addition, we compared outcomes in only one direction, i.e. from abstract to record. We did not do the reverse analysis (i.e. comparing all outcomes reported in with those reported in the abstract) because we would not expect an abstract to include all outcomes reported in In so doing, we did find that sometimes the outcome reported in the abstract as the “primary” outcomes was a secondary outcome in the record. In the future it will be of interest to compare outcomes reported as results with the specified outcomes in the record. Results reporting in opened in September 2008. When we retrieved the records in May and June 2009, none of the records included any study results when downloaded.

Other studies have also found discrepancies between the description of an RCT as reported in an abstract with what was reported in a full length publication [1721]. It would be of interest to make a three-way comparison of a conference abstract, trial register and results record, and full length publication to determine congruence between these sources of information. It would also be useful for authors to include conference abstracts in the record as a publication, in addition to full length publications.

Whether the study results reported in an abstract should be included in a systematic review or used to make clinical decisions is unclear. That results from abstracts are being included in systematic reviews has been reported in a sample of meta-analyses indexed in Medline [7], for health technology assessments [5], and for drug formulary decision-making [22]. Including abstracts in systematic reviews is recommended by IOM [1], the Cochrane Collaboration [23], and the Agency for Health Care Research and Quality [24], although caution is urged due to the preliminary nature of abstract results [25, 26]. Results from a Cochrane systematic review found that there was an overall smaller treatment effect if systematic reviewers included abstract results in the review compared to the treatment effect size without abstracts included, although this finding was not statistically significant [3]. Additional investigators report similar results, showing small reductions in the size of the treatment effect when abstract results are included in a systematic review [27, 28].

Overall, given our findings, we would encourage systematic reviewers to take advantage of the information provided in the record to supplement the information provided in an abstract. We did not record the amount of time it took to extract information for each record. Because the trials register number was available, matching the abstract to the registry report was straightforward. Extracting all the information we did for our study was somewhat time consuming, and we found that it was more efficient to obtain the information from the tabular view in rather than the full text view. Most likely, it would take much less time for a systematic reviewer looking for specific information. We would caution systematic reviewers to exercise care in using data from a trials register record, however, as with any unpublished study result. In many cases, systematic reviewers may not be able to distinguish between conflicting reports and still find it necessary to contact the study investigators. For this situation, the record frequently provides contact information that is almost always missing from a conference abstract.


Systematic reviewers may find additional information about an RCT in the record to supplement the information provided in a conference abstract. However, it may still be necessary to contact study investigators for information not included in either the abstract or record, or that is in conflict between the two sources.



Association for Research in Vision and Ophthalmology


Consolidated Standards of Reporting Trials


Institute of Medicine


Randomized controlled trial.


  1. IOM (Institute of Medicine): Knowing What Works in Health Care: a Roadmap for the Nation. Committee on Reviewing Evidence to Identify Highly Effective Clinical Services. 2011, Washington DC: The National Academic Press

    Google Scholar 

  2. Scherer RW, Langenberg P, Von Elm E: Full publication of results initially presented in abstracts. Cochrane Database Syst Rev. 2007, 2: 10.1002/14651858.MR000005.pub3. Art. No.: MR000005

    Google Scholar 

  3. Hopewell S, McDonald S, Clarke MJ, Egger M: Grey literature in meta-analyses of randomized trials of health care interventions. Cochrane Database Syst Rev. 2007, 2: 10.1002/14651858.MR000010.pub3. Art. No.:MR000010

    Google Scholar 

  4. Hopewell S, Eisinga A, Clarke M: Better reporting of randomized trials in biomedical journal and conference abstracts. J Information Science. 2008, 34: 162-173.

    Article  Google Scholar 

  5. Dundar Y, Dodd S, Williamson P, Walley T, Dickson R: Searching for and use of conference abstracts in health technology assessments: Policy and practice. Int J Technol Assess Health Care. 2006, 22: 283-287.

    PubMed  Google Scholar 

  6. Egger M, Juni P, Bartlett C, Holensteing F, Sterne J: How important are comprehensive literature searches and the assessment of trial quality in systematic reviews? Empirical study. Health Technol Assess. 2003, 7 (Issue 1): 1-76.

    CAS  PubMed  Google Scholar 

  7. Cook DJ, Guyatt GH, Ryan G, Clifton J, Buckingham L, Willan A, McIlroy , Oxman AD: Should unpublished data be included in meta-analyses?. JAMA. 1993, 269: 2749-2753. 10.1001/jama.1993.03500210049030.

    Article  CAS  PubMed  Google Scholar 

  8. Association for Research in Vision and Ophthalmology Statement on Registering Clinical Trials:,

  9. Scherer RW, Sieving P, Dickersin K: Can we depend on investigators to identify and register randomized controlled trials?. PLoS One. 2012, 7: e44183-10.1371/journal.pone.0044183.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Training Manual for Handsearchers, US Cochrane Center: []

  11. Dwan K, Altman DG, Cresswell L, Blundell M, Gamble CL, Williamson PR: Comparison of protocols and registry entries to published reports for randomised controlled trials. Cochrane Database Syst Rev. 2011, 10.1002/14651858.MR000031.pub2. Art. No.: MR000031, Issue 1

    Book  Google Scholar 

  12. Ghandi R, Jan M, Smith HN, Mahomed NN, Bhandari M: Comparison of published orthopaedic trauma trials following registration in BMC Musculoskelet Disord. 2011, 12: 278-10.1186/1471-2474-12-278. At.

    Article  Google Scholar 

  13. You B, Gan HK, Pnd G, Chen EX: Consistency in the analysis and reporting of primary end points in oncology randomized controlled trials form registration to publication: a systematic review. J Clin Oncol. 2012, 30: 210-216. 10.1200/JCO.2011.37.0890.

    Article  PubMed  Google Scholar 

  14. Mathieu S, Boutron I, Moher D, Altman DG, Ravaud P: Comparison of registered and published primary outcomes in randomized controlled trials. JAMA. 2009, 302: 977-984. 10.1001/jama.2009.1242.

    Article  CAS  PubMed  Google Scholar 

  15. Millette K, Roseman M, Thombs BD: Transparency of outcome reporting and trial registration of randomized controlled trials in top psychosomatic and behavioral health journals: a systematic review. J Psychosomatic Research. 2011, 70: 205-217. 10.1016/j.jpsychores.2010.09.015.

    Article  Google Scholar 

  16. Bourgeois F, Murthy S, Mandl KD: Outcome reporting among drug trials registered in Ann Intern Med. 2010, 153: 158-166. 10.7326/0003-4819-153-3-201008030-00006.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Toma M, McAlister FA, Bialy L, Adams D, Vendermeer B, Armstrong PW: Transition from meeting abstract to full-length journal article for randomized controlled trials. JAMA. 2006, 295: 1281-1287. 10.1001/jama.295.11.1281.

    Article  CAS  PubMed  Google Scholar 

  18. Dundar Y, Dodd S, Williamson P, Dickson R, Walley T: Case study of the comparison of data from conference abstracts and full-text articles in health technology assessment of rapidly evolving technologies: Does it make a difference?. Int J Tech Assess Health Care. 2006, 22: 288-294.

    Google Scholar 

  19. Prasad S, Lee DJ, Yuan JC-C, Barao VAR, Shyamsundeer N, Sukotjo C: Discrepancies between abstracts presented at International Association for Dental Research annual sessions from 2004 to 2005 and full-text publications. International J Dentistry. 2012, 10.1155/2012/859561. 859561

  20. Tam VC, Hotte SJ: Consistency of phase III clinical trial abstracts presented at an annual meeting of the American Society of Clinical Oncology compared with their subsequent full-text publications. J Clin Oncol. 2008, 26: 2205-2211. 10.1200/JCO.2007.14.6795.

    Article  PubMed  Google Scholar 

  21. Snedeker KG, Campbell M, Totton SC, Guthrie A, Sargeant JM: Comparison of outcomes and other variables between conference abstracts and subsequent peer-reviewed papers involving pre-harvest or abbatoir-level interventions against foodborne pathogens. Prev Vet Med. 2010, 9: 67-76.

    Article  Google Scholar 

  22. Weizman AV, Griesman J, Bell CM: The use of research abstracts in formulary decision making by the Joint Oncology Drug Review of Canada. Appl Health Econ Policy. 2010, 8: 387-391. 10.2165/11530510-000000000-00000.

    Article  Google Scholar 

  23. Cochrane Handbook of Systematic Reviews of Interventions: []

  24. Relevo R, Balshem H: Finding evidence for comparing medical interventions: AHRQ and the Effective Health Care Program. J Clin Epidem. 2011, 64: 168-1177.

    Article  Google Scholar 

  25. Falagas ME, Rosmarakis ES: Clinical decision-making based on findings presented in conference abstracts: is it safe for our patients?. Eur Heart J. 2006, 27: 2038-2039. 10.1093/eurheartj/ehl175.

    Article  PubMed  Google Scholar 

  26. Wong SS-W, Fraser C, Lurenco T, Barnett D, Avenell A, Glazener C, Cuthberston B, N’Dow J: The fate of conference abstracts: systematic review and meta-analysis of surgical treatments for men with benign prostatic enlargement. World J Urol. 2010, 28: 63-69. 10.1007/s00345-009-0500-3.

    Article  PubMed  Google Scholar 

  27. Hoehner C, Soares J, Parra DC, Ribeiro IC, Pratt M, Bracco M, Hallal PC, Bownson RC: Physical activity interventions in Latin America: What value might be added by including conference abstracts in a literature review?. J Physical Activity and Health. 2010, 7 (Suppl 2): S265-S278.

    Google Scholar 

  28. Hopewell S, Clarke M, Moher D, Wager E, Middleton P, Altman DG, Schulz KF, CONSORT Group: CONSORT for reporting randomized controlled trials in journal and conference abstracts: Explanation and elaboration. PLoS Med. 2008, 5: e20-10.1371/journal.pmed.0050020.

    Article  PubMed  PubMed Central  Google Scholar 

Pre-publication history

Download references


Funding for this study was provided by the National Eye Institute, National Institutes of Health, Grant 1U01EY020522. Publication of this article was funded in part of the Open Access Promotion Fund of the Johns Hopkins University Libraries.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Roberta W Scherer.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

RWS conceived of the study, participated in the study design and data extraction, entered the data in the Access database, participated in the analysis and interpretation of results, and drafted the manuscript. LH participated in the study design, data extraction, analysis and interpretation of results. AE participated in the study design, data extraction, and interpretation of results. JT participated in the study design and data extraction. KD participated in data extraction and interpretation of results. All authors read, edited, and approved the final manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Authors’ original file for figure 3

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Scherer, R.W., Huynh, L., Ervin, AM. et al. registration can supplement information in abstracts for systematic reviews: a comparison study. BMC Med Res Methodol 13, 79 (2013).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: