Challenges of self-reported medical conditions and electronic medical records among members of a large military cohort
© Smith et al; licensee BioMed Central Ltd. 2008
Received: 03 January 2008
Accepted: 05 June 2008
Published: 05 June 2008
Self-reported medical history data are frequently used in epidemiological studies. Self-reported diagnoses may differ from medical record diagnoses due to poor patient-clinician communication, self-diagnosis in the absence of a satisfactory explanation for symptoms, or the "health literacy" of the patient.
The US Department of Defense military health system offers a unique opportunity to evaluate electronic medical records with near complete ascertainment while on active duty. This study compared 38 self-reported medical conditions to electronic medical record data in a large population-based US military cohort. The objective of this study was to better understand challenges and strengths in self-reporting of medical conditions.
Using positive and negative agreement statistics for less-prevalent conditions, near-perfect negative agreement and moderate positive agreement were found for the 38 diagnoses.
This report highlights the challenges of using self-reported medical data and electronic medical records data, but illustrates that agreement between the two data sources increases with increased surveillance period of medical records. Self-reported medical data may be sufficient for ruling out history of a particular condition whereas prevalence studies may be best served by using an objective measure of medical conditions found in electronic healthcare records. Defining medical conditions from multiple sources in large, long-term prospective cohorts will reinforce the value of the study, particularly during the initial years when prevalence for many conditions may still be low.
Epidemiological studies often rely on self-reported medical history for both exposure and outcome information. A number of studies have addressed the reliability of these data by comparing self-reported information with objective sources, such as medical records. Results vary by study population and by diagnosis, as well as study design [1–13]. Previous research on the accuracy of self-reported angina shows low agreement (kappa [κ] = 0.57) in elderly patients , but substantial agreement (κ = 0.72) in men participating in the British Regional Heart Study , with differences possibly attributed to dissimilarities in study populations and/or study design [5, 10, 14]. Variability by medical condition within the same study population has also been noted [1, 3, 5, 10, 12]. In a study of chronic diseases in elderly patients, researchers found high rates of agreement between self-reported and recorded diagnoses using kappa statistics for diabetes (κ = 0.84) and hypertension (κ = 0.70), but moderate to poor agreement for chronic lung disease (κ = 0.55), osteoarthritis of the knee (κ = 0.54), and chronic low back pain (κ = 0.36) . Several studies suggest that this variability may be due to poor communication between the health care provider and the patient, since diseases with clear diagnostic criteria (e.g., diabetes, hypertension, myocardial infarction) tend to have higher rates of agreement than those that may be more complicated to diagnose by the physician or more difficult for the patient to understand (e.g., heart failure) [1, 3, 10].
Using data from the Millennium Cohort Study, a longitudinal study designed to assess the long-term health effects of military service , self-reported clinician-diagnosed medical conditions were compared with diagnostic codes from available electronic medical records. Unlike previous studies of this kind, the current study investigated a constellation of medical conditions that, to the best of our knowledge, have not been previously examined. Understanding the potential limitations of self-reported versus objective medical record data for a broad array of medical conditions will yield greater understanding of the results of future epidemiological studies based on comparable data sources for similar health outcomes.
The Millennium Cohort Study is a large, 21-year prospective study aimed at evaluating the potential effects of deployment and other military occupational exposures on long-term health outcomes using self-reported and electronic military health care data [15, 16]. The invited participants were randomly selected from over 2 million US military personnel on active rosters in October 2000, with oversampling of Reserve and National Guard personnel, female service members, and those recently deployed, to ensure adequate statistical power to detect differences in even relatively rare outcomes in these subgroups. The baseline enrollment ended with 36% of those invited consenting to participate in the 21-year study. When compared with the 2000 US military at large, Cohort members were slightly more likely to be female, older, better educated, married, officers, in the Air Force, and from health care occupations . The higher enrollment of women and those recently deployed reflects the intended oversampling . Analyses to investigate potential reporting biases show no differential in responder health with respect to hospitalization and outpatient encounters in the year prior to enrollment , strong test-retest reliability , reliable vaccination reporting [19, 20], occupation reporting , and deployment reporting  and minimal differences between participants choosing web submission in comparison to paper submission .
Demographic and military data for the Cohort, as of October 1, 2000, included sex, date of birth, education, marital status, race/ethnicity, previous deployment experience (January 1, 1998, to September 1, 2000), pay grade, service component (active duty and Reserve/Guard), service branch (Army, Navy/Coast Guard, Air Force, and Marine Corps), and occupation.
The population for this study consisted of participants from the first panel of Millennium Cohort participants who voluntarily consented and completed a baseline questionnaire (n = 77,047) between 2001 and 2003. Reservists and National Guard members (n = 39,028) were excluded because their electronic medical records are not fully available within the Department of Defense (DoD) medical record system. Additionally, Cohort members who failed to respond to any of the questionnaire items related to the medical conditions of interest (n = 124) or who had missing covariate data (n = 97) were excluded. The remaining 37,798 (49 percent of the first panel) comprise the study population for these analyses.
The Millennium Cohort survey included a number of more serious diseases often associated with age . Though the population was fairly young at baseline (54% of cohort members were younger than age 35), by the end of the 21-year study, many will have reached an age associated with increased risk for chronic diseases. Self-reported medical conditions listed in Additional file 1 were based on responses to the question: "Has your doctor or other health professional EVER told you that you have any of the following conditions?" "Yes" or "No" response choices were provided for each condition.
Individual, electronic hospitalization and ambulatory data included diagnoses using International Classification of Diseases, Ninth Revision, Clinical Modification (ICD-9-CM) codes . These data were acquired from three sources: (1) the Standard Inpatient Data Record (SIDR), (2) the Standard Ambulatory Data Record (SADR), and (3) the Health Care Service Record (HCSR). SIDR contains up to eight ICD-9-CM discharge diagnoses for individual inpatient care at any DoD medical treatment facility worldwide since October 1988. SADR contains up to four ICD-9-CM diagnoses for individual outpatient encounters at any DoD health care facility since October 1998. HCSR contains up to 10 ICD-9-CM diagnoses for encounters at civilian facilities that are reimbursed by the DoD insurance system. These files contain historical inpatient data from October 1993 and outpatient data from October 1999. For each participant, all electronic data were scanned for ICD-9-CM codes corresponding to medical conditions from the time earliest records were available up to and including the date of survey submission.
In order to compare self-reported and electronic data, ICD-9-CM codes were selected to best represent the 38 medical conditions included in the questionnaire. Selection of one or more codes representing each medical condition evaluated was accomplished by several groups of paired clinician researchers, each blinded to the diagnostic codes selected by the other. Any discrepancies in ICD-9-CM codes selected were resolved through discussion. In addition, annual changes in ICD-9-CM coding up to 2003 (last year of survey submission) were accounted for in the final list of codes (Additional file 1) . Electronic medical records were scanned in chronological order, and diagnostic fields were scanned in numerical order for the selected diagnostic codes. Any diagnostic code in any portion of the medical record indicated agreement with a self-reported medical condition.
The prevalence of each condition was computed for both the self-reported and electronically maintained data. Statistical comparisons of these frequencies were performed using the chi-square test. Prevalence of conditions identified exclusively through the electronically maintained medical records was reported to estimate what might be lost by use of self-report alone.
Several measures of agreement were considered for interpretation of results in this study. An omnibus index, such as the kappa statistic, is often utilized in validity studies, but is appropriate only if the sole purpose of the research is to compare responses over time or to previous studies. If the results are intended for use in future studies, an omnibus index is unsatisfactory, and thus both measurements being compared should be presented. Furthermore, the kappa statistic is strongly affected by prevalence (i.e., when the prevalence is low, kappa approaches zero). Many of the medical conditions in the current study had low prevalence, therefore, the kappa statistic was not deemed appropriate for these analyses.
Sensitivity and specificity were considered as an alternative approach. Sensitivity and specificity might be used to explain how a positive or negative self-report of a particular medical condition compares with a documented diagnostic code in the medical record. However, these measures of diagnostic test performance become inappropriate to use when there is no "gold standard." Although the electronic medical records are thought to accurately reflect actual diagnoses for the active-duty study population, they are subject to coding errors and/or omissions, as well as potential biases, for example, those related to reimbursement issues. In addition, while participants are asked to consider their lifetime when answering the survey question, the electronic data only contain records beginning in October 1988. Furthermore, electronic data may not capture conditions diagnosed prior to active-duty military service. Since neither self-reported nor electronic medical record data were considered the gold standard for the existence of a medical condition, we chose not to report sensitivity and specificity.
After considering these alternatives, the present study used an approach similar to a previous investigation of cardiovascular patients in which positive and negative agreement was used to compare self-reported data on medical conditions with electronically available medical records . Positive and negative agreement was selected as our analytic approach to resolve the omnibus issue  and the lack of a diagnostic gold standard. Positive and negative agreement, unlike the kappa statistic, is unaffected by imbalances in marginal totals caused by high or low prevalence. This approach may provide a better understanding for analyses based on various data sources, including insight into the limitations of both self-reported and objective electronic data pertaining to a large number of medical conditions.
Self-reported medical conditions from questionnaire responses and objective health encounter data for a total of 37,798 Millennium Cohort participants were available for analysis. As previously described, participants who skipped all 38 medical conditions listed in the questionnaire (n = 124) were excluded from these analyses. Approximately 88 percent of the remaining answered all 38 conditions. In order to maximize the numbers available to assess each medical condition, participants failing to answer an individual medical condition were removed only from the analysis of that particular condition. This resulted in sample sizes that varied from 37,328 to 37,696 for each individual medical condition.
Demographic and military characteristics of active-duty Millennium Cohort participants (2001–2003) self-reporting medical conditions
Study population† N = 37,798 n (%)
Subjects who reported one or more conditions n = 18,581 n (%)
Some high school
High school graduate
Military pay grade
Previous deployment experience
Length of service, years
Electronic equipment repair
Health care specialists
Other technical & allied specialists
Functional support specialists
Service & supply handlers
Prevalence, positive agreement, and negative agreement of active-duty Millennium Cohort participant self-reported and electronic medical record data
% and 95% CI Self-reported
% and 95% CI Electronic- recorded
% Exclusively Electronic Recorded
More Likely Chronic
Hypertension (high blood pressure)
10.4 (10.1, 10.7)
8.1 (7.8, 8.3)
Significant hearing loss
9.4 (9.1, 9.7)
5.9 (5.7, 6.1)
3.3 (3.1, 3.4)
4.0 (3.8, 4.2)
2.7 (2.6, 2.9)
1.5 (1.4, 1.6)
2.4 (2.2, 2.5)
0.3 (0.2, 0.3)
Thyroid condition other than cancer
2.2 (2.1, 2.4)
1.9 (1.8, 2.0)
2.0 (1.9, 2.2)
1.8 (1.7, 2.0)
Chronic fatigue syndrome
1.3 (1.1, 1.4)
0.2 (0.1, 0.2)
Diabetes or sugar diabetes
1.2 (1.1, 1.3)
1.1 (0.9, 1.2)
Ulcerative colitis or proctitis
0.9 (0.8, 1.0)
0.3 (0.2, 0.3)
0.9 (0.8, 0.9)
2.1 (2.0, 2.3)
0.8 (0.7, 0.9)
0.2 (0.2, 0.3)
Coronary heart disease
0.7 (0.6, 0.8)
0.6 (0.5, 0.6)
0.6 (0.5, 0.7)
0.2 (0.1, 0.2)
0.6 (0.5, 0.7)
0.1 (0.1, 0.2)
0.6 (0.5, 0.7)
0.1 (0.1, 0.1)
0.6 (0.5, 0.7)
0.2 (0.2, 0.2)
Schizophrenia or psychosis
0.6 (0.5, 0.7)
0.8 (0.7, 0.8)
Kidney failure requiring dialysis
0.6 (0.5, 0.6)
0.0 (0.0, 0.0)
0.5 (0.5, 0.6)
0.0 (0.0, 0.0)
More Likely Acute or Transient
14.8 (14.5, 15.1)
13.9 (13.5, 14.2)
10.8 (10.5, 11.1)
4.3 (4.1, 4.5)
7.0 (6.8, 7.3)
6.7 (6.4, 6.9)
6.9 (6.6, 7.1)
0.6 (0.5, 0.6)
5.8 (5.6, 6.0)
3.9 (3.7, 4.1)
Any other heart condition
5.1 (4.9, 5.3)
2.1 (1.9, 2.2)
4.6 (4.4, 4.8)
3.0 (2.8, 3.2)
Stomach, duodenal, or peptic ulcer
3.3 (3.1, 3.5)
0.6 (0.5, 0.6)
Angina (chest pain)
2.6 (2.4, 2.8)
0.3 (0.2, 0.3)
2.2 (2.0, 2.3)
4.5 (4.3, 4.7)
1.7 (1.6, 1.8)
0.7 (0.7, 0.8)
Posttraumatic stress disorder
1.5 (1.4, 1.7)
1.3 (1.2, 1.4)
Any other hepatitis
1.3 (1.2, 1.5)
0.5 (0.4, 0.6)
1.1 (1.0, 1.3)
0.1 (0.1, 0.2)
0.9 (0.8, 1.0)
0.4 (0.4, 0.5)
0.7 (0.6, 0.8)
0.1 (0.1, 0.2)
0.7 (0.6, 0.8)
0.2 (0.1, 0.2)
0.5 (0.5, 0.6)
0.2 (0.2, 0.3)
Prevalence by length of service of most commonly reported medical conditions via self-report, electronic medical record, either self-report or electronic medical record, and both self-report and electronic medical record among active-duty Millennium Cohort participants
Length of Service (years)
0–5 n = 7,183†
6–10 n = 8,516†
11–15 n = 10,179†
≥ 16 n = 10,574†
Significant hearing loss
Any other heart condition
Stomach, duodenal, peptic ulcer
Health survey research obtaining outcome and risk factor information relies heavily on the ability of participants to correctly and specifically self-report their medical histories. Previous studies that have looked at the reliability of these data have focused on one or few conditions simultaneously. The Millennium Cohort questionnaire contains 38 clinician-diagnosed medical conditions self-reported by the participant, which were compared with diagnostic codes from available electronic medical records. The most commonly observed conditions from both data sources were sinusitis, migraine headaches, hypertension, hearing loss, and depression. Prevalence for most conditions was found consistently lower in the electronic medical records than by self-report. Negative agreement between self-report of medical conditions and electronic medical record data was quite high, whereas positive agreement was relatively low, increasing with longer observation periods of objective data.
The choice to use positive and negative agreement rather than other measures, such as the kappa statistic or sensitivity and specificity, was driven by inherent limitations in the applicability of these measures to the current study. The results of this study provide insight into the degree of concordance between self-reported and electronic medical record data in a predominantly healthy, young, working population. Results also illustrate changes in positive and negative agreement with length of time in military service, which, in this study, is equivalent to length of time of accrued medical record data. The observed variability in positive and negative agreement across diagnostic categories highlights the importance of using multiple data sources to assess health outcomes when possible. However, for those cases in which objective electronic data are not available, our assessment of diagnostic codes found exclusively in electronic data sources offers information on what would be missed (magnitude and direction of possible biases) using self-reported data alone.
For the most-prevalent medical conditions, positive agreement increased with length of time in service, illustrating that self-reported diagnoses are likely to be reflected in electronic medical records given enough opportunity for capture in health encounter data. However, time in service is largely associated with age. Thus, increasing diagnoses over time is likely the result of a combination of increasing age and increasing data capture, as well as increasing patient understanding of their medical condition(s). Perhaps chronic conditions are a more appropriate assessment of reliability in reporting. Chronic conditions, if diagnosed early, would persist into adulthood and would be reflected in military healthcare databases and thus be concordant with self-report. A diagnosis in childhood such as sinusitis, however, would not be in the military healthcare databases and would likely be reported on the survey as a diagnosis by the Cohort member thus explaining the reason for lower positive agreement. The lower positive agreement with acute conditions such as asthma may also be a result of diagnosis prior to military service. Alternatively, conditions such as kidney failure requiring dialysis, cirrhosis, and emphysema should be obvious to the person and happen later in life where surveillance with medical healthcare data would indicate such a disorder. These data suggest differing methods of ascertainment dependent upon the condition being studied.
There are practical reasons that may explain some of the lower concordance measures between self-report and electronic medical outcomes. A complete description of the medical condition(s) may not have been sufficiently addressed by the medical practitioner at the time of diagnosis, allowing the patient's perception to differ from what was medically coded. Patients may be in the diagnostic or 'rule-out' phase of an explanation for their ill health and may report conditions for which they have been tested, but not diagnosed, or may self-diagnose in the absence of a satisfactory explanation for their health complaint. Inadequate patient-clinician communication may also account for some of the disagreement noted in this study. Additionally, the "health literacy" of patients may also explain reduced recognition of listed medical outcomes in the survey if the knowledge proficiency for some medical conditions vary . Still, one would argue that with good patient-clinician communication, patients will recognize their diagnosed conditions, and may or may not recognize conditions with which they have not been diagnosed, both leading to an increase in positive and negative agreement. Further, it is not possible to know if an individual self- managed a condition such as migraines without consulting a medical professional. Although the question stated, "Has your doctor or other health professional EVER told...", it is possible that the participant marked affirmative thus impacting agreement between data sources. To a lesser degree, low concordance in other rare conditions may represent inaccurate ICD-9-CM codes within the medical records.
The method for selecting ICD-9-CM codes to represent each condition may have also affected the rates of agreement. The codes presented in Additional file 1 were chosen by two expert reviewers to reflect, in their opinion, the presence of each medical condition, while being neither too broad nor too narrow in definition. If, for example, the selected codes more broadly defined a condition, more cases would be identified in the electronic record, thereby increasing positive agreement but decreasing negative agreement. Conversely, if the codes were narrowly defined, fewer cases would be identified in the electronic record, increasing negative agreement and decreasing positive agreement. To best objectively accomplish this task, the expert clinician researchers chose codes most appropriate for each condition blinded to the diagnostic codes selected by the other. Discrepancies were resolved through discussion.
Another inherent problem with using ICD-9-CM coding for concordance studies is that ICD-9-CM codes are not uniquely related to only one condition. Furthermore, the medical conditions used in this study are not mutually exclusive of each other. For example, myocardial infarctions are a result of coronary heart disease, and therefore, a number of codes used to identify a myocardial infarction (i.e., 410.xx, 411.0, and 412) are also associated with coronary heart disease. If a patient had an ICD-9-CM code of 412 in his or her medical record, and reported having had a heart attack, but not coronary heart disease, the net effect would be an increase in positive agreement for heart attack, but a decrease in this same measure for coronary heart disease. This limitation may help to explain the low concordance for some of the medical conditions analyzed in this study.
Further explanation for disagreement between self-report and objective record data may be due, in part, to the electronic database itself. Electronic medical records contain historical data beginning with October 1988 for active-duty service members. Conditions diagnosed prior to this time or prior to military service may not be identified using these records. To reduce the effect of these limitations, the study population was restricted to active-duty service members, and a lengthy period of observation was used. The probability of many of these medical conditions increases with age and age was found to be associated with length of service (correlation = 91.4%). However, length of service in the military will also affect the likelihood of ascertaining a medical encounter in available military electronic records. Individuals with a shorter service length have less opportunity for medical events to be captured in military health system records than those with longer service time. The mean age and length of service for this study population was 31.9 and 11.0 years respectively, whereas the mean age and length of service for those with at least one reported condition was 33.4 and 12.3 years respectively. To adjust for this factor, prevalence and positive agreement for the most commonly reported conditions were stratified by length of service. We further tried to mitigate the vulnerability to prevalence that the kappa statistic would have resulted in by choosing a measure of positive and negative agreement that is independent of the outcome prevalence. These problems should diminish in general and particularly with future data collection efforts in this population. Furthermore, conditions diagnosed prior to military service or inception of electronic medical records that recur or for which follow-up is ongoing will be captured in available data sources. Validation studies of medical records have found evidence of non-reporting and mis-reporting of diagnoses by physicians [25–27]. When experienced coders are employed, ICD-9-CM codes are assigned using information recorded in the medical record, and therefore, any inaccuracies in the medical record will also be reflected in the coding. In addition, other mis-reporting and non-reporting may occur during the coding process. Other issues will remain that will cause misclassification on the part of the study subject and medical record. Such errors, if random, will serve to diminish associations between outcomes and exposures, thus biasing findings toward the null hypothesis. However, given the large sample size of this cohort, this may not be a major problem with regard to missing significant associations since the high level of statistical power will outweigh the potentially smaller effect due to nondifferential misclassification.
The study population used in this investigation is a subset of Millennium Cohort responders and may not be representative of the military or the Cohort. Analyses were limited to active-duty participants because electronic medical record data are not fully available for Reserve and National Guard members. Further, while on active duty, service members have ready access to essentially free medical care in Defense Department facilities and they seldom seek medical care outside the Defense Department health care system. However, it is possible that for some conditions such as mental health disorders where the person may not want the diagnosis on their military healthcare record, a person may go to an outside provider and incur the cost of treatment.
Despite these limitations, this study has many strengths. Pairing Millennium Cohort data with available electronic medical records allowed examination of a wide range of self-reported medical conditions in a large, working population. Unlike most previous studies of this kind that have focused on older populations, the current study was conducted on a relatively young adult working population. Few epidemiological studies to date have had the resources to investigate the concordance of self-reported and electronic medical record data on such a broad range of conditions in a population of this size. In addition, the electronic medical record data were relatively complete for the available time frame among active-duty personnel. What may be most important about this current report is what these data suggest, in a broad way, that self-reported medical data may be sufficient for ruling out history of a particular condition as suggested by the high negative agreement values. Further, that prevalence studies may be best served by using an objective measure of medical conditions found in electronic healthcare records. Finally, the Cohort itself has been shown to be well-representative of its target population, relatively free of response biases, and to have strong reliability metrics [15, 17, 18, 20–23, 28].
In summary, this article highlights the research challenges of self-reported medical outcomes data and also underscores the potential limitations of electronic medical record data. Data integrity increased with length of observation within the medical record data. This study demonstrated that electronic diagnoses generally agreed with self-reported medical conditions, but more accurately represented the absence of disease over the presence of disease. As in the Millennium Cohort Study, health researchers who rely on self-reported medical conditions should consider using multiple data sources to assess health outcomes when possible, particularly in young or healthy populations.
Department of Defense
International Classification of Diseases, Ninth Revision, Clinical Modification
Health Care Service Record
Standard Ambulatory Data Record
Standard Inpatient Data Record.
In addition to the authors, the Millennium Cohort Study Team includes Gregory C. Gray, MD, MPH, College of Public Health, University of Iowa, Iowa City, IA, USA; and James R. Riddle, DVM and Timothy S. Wells, DVM, MPH, PhD, Air Force Research Laboratory, Wright-Patterson Air Force Base, OH, USA.
We are indebted to the Millennium Cohort Study participants, without whom these analyses would not be possible. We thank Scott L. Seggerman from the Management Information Division, Defense Manpower Data Center, Seaside, California. Additionally, we thank Lacy Farnell; Isabel Jacobson, MPH; Cynthia LeardMann, MPH; Travis Leleu; Robert Reed, MS; Katherine Snell, Steven Spiegel; Kari Welch, MA; James Whitmer, and Charlene Wong, MPH, from the Department of Defense Center for Deployment Health Research, and Michelle Stoia from the Naval Health Research Center, San Diego, California. We also thank the professionals from the US Army Medical Research and Materiel Command, especially those from the Military Operational Medicine Research Program, Fort Detrick, Maryland. We appreciate the support of the Henry M. Jackson Foundation for the Advancement of Military Medicine, Rockville, Maryland.
This represents report 07–14, supported by the Department of Defense, under work unit no. 60002. The views expressed in this article are those of the authors and do not reflect the official policy or position of the Department of the Navy, Department of the Army, Department of the Air Force, Department of Defense, Department of Veterans Affairs, or the US Government. This research has been conducted in compliance with all applicable federal regulations governing the protection of human subjects in research (Protocol NHRC 2000.007).
- Colditz GA, Martin P, Stampfer MJ, Willett WC, Sampson L, Rosner B, et al: Validation of questionnaire information on risk factors and disease outcomes in a prospective cohort study of women. Am J Epidemiol. 1986, 123 (5): 894-900.PubMedGoogle Scholar
- Bush TL, Miller SR, Golden AL, Hale WE: Self-report and medical record report agreement of selected medical conditions in the elderly. Am J Public Health. 1989, 79 (11): 1554-6.View ArticlePubMedPubMed CentralGoogle Scholar
- Paganini-Hill A, Chao A: Accuracy of recall of hip fracture, heart attack, and cancer: a comparison of postal survey data and medical records. Am J Epidemiol. 1993, 138 (2): 101-6.PubMedGoogle Scholar
- Haapanen N, Miilunpalo S, Pasanen M, Oja P, Vuori I: Agreement between questionnaire data and medical records of chronic diseases in middle-aged and elderly Finnish men and women. Am J Epidemiol. 1997, 145 (8): 762-9.View ArticlePubMedGoogle Scholar
- Bergmann MM, Byers T, Freedman DS, Mokdad A: Validity of self-reported diagnoses leading to hospitalization: a comparison of self-reports with hospital records in a prospective study of American adults. Am J Epidemiol. 1998, 147 (10): 969-77.View ArticlePubMedGoogle Scholar
- Walker MK, Whincup PH, Shaper AG, Lennon LT, Thomson AG: Validation of patient recall of doctor-diagnosed heart attack and stroke: a postal questionnaire and record review comparison. Am J Epidemiol. 1998, 148 (4): 355-61.View ArticlePubMedGoogle Scholar
- Lampe FC, Walker M, Lennon LT, Whincup PH, Ebrahim S: Validity of a self-reported history of doctor-diagnosed angina. J Clin Epidemiol. 1999, 52 (1): 73-81. 10.1016/S0895-4356(98)00146-2.View ArticlePubMedGoogle Scholar
- Martin LM, Leff M, Calonge N, Garrett C, Nelson DE: Validation of self-reported chronic conditions and health services in a managed care population. Am J Prev Med. 2000, 18 (3): 215-8. 10.1016/S0749-3797(99)00158-0.View ArticlePubMedGoogle Scholar
- Horner RD, Cohen HJ, Blazer DG: Accuracy of self-reported stroke among elderly veterans. Aging Ment Health. 2001, 5 (3): 275-81. 10.1080/13607860120065041.View ArticlePubMedGoogle Scholar
- Okura Y, Urban LH, Mahoney DW, Jacobsen SJ, Rodeheffer RJ: Agreement between self-report questionnaires and medical record data was substantial for diabetes, hypertension, myocardial infarction and stroke but not for heart failure. J Clin Epidemiol. 2004, 57 (10): 1096-103. 10.1016/j.jclinepi.2004.04.005.View ArticlePubMedGoogle Scholar
- St Sauver JL, Hagen PT, Cha SS, Bagniewski SM, Mandrekar JN, Curoe AM, et al: Agreement between patient reports of cardiovascular disease and patient medical records. Mayo Clin Proc. 2005, 80 (2): 203-10.View ArticlePubMedGoogle Scholar
- Skinner KM, Miller DR, Lincoln E, Lee A, Kazis LE: Concordance between respondent self-reports and medical records for chronic conditions: experience from the Veterans Health Study. J Ambul Care Manage. 2005, 28 (2): 102-10.View ArticlePubMedGoogle Scholar
- Rahman A, Gibney L, Person SD, Williams OD, Kiefe C, Jolly P, et al: Validity of self-reports of reasons for hospitalization by young adults and risk factors for discordance with medical records: the Coronary Artery Risk Development in Young Adults (CARDIA) Study. Am J Epidemiol. 2005, 162 (5): 491-8. 10.1093/aje/kwi215.View ArticlePubMedGoogle Scholar
- Goldman N, Lin IF, Weinstein M, Lin YH: Evaluating the quality of self-reports of hypertension and diabetes. J Clin Epidemiol. 2003, 56 (2): 148-54. 10.1016/S0895-4356(02)00580-2.View ArticlePubMedGoogle Scholar
- Ryan MA, Smith TC, Smith B, Amoroso P, Boyko EJ, Gray GC, et al: Millennium Cohort: enrollment begins a 21-year contribution to understanding the impact of military service. J Clin Epidemiol. 2007, 60 (2): 181-91. 10.1016/j.jclinepi.2006.05.009.View ArticlePubMedGoogle Scholar
- Gray GC, Chesbrough KB, Ryan MA, Amoroso P, Boyko EJ, Gackstetter GD, et al: The Millennium Cohort Study: a 21-year prospective cohort study of 140,000 military personnel. Mil Med. 2002, 167 (6): 483-8.PubMedGoogle Scholar
- Wells TS, Jacobson IG, Smith TC, Spooner CN, Smith B, Reed RJ, et al: Prior health care utilization as a determinant to enrollment in a 22-year prospective study; the Millennium Cohort Study. Eur J Epidemiol. 2008, 23 (2): 79-87. 10.1007/s10654-007-9216-0.View ArticlePubMedGoogle Scholar
- Smith TC, Smith B, Jacobson IG, Corbeil TE, Ryan MA: Reliability of standard health assessment instruments in a large, population-based cohort study. Ann Epidemiol. 2007, 17 (7): 525-32. 10.1016/j.annepidem.2006.12.002.View ArticlePubMedGoogle Scholar
- LeardMann CA, Smith B, Smith TC, Wells TS, Ryan MA: Smallpox vaccination: comparison of self-reported and electronic vaccine records in the Millennium Cohort Study. Hum Vaccin. 2007, 3 (6): 245-51.View ArticlePubMedGoogle Scholar
- Smith B, Leard CA, Smith TC, Reed RJ, Ryan MA: Anthrax vaccination in the Millennium Cohort: validation and measures of health. Am J Prev Med. 2007, 32 (4): 347-53. 10.1016/j.amepre.2006.12.015.View ArticlePubMedGoogle Scholar
- Smith TC, Jacobson IG, Smith B, Hooper TI, Ryan MA: The occupational role of women in military service: validation of occupation and prevalence of exposures in the Millennium Cohort Study. Int J Environ Health Res. 2007, 17 (4): 271-84. 10.1080/09603120701372243.View ArticlePubMedGoogle Scholar
- Smith B, Wingard DL, Ryan MA, Macera CA, Patterson TL, Slymen DJ: U.S. military deployment during 2001–2006: comparison of subjective and objective data sources in a large prospective health study. Ann Epidemiol. 2007, 17 (12): 976-82. 10.1016/j.annepidem.2007.07.102.View ArticlePubMedGoogle Scholar
- Smith B, Smith TC, Gray GC, Ryan MAK: When epidemiology meets the Internet: Web-based surveys in the Millennium Cohort Study. Am J Epidemiol. 2007, 166 (11): 1345-54. 10.1093/aje/kwm212.View ArticlePubMedGoogle Scholar
- International classification of diseases, 9th revision, clinical modification: 2005, Salt Lake City: Medicode Publication
- National Center for Health Statistics: Conversion Table of New ICD-9-CM Codes. 2005Google Scholar
- Cicchetti DV, Feinstein AR: High agreement but low kappa: II. Resolving the paradoxes. J Clin Epidemiol. 1990, 43 (6): 551-8. 10.1016/0895-4356(90)90159-M.View ArticlePubMedGoogle Scholar
- American Medical Association: Ad Hoc Committee on Health Literacy for the Council on Scientific Affairs. JAMA. 1999, 281: 552-57. 10.1001/jama.281.6.552.View ArticleGoogle Scholar
- Riddle JR, Smith TC, Smith B, Corbeil TE, Engel CC, Wells TS, et al: Millennium Cohort: the 2001–2003 baseline prevalence of mental disorders in the U.S. military. J Clin Epidemiol. 2007, 60 (2): 192-201. 10.1016/j.jclinepi.2006.04.008.View ArticlePubMedGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1471-2288/8/37/prepub
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.