Skip to main content

Table 5 Simulated chart-specific inter-rater reliability coefficients

From: Examining intra-rater and inter-rater response agreement: A medical chart abstraction study of a community-based asthma care program

   Kappa* Sensitivity§ Specificity§
Simulated Charts Percentage agreement (%) κ 95% CI Estimate 95% CI Estimate 95% CI
1 88 0.75 (0.72–0.78) 0.92 (0.86–0.96) 0.90 (0.85–0.94)
2 91 0.78 (0.75–0.82) 0.99 (0.91–1.00) 0.90 (0.85–0.93)
3 90 0.70 (0.66–0.74) 0.84 (0.73–0.91) 0.96 (0.92–0.98)
4 88 0.76 (0.73–0.79) 0.96 (0.90–0.99) 0.87 (0.82–0.91)
5 88 0.74 (0.71–0.78) 0.91 (0.82–0.96) 0.88 (0.83–0.92)
6 85 0.69 (0.65–0.73) 0.93 (0.85–0.97) 0.85 (0.79–0.90)
7 88 0.76 (0.73–0.79) 0.91 (0.85–0.95) 0.91 (0.85–0.95)
8 86 0.72 (0.69–0.76) 0.85 (0.77–0.90) 0.85 (0.79–0.90)
  1. * Chi-square test statistic for homogeneity = 22, df = 7, p = 0.003
  2. § Sensitivity and specificity are calculated to compare assessments of all raters against the gold standard.
  3. Abbreviations: κ kappa statistic, CI confidence interval