Weighted specific-category kappa measure of interobserver agreement

Psychol Rep. 2003 Dec;93(3 Pt 2):1283-90. doi: 10.2466/pr0.2003.93.3f.1283.

Abstract

When two observers classify a sample of items using the same categorical scale, and when different disagreements are differentially weighted, the weighted Kappa (Kw) by Cohen may serve as a measure of interobserver agreement. We propose a Kappa-based weighted measure (K(ws)) of agreement on some specific category s, with Kw being a weighted average of all K(ws)s. Therefore, while Cohen's Kw is a summary measure of the overall agreement, the proposed K(ws) provides a measure of the extent to which the observers agree on the specific categories, with both measures being suitable for ordinal categories because of the weights being used. Statistical inferences for K(ws) and its unweighted counterpart are also discussed. A numerical example is provided.

MeSH terms

  • Humans
  • Models, Psychological*
  • Observer Variation*
  • Psychology / methods