Measuring nominal scale agreement among many raters

Author(s)
Fleiss, J.L.
Year

The statistic kappa was introduced to measure nominal scale agreement between a fixed pair of raters. In this paper kappa is generalized to the case where each of a sample of subjects is rated on a nominal scale by the same number of raters, but where the raters rating one subject are not necessarily the same as those rating another. Large sample standard errors are derived, and a numerical example is given.

Request publication

11 + 7 =
Solve this simple math problem and enter the result. E.g. for 1+3, enter 4.
Pages
378-382
Published in
Psychological Bulletin
76 (5)
Library number
20220128 ST [electronic version only]

Our collection

This publication is one of our other publications, and part of our extensive collection of road safety literature, that also includes the SWOV publications.