Fleiss' kappa

Fleiss' kappa (named after Joseph L. Fleiss) is a statistical measure for assessing the reliability of agreement between a fixed number of raters when assigning categorical ratings to a number of items or classifying items. This contrasts with other kappas such as Cohen's kappa, which only work when assessing the agreement between two raters. The measure calculates the degree of agreement in classification over that which would be expected by chance. There is no generally agreed-upon measure of significance, although guidelines have been given.

Fleiss' kappa can be used only with binary or nominal-scale ratings. No version is available for ordered-categorical ratings.

Introduction

Fleiss' kappa is a generalisation of Scott's pi statistic, a statistical measure of inter-rater reliability. It is also related to Cohen's kappa statistic and Youden's J statistic which may be more appropriate in certain instances. Whereas Scott's pi and Cohen's kappa work for only two raters, Fleiss' kappa works for any number of raters giving categorical ratings, to a fixed number of items. It can be interpreted as expressing the extent to which the observed amount of agreement among raters exceeds what would be expected if all raters made their ratings completely randomly. It is important to note that whereas Cohen's kappa assumes the same two raters have rated a set of items, Fleiss' kappa specifically allows that although there are a fixed number of raters (e.g., three), different items may be rated by different individuals (Fleiss, 1971, p.378). That is, Item 1 is rated by Raters A, B, and C; but Item 2 could be rated by Raters D, E, and F.

Agreement can be thought of as follows, if a fixed number of people assign numerical ratings to a number of items then the kappa will give a measure for how consistent the ratings are. The kappa, \kappa\,, can be defined as,

(1)

\kappa = \frac{\bar{P} - \bar{P_e}}{1 - \bar{P_e}}

The factor 1 - \bar{P_e} gives the degree of agreement that is attainable above chance, and, \bar{P} - \bar{P_e} gives the degree of agreement actually achieved above chance. If the raters are in complete agreement then \kappa = 1~. If there is no agreement among the raters (other than what would be expected by chance) then \kappa \le 0.

An example of the use of Fleiss' kappa may be the following: Consider fourteen psychiatrists are asked to look at ten patients. Each psychiatrist gives one of possibly five diagnoses to each patient. These are compiled into a matrix, and Fleiss' kappa can be computed from this matrix (see example below) to show the degree of agreement between the psychiatrists above the level of agreement expected by chance.

Equations

Let N be the total number of subjects, let n be the number of ratings per subject, and let k be the number of categories into which assignments are made. The subjects are indexed by i = 1, ... N and the categories are indexed by j = 1, ... k. Let nij represent the number of raters who assigned the i-th subject to the j-th category.

First calculate pj, the proportion of all assignments which were to the j-th category:

(2)

p_{j} = \frac{1}{N n} \sum_{i=1}^N n_{i j},\quad\quad 1 = \frac{1}{n} \sum_{j=1}^k n_{i j}

Now calculate P_{i}\,, the extent to which raters agree for the i-th subject (i.e., compute how many rater--rater pairs are in agreement, relative to the number of all possible rater--rater pairs):

(3)

P_{i} = \frac{1}{n(n - 1)} \sum_{j=1}^k n_{i j} (n_{i j} - 1)
      = \frac{1}{n(n - 1)} \sum_{j=1}^k (n_{i j}^2 - n_{i j})
      = \frac{1}{n(n - 1)} [(\sum_{j=1}^k n_{i j}^2) - (n)]

Now compute \bar{P}, the mean of the P_i\,'s, and \bar{P_e} which go into the formula for \kappa\,:

(4)

\bar{P} = \frac{1}{N} \sum_{i=1}^N P_{i}
       = \frac{1}{N n (n - 1)} (\sum_{i=1}^N \sum_{j=1}^k n_{i j}^2 - N n)

(5)

\bar{P_e} = \sum_{j=1}^k p_{j} ^2

Worked example

n_{ij} 1 2 3 4 5 P_i\,
1 0 0 0 0 14 1.000
2 0 2 6 4 2 0.253
3 0 0 3 5 6 0.308
4 0 3 9 2 0 0.440
5 2 2 8 1 1 0.330
6 7 7 0 0 0 0.462
7 3 2 6 3 0 0.242
8 2 5 3 2 2 0.176
9 6 5 2 1 0 0.286
10 0 2 2 3 7 0.286
Total 20 28 39 21 32
p_j\, 0.143 0.200 0.279 0.150 0.229
Table of values for computing the worked example

In the following example, fourteen raters (n) assign ten "subjects" (N) to a total of five categories (k). The categories are presented in the columns, while the subjects are presented in the rows. Each cell is filled with the number of raters who agreed that a certain subject belongs to a certain category.

Data

See table to the right.

N = 10, n = 14, k = 5

Sum of all cells = 140
Sum of P_{i}\, = 3.780

Calculations

For example, taking the first column,

p_1 = \frac{ 0+0+0+0+2+7+3+2+6+0 }{140} = 0.143

And taking the second row,

P_2 = \frac{1}{14(14 - 1)} \left(0^2 + 2^2 + 6^2 + 4^2 + 2^2 - 14\right) = 0.253

In order to calculate \bar{P}, we need to know the sum of P_i,

\sum_{i=1}^N P_{i}= 1.000 + 0.253 + \cdots + 0.286 + 0.286 = 3.780

Over the whole sheet,

\bar{P} = \frac{1}{(10)} (3.780) = 0.378
\bar{P}_{e} = 0.143^2 + 0.200^2 + 0.279^2 + 0.150^2 + 0.229^2 = 0.213
\kappa = \frac{0.378 - 0.213}{1 - 0.213} = 0.210

Interpretation

Landis and Koch (1977) gave the following table for interpreting \kappa values. This table is however by no means universally accepted. They supplied no evidence to support it, basing it instead on personal opinion. It has been noted that these guidelines may be more harmful than helpful, as the number of categories and subjects will affect the magnitude of the value. The kappa will be higher when there are fewer categories.

\kappa Interpretation
< 0 Poor agreement
0.01 0.20 Slight agreement
0.21 0.40 Fair agreement
0.41 0.60 Moderate agreement
0.61 0.80 Substantial agreement
0.81 1.00 Almost perfect agreement

See also

The Wikibook Algorithm implementation has a page on the topic of: Fleiss' kappa

Notes

  1. ^ Fleiss, J. L. (1971) pp. 378382
  2. ^ Scott, W. (1955) pp. 321325
  3. ^ Powers, D. M. W. (2011)
  4. ^ Powers, D. M. W. (2012)
  5. ^ Landis, J. R. and Koch, G. G. (1977) pp. 159174
  6. ^ Gwet, K. L. (2014, chapter 6)
  7. ^ Sim, J. and Wright, C. C. (2005) pp. 257268

References

Further reading

External links

This article is issued from Wikipedia - version of the Thursday, January 21, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.