Reliability

From WikiMSK

This article is still missing information.

Interobserver reliability is the agreement between two observers on the findings of a given diagnostic test. The 2x2 table is used for evaluating reliability with tests with two possible categorical outcomes (yes/no, positive/negative etc). The table can be used to calculate a simple measure of agreement called the observed agreement (po). However the observed agreement doesn't take into account chance, and what we really want to know is the agreement beyond chance.

B
A
Yes No
Yes a b
No c d

Cohen's kappa

Definition.png
Definition
Cohen's kappa
A chance corrected standardised measure of agreement between categorical ratings produced by two observers

Cohen's kappa (κ) coefficient is a commonly used mathematical method of taking into account chance. The κ value ranges from 0 to 1, with 0 representing the amount of agreement expected by chance. It can theoretically be negative, signifying worse than chance, but this is unusual. The score gives some sort of indicator as to the interobserver reliability by taking into account chance. The numbers are often assigned qualitative descriptors to facilitate understanding. Good tests have a kappa score of at least 0.6.

Qualitative descriptions of Kappa Values
Kappa Value Bogduk Landis & Koch Cicchetti & Sparrow
0.8 - 1.0 Very good Almost perfect
0.6 - 0.8 Good Substantial Excellent
0.4 - 0.6 Moderate Moderate Fair
0.2 - 0.4 Slight Fair
0.0 - 0.2 Poor Slight
<0.0 Less than chance Poor Poor

A hypothetical narrative illustration of why chance is important is provided. Observer A and observer B agree by chance in 60% of cases. We want to know their agreement in the remaining 40% of cases. Let's say their observed agreement overall is 80%, we need to subtract the 60% that is due to chance, leaving 20% agreement above chance. Now we had 40% remaining in total above chance, but the observers only agreed 20% of the time above chance. Therefore the true agreement is 20%/40%, i.e. 50%, much less than the observed 80%.

Mathematics

Interobserver Agreement and the κ (Kappa) Statistic. Top: Conventional 2 x 2 table for Kappa calculation. Bottom: Sample problem, with observed agreement of 80%, chance agreement of 74%, and Kappa statistic of 0.23

If two observers examine the same patients independently this can be displayed in a 2 x 2 table. Observer A finds a positive sign in w1 patients, and negative sign in w2 patients. Observer B finds a positive sign in y1 patients and negative sign in y2 patients. The two observers agree that the sign if positive in a patients, and is negative in d patients. The observed agreement (po) is calculated:

[math]\displaystyle{ p_o=\frac{a+d}{N} }[/math]

To calculate the κ-statistic the first step is calculating the agreement that would have occurred simply due to chance. Observer A found that among all patients W1/N have the sign. Observer B finds that the patients y1 have the sign, and among y1 patients Observer A would find the sign by chance in (w1y1/N) patients. This number represents the number of patients in which both observers agree by chance. Likewise, both observers would agree that the sign is negative by chance in (w2y2/N) patients.

The expected chance of agreement (pe) is the sum of these two values, divided by N.

[math]\displaystyle{ p_e=\frac{w_1y_1+w_2y_2}{N^2} }[/math]

If both observers agree that a finding is rare (approaching 0) or very common (approaching N) then the expected chance of agreement (pe) approaches 100%

The κ-statistic is the difference in observed agreement (po) and the expected chance of agreement (pe) x (po - pe), divided by the maximal increment that could have been observed if the observed agreement (po) was 100% (1 - pe)

[math]\displaystyle{ \kappa=\frac{p_o-p_e}{1-p_e} }[/math]

The sample problem in the figure represents a fictional study of 100 patients with ischial tuberosity tenderness. Both observers agree that tenderness is present in 5 patients and absent in 75 patients.

Observed agreement (po) is (5 + 75)/100 = 0.8.
By chance the observers would agree the sign was present in ((5+5) x (15+5))/100 patients = 2 patients
By chance the observers would agree the sign was not present in ((75+15) x (75+5))/100 patients = 72 patients.
Expected chance of agreement (pe) is (2 + 72)/100 patients = 0.74.
The κ-statistic is (0.80 – 0.74)/(1 – 0.74) = (0.06)/(0.26) = 0.23.

Limitations

The theoretical range of kappa is -1 to +1, however this is only theoretically true. The assumptions that the formula makes about the independence of rater observations is not well supported, and it may underestimate the agreement. It is also less useful when the condition tested is very high or very low, but it is still a very useful statistic.

Resources

See Also

Validity

Bibliography

  • McGee, Steven R. Evidence-based physical diagnosis. Philadelphia: Elsevier/Saunders, 2012.