场景该定义可能会截然不同,所以下面这个只是供大家一个参考,帮助大家了解Kappa检验的作用,请勿作为Best Practice.
Kappa measures the strength of agreement of the row and column variables, which typically represent the same categorical rating variable as applied by two raters to a set of subjects or items. Note that the minimum value of kappa, when there is complete disagreement, is negative. When there is perfect agreement, all cell counts off the diagonal are 0 and kappa is 1. Kappa is zero when there is no more agreement than would be expected under independence of the row and column variables. Landis and Koch ( Biometrics, 1977) give this interpretation of the range of kappa: