Fleiss' kappa is a generalisation of Scott's pi statistic, a statistical measure of inter-rater reliability. It is also related to Cohen's kappa statistic. Whereas Scott's pi and Cohen's kappa work for only two raters, Fleiss' kappa works for any number of raters giving categorical ratings (see nominal data), to a fixed number of items. It can be interpreted as expressing the extent to which the observed amount of agreement among raters exceeds what would be expected if all raters made their ratings completely randomly. Agreement can be thought of as follows, if a fixed number of people assign numerical ratings to a number of items then the kappa will give a measure for how consistent the ratings are.
意思是:
Fleiss kappa和Cohen Kappa都是描述评价者之间的一致性,但传统的Cohen kappa适用于2个评价者,而Fleiss kappa适用于多个评价者,评价结果为分类变量的一致性。
数据如下,松哥模拟虚构仅做演示:3个专家对某10位受试者打分情况,是分析三位专家评价结果的一致性?
SPSS26.0实现过程如下:
分析-刻度-可靠性分析,弹出下图,将3位专家放入评级,确定即可。
结果呈现:Flsiss kappa=0.464,P=0.011。
Kappa取值的一致性解释:
联系客服