Variance Estimation of Nominal-Scale Inter-Rater Reliability with Random Selection of Raters

作者: Kilem Li Gwet

DOI: 10.1007/S11336-007-9054-8

关键词:

摘要: Most inter-rater reliability studies using nominal scales suggest the existence of two populations inference: population subjects (collection objects or persons to be rated) and that raters. Consequently, sampling variance coefficient can seen as a result combined effect However, all estimators proposed in literature only account for subject variability, ignoring extra due raters, even though latter may biggest components. Such make statistical inference possible universe. This paper proposes will it infer both universes The consistency these is proved well their validity confidence interval construction. These results are applicable fully crossed designs where each rater must rate subject. A small Monte Carlo simulation study presented demonstrate accuracy large-sample approximations on reasonably samples.

参考文章(38)
Wayne A. Fuller, Cary T. Isaki, SURVEY DESIGN UNDER SUPERPOPULATION MODELS Current Topics in Survey Sampling#R##N#Proceedings of the International Symposium on Survey Sampling Held in Ottawa, Canada, May 7–9, 1980. pp. 199- 226 ,(1981) , 10.1016/B978-0-12-426280-5.50019-8
Richard J. Cook, Kappa and Its Dependence on Marginal Rates Wiley StatsRef: Statistics Reference Online. ,(2014) , 10.1002/9781118445112.STAT05243
John S. Uebersax, William M. Grove, Latent class analysis of diagnostic agreement Statistics in Medicine. ,vol. 9, pp. 559- 572 ,(1990) , 10.1002/SIM.4780090509
Joseph L. Fleiss, Measuring nominal scale agreement among many raters. Psychological Bulletin. ,vol. 76, pp. 378- 382 ,(1971) , 10.1037/H0031619
Rebecca Zwick, Another look at interrater agreement. Psychological Bulletin. ,vol. 103, pp. 374- 378 ,(1988) , 10.1037/0033-2909.103.3.374
Anthony J. Conger, Integration and generalization of kappas for multiple raters. Psychological Bulletin. ,vol. 88, pp. 322- 328 ,(1980) , 10.1037/0033-2909.88.2.322
Emma Bartfay, Allan Donner, Statistical Inferences For Interobserver Agreement Studies With Nominal Outcome Data The Statistician. ,vol. 50, pp. 135- 146 ,(2001) , 10.1111/1467-9884.00266
Kenneth J. Berry, Paul W. Mielke, A Generalization of Cohen's Kappa Agreement Measure to Interval Measurement and Multiple Raters Educational and Psychological Measurement. ,vol. 48, pp. 921- 933 ,(1988) , 10.1177/0013164488484007
Alvan R. Feinstein, Domenic V. Cicchetti, High agreement but low kappa: I. The problems of two paradoxes. Journal of Clinical Epidemiology. ,vol. 43, pp. 543- 549 ,(1990) , 10.1016/0895-4356(90)90158-L