메뉴 건너뛰기




Volumn 82, Issue 6, 2014, Pages 1219-1227

Interrater agreement statistics with skewed data: Evaluation of alternatives to Cohen's kappa

Author keywords

Behavior observation; Diagnosis; Interrater agreement; Low base rate; Skew

Indexed keywords

ARTICLE; BEHAVIORAL SCIENCE; CLINICAL RESEARCH; COHEN KAPPA; HUMAN; INTERRATER AGREEMENT STATISTICS; INTERRATER RELIABILITY; KAPPA STATISTICS; MONTE CARLO METHOD; SIMULATION; OBSERVER VARIATION; REPRODUCIBILITY; STATISTICS;

EID: 84925639509     PISSN: 0022006X     EISSN: 19392117     Source Type: Journal    
DOI: 10.1037/a0037489     Document Type: Article
Times cited : (73)

References (31)
  • 2
    • 0013565775 scopus 로고    scopus 로고
    • Detecting sequential patterns and determining their reliability with fallible observers
    • Bakeman, R., McArthur, D., Quera, V., & Robinson, B. F. (1997). Detecting sequential patterns and determining their reliability with fallible observers. Psychological Methods, 2, 357-370. doi:10.1037/1082-989X .2.4.357
    • (1997) Psychological Methods , vol.2 , pp. 357-370
    • Bakeman, R.1    McArthur, D.2    Quera, V.3    Robinson, B.F.4
  • 3
    • 33845891920 scopus 로고    scopus 로고
    • The design of simulation studies in medical statistics
    • Burton, A., Altman, D. G., Royston, P., & Holder, R. L. (2006). The design of simulation studies in medical statistics. Statistics in Medicine, 25, 4279-4292. doi:10.1002/sim.2673
    • (2006) Statistics in Medicine , vol.25 , pp. 4279-4292
    • Burton, A.1    Altman, D.G.2    Royston, P.3    Holder, R.L.4
  • 4
    • 0028600665 scopus 로고
    • Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology
    • Cicchetti, D. V. (1994). Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychological Assessment, 6, 284-290. doi:10.1037/1040-3590.6.4.284
    • (1994) Psychological Assessment , vol.6 , pp. 284-290
    • Cicchetti, D.V.1
  • 6
    • 84973587732 scopus 로고
    • A coefficient of agreement for nominal scales
    • Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20, 37-46. doi:10.1177/ 00131644600
    • (1960) Educational and Psychological Measurement , vol.20 , pp. 37-46
    • Cohen, J.1
  • 7
    • 58149412516 scopus 로고
    • Weighed kappa: Nominal scale agreement with provision for scaled disagreement or partial credit
    • Cohen, J. (1968). Weighed kappa: Nominal scale agreement with provision for scaled disagreement or partial credit. Psychological Bulletin, 70, 213-220. doi:10.1037/h0026256
    • (1968) Psychological Bulletin , vol.70 , pp. 213-220
    • Cohen, J.1
  • 8
    • 84879319237 scopus 로고    scopus 로고
    • Factors affecting intercoder reliability: A Monte Carlo experiment
    • Feng, G. C. (2013). Factors affecting intercoder reliability: A Monte Carlo experiment. Quality & Quantity, 47, 2959-2982. doi:10.1007/s11135-012-9745-9
    • (2013) Quality & Quantity , vol.47 , pp. 2959-2982
    • Feng, G.C.1
  • 10
    • 84890425821 scopus 로고    scopus 로고
    • Diagnostic system for personality disorders in an outpatient clinical sample
    • diagnostic system for personality disorders in an outpatient clinical sample. Journal of Abnormal Psychology, 122, 1057-1069. doi:10.1037/ a0034878
    • Journal of Abnormal Psychology , vol.122 , pp. 1057-1069
  • 11
    • 84877819479 scopus 로고    scopus 로고
    • Interrater agreement and interrater reliability: Key concepts, approaches, and applications
    • Gisev, N., Bell, J. S., & Chen, T. F. (2013). Interrater agreement and interrater reliability: Key concepts, approaches, and applications. Research in Social and Administrative Pharmacy, 9, 330-338. doi: 10.1016/j.sapharm.2012.04.004
    • (2013) Research in Social and Administrative Pharmacy , vol.9 , pp. 330-338
    • Gisev, N.1    Bell, J.S.2    Chen, T.F.3
  • 12
    • 85006639034 scopus 로고    scopus 로고
    • Kappa statistic is not satisfactory for assessing the extent of agreement between raters
    • Gwet, K. (2002). Kappa statistic is not satisfactory for assessing the extent of agreement between raters. Retrieved from http://www.agreestat.com/ research-papers/kappa-statistic-is-not-satisfactory.pdf
    • (2002) Retrieved from
    • Gwet, K.1
  • 13
    • 44349101004 scopus 로고    scopus 로고
    • Computing inter-rater reliability and its variance in the presence of high agreement
    • Gwet, K. L. (2008). Computing inter-rater reliability and its variance in the presence of high agreement. British Journal of Mathematical and Statistical Psychology, 61, 29-48. doi:10.1348/000711006X1
    • (2008) British Journal of Mathematical and Statistical Psychology , vol.61 , pp. 29-48
    • Gwet, K.L.1
  • 16
    • 84890691291 scopus 로고    scopus 로고
    • The effect of the raters' marginal distributions on their matched agreement: A rescaling framework for interpreting kappa
    • Karelitz, T. M., & Budescu, D. V. (2013). The effect of the raters' marginal distributions on their matched agreement: A rescaling framework for interpreting kappa. Multivariate Behavioral Research, 48, 923-952. doi:10.1080/00273171.2013.830064
    • (2013) Multivariate Behavioral Research , vol.48 , pp. 923-952
    • Karelitz, T.M.1    Budescu, D.V.2
  • 19
    • 84995012914 scopus 로고
    • A review of statistical methods in the analysis of data arising from observer reliability studies (Part I)
    • Landis, J. R., & Koch, G. G. (1975a). A review of statistical methods in the analysis of data arising from observer reliability studies (Part I). Statistica Neerlandica, 29, 101-123. doi:10.1111/j.1467-9574.1975.tb00254.x
    • (1975) Statistica Neerlandica , vol.29 , pp. 101-123
    • Landis, J.R.1    Koch, G.G.2
  • 20
    • 84995114573 scopus 로고
    • A review of statistical methods in the analysis of data arising from observer reliability studies (Part II)
    • Landis, J. R., & Koch, G. G. (1975b). A review of statistical methods in the analysis of data arising from observer reliability studies (Part II). Statistica Neerlandica, 29, 151-161. doi:10.1111/j.1467-9574.1975.tb00259.x
    • (1975) Statistica Neerlandica , vol.29 , pp. 151-161
    • Landis, J.R.1    Koch, G.G.2
  • 21
    • 0017381382 scopus 로고
    • An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers
    • Landis, J. R., & Koch, G. G. (1977). An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers. Biometrics, 33, 363-374. doi:10.2307/2529786
    • (1977) Biometrics , vol.33 , pp. 363-374
    • Landis, J.R.1    Koch, G.G.2
  • 22
    • 33745784824 scopus 로고    scopus 로고
    • Can minimally trained observers provide valid global ratings
    • Lorber, M. F. (2006). Can minimally trained observers provide valid global ratings? Journal of Family Psychology, 20, 335-338. doi:10.1037/0893-3200.20.2.335
    • (2006) Journal of Family Psychology , vol.20 , pp. 335-338
    • Lorber, M.F.1
  • 23
    • 85043870731 scopus 로고    scopus 로고
    • On the practice of dichotomization of quantitative variables
    • MacCallum, R. C., Zhang, S., Preacher, K. J., & Rucker, D. D. (2002). On the practice of dichotomization of quantitative variables. Psychological Methods, 7, 19-40. doi:10.1037/1082-989X.7.1.19
    • (2002) Psychological Methods , vol.7 , pp. 19-40
    • Maccallum, R.C.1    Zhang, S.2    Preacher, K.J.3    Rucker, D.D.4
  • 24
    • 0000135659 scopus 로고    scopus 로고
    • Forming inferences about some intraclass correlation coefficients
    • McGraw, K. O., & Wong, S. P. (1996). Forming inferences about some intraclass correlation coefficients. Psychological Methods, 1, 30-46. doi:10.1037/1082-989X.1.1.30
    • (1996) Psychological Methods , vol.1 , pp. 30-46
    • McGraw, K.O.1    Wong, S.P.2
  • 25
    • 0000241635 scopus 로고
    • Interobserver agreement, reliability, and generalizability of data collected in observational studies
    • Mitchell, S. (1979). Interobserver agreement, reliability, and generalizability of data collected in observational studies. Psychological Bulletin, 86, 376-390. doi:10.1037/0033-2909.86.2.376
    • (1979) Psychological Bulletin , vol.86 , pp. 376-390
    • Mitchell, S.1
  • 26
    • 33748063869 scopus 로고
    • Reliability of content analysis: The case of nominal scale coding
    • Scott, W. A. (1955). Reliability of content analysis: The case of nominal scale coding. Public Opinion Quarterly, 19, 321-325. doi:10.1086/ 266577
    • (1955) Public Opinion Quarterly , vol.19 , pp. 321-325
    • Scott, W.A.1
  • 27
    • 33748088163 scopus 로고    scopus 로고
    • Including omission mistakes in the calculation of Cohen's kappa and an analysis of the coefficient's paradox features
    • Simon, P. (2006). Including omission mistakes in the calculation of Cohen's kappa and an analysis of the coefficient's paradox features. Educational and Psychological Measurement, 66, 765-777. doi:10.1177/ 00131644052
    • (2006) Educational and Psychological Measurement , vol.66 , pp. 765-777
    • Simon, P.1
  • 28
    • 33745618426 scopus 로고    scopus 로고
    • Creating and field-testing child maltreatment definitions: Improving the reliability of substantiation determinations
    • Slep, A. M. S., & Heyman, R. E. (2006). Creating and field-testing child maltreatment definitions: Improving the reliability of substantiation determinations. Child Maltreatment, 11, 217-236. doi:10.1177/ 10775595062
    • (2006) Child Maltreatment , vol.11 , pp. 217-236
    • Slep, A.M.S.1    Heyman, R.E.2
  • 29
    • 0021845522 scopus 로고
    • A proposed solution to the base rate problem in the kappa statistic
    • Spitznagel, E. L., & Helzer, J. E. (1985). A proposed solution to the base rate problem in the kappa statistic. Archives of General Psychiatry, 42, 725-728. doi:10.1001/archpsyc.1985.01790300093
    • (1985) Archives of General Psychiatry , vol.42 , pp. 725-728
    • Spitznagel, E.L.1    Helzer, J.E.2
  • 30
    • 0001319997 scopus 로고
    • Diversity of decision-making models and the measurement of interrater agreement
    • Uebersax, J. S. (1987). Diversity of decision-making models and the measurement of interrater agreement. Psychological Bulletin, 101, 140-146. doi:10.1037/0033-2909.101.1.140
    • (1987) Psychological Bulletin , vol.101 , pp. 140-146
    • Uebersax, J.S.1
  • 31
    • 20444420190 scopus 로고    scopus 로고
    • The dependence of Cohen's kappa on the prevalence does not matter
    • Vach, W. (2005). The dependence of Cohen's kappa on the prevalence does not matter. Journal of Clinical Epidemiology, 58, 655-661. doi: 10.1016/j.jclinepi.2004.02.021
    • (2005) Journal of Clinical Epidemiology , vol.58 , pp. 655-661
    • Vach, W.1


* 이 정보는 Elsevier사의 SCOPUS DB에서 KISTI가 분석하여 추출한 것입니다.