Interrater Agreement Reconsidered: An Alternative to the rwg Indices
- 1 April 2005
- journal article
- research article
- Published by SAGE Publications in Organizational Research Methods
- Vol. 8 (2), 165-184
- https://doi.org/10.1177/1094428105275376
Abstract
For continuous constructs, the most frequently used index of interrater agreement (r wg(1))can be problematic. Typically, rwg(1) is estimated with the assumption that a uniform distribution represents no agreement. The authors review the limitations of this uniform nullr wg(1) index and discuss alternative methods for measuring interrater agreement. A new interrater agreement statistic,a wg(1),is proposed. The authors derive thea wg(1)statistic and demonstrate thatawg(1) is an analogue to Cohen’s kappa, an interrater agreement index for nominal data. A comparison is made between agreement estimates based on the uniformr wg(1)and a wg(1), and issues such as minimum sample size and practical significance levels are discussed. The authors close with recommendations regarding the use ofr wg(1)/rwg(J) when a uniform null is assumed,r wg(1)/rwg(J) indices that do not assume a uniform null,awg(1) / a wg(J)indices, and generalizability estimates of interrater agreement.Keywords
This publication has 44 references indexed in Scilit:
- Accurate tests of statistical significance for rWG and average deviation interrater agreement indexes.Journal of Applied Psychology, 2003
- Estimating Interrater Agreement with the Average Deviation Index: A User’s GuideOrganizational Research Methods, 2002
- Statistical properties of the rWG(J) index of agreement.Psychological Methods, 2001
- On Average Deviation Indices for Estimating Interrater AgreementOrganizational Research Methods, 1999
- Interrater agreement in multi-source performance appraisal: a commentaryJournal of Organizational Behavior, 1997
- Collectivistic orientation in teams: an individual and group-level analysisJournal of Organizational Behavior, 1997
- Confounding among Measures of Leniency and HaloEducational and Psychological Measurement, 1989
- A Generalization of Cohen's Kappa Agreement Measure to Interval Measurement and Multiple RatersEducational and Psychological Measurement, 1988
- Personal constructs, performance schemata, and “folk theories” of subordinate effectiveness: Explorations in an army officer sampleOrganizational Behavior and Human Decision Processes, 1987
- A Coefficient of Agreement for Nominal ScalesEducational and Psychological Measurement, 1960