TY - JOUR
T1 - Evaluating random error in clinician-administered surveys
T2 - Theoretical considerations and clinical applications of interobserver reliability and agreement
AU - Bennett, Rebecca J.
AU - Taljaard, Dunay S.
AU - Olaithe, Michelle
AU - Brennan-Jones, Chris
AU - Eikelboom, Robert H.
PY - 2017/9/1
Y1 - 2017/9/1
N2 - Purpose: The purpose of this study is to raise awareness of interobserver concordance and the differences between interobserver reliability and agreement when evaluating the responsiveness of a clinician-administered survey and, specifically, to demonstrate the clinical implications of data types (nominal/categorical, ordinal, interval, or ratio) and statistical index selection (for example, Cohen’s kappa, Krippendorff’s alpha, or interclass correlation). Methods: In this prospective cohort study, 3 clinical audiologists, who were masked to each other’s scores, administered the Practical Hearing Aid Skills Test–Revised to 18 adult owners of hearing aids. Interobserver concordance was examined using a range of reliability and agreement statistical indices. Results: The importance of selecting statistical measures of concordance was demonstrated with a worked example, wherein the level of interobserver concordance achieved varied from “no agreement” to “almost perfect agreement” depending on data types and statistical index selected. Conclusions: This study demonstrates that the methodology used to evaluate survey score concordance can influence the statistical results obtained and thus affect clinical interpretations.
AB - Purpose: The purpose of this study is to raise awareness of interobserver concordance and the differences between interobserver reliability and agreement when evaluating the responsiveness of a clinician-administered survey and, specifically, to demonstrate the clinical implications of data types (nominal/categorical, ordinal, interval, or ratio) and statistical index selection (for example, Cohen’s kappa, Krippendorff’s alpha, or interclass correlation). Methods: In this prospective cohort study, 3 clinical audiologists, who were masked to each other’s scores, administered the Practical Hearing Aid Skills Test–Revised to 18 adult owners of hearing aids. Interobserver concordance was examined using a range of reliability and agreement statistical indices. Results: The importance of selecting statistical measures of concordance was demonstrated with a worked example, wherein the level of interobserver concordance achieved varied from “no agreement” to “almost perfect agreement” depending on data types and statistical index selected. Conclusions: This study demonstrates that the methodology used to evaluate survey score concordance can influence the statistical results obtained and thus affect clinical interpretations.
UR - http://www.scopus.com/inward/record.url?scp=85029717708&partnerID=8YFLogxK
U2 - 10.1044/2017_AJA-16-0100
DO - 10.1044/2017_AJA-16-0100
M3 - Article
C2 - 28768319
AN - SCOPUS:85029717708
SN - 1059-0889
VL - 26
SP - 191
EP - 201
JO - American Journal of Audiology
JF - American Journal of Audiology
IS - 3
ER -