Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Answering the Call for a Standard Reliability Measure for Coding Data
4.063
Zitationen
2
Autoren
2007
Jahr
Abstract
In content analysis and similar methods, data are typically generated by trained human observers who record or transcribe textual, pictorial, or audible matter in terms suitable for analysis. Conclusions from such data can be trusted only after demonstrating their reliability. Unfortunately, the content analysis literature is full of proposals for so-called reliability coefficients, leaving investigators easily confused, not knowing which to choose. After describing the criteria for a good measure of reliability, we propose Krippendorff's alpha as the standard reliability measure. It is general in that it can be used regardless of the number of observers, levels of measurement, sample sizes, and presence or absence of missing data. To facilitate the adoption of this recommendation, we describe a freely available macro written for SPSS and SAS to calculate Krippendorff's alpha and illustrate its use with a simple example.
Ähnliche Arbeiten
The Measurement of Observer Agreement for Categorical Data
1977 · 77.129 Zit.
STATISTICAL METHODS FOR ASSESSING AGREEMENT BETWEEN TWO METHODS OF CLINICAL MEASUREMENT
1986 · 47.144 Zit.
A Coefficient of Agreement for Nominal Scales
1960 · 40.476 Zit.
A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research
2016 · 26.117 Zit.
Intraclass correlations: Uses in assessing rater reliability.
1979 · 22.736 Zit.