Currently, computational linguists and cognitive scientists working in the area of discourse and dialogue argue that their subjective judgments are reliable using several different statistics, none of which are easily interpretable or comparable to each other. Meanwhile, researchers in content analysis have already experienced the same difficulties and come up with a solution in the kappa statistic. We discuss what is wrong with reliability measures as they are currently used for discourse and dialogue work in computational linguistics and cognitive science, and argue that we would be better off as a field adopting techniques from content analysis.
Assessing Agreement on Classification Tasks: The Kappa Statistic
Published 1996 in International Conference on Computational Logic
ABSTRACT
PUBLICATION RECORD
- Publication year
1996
- Venue
International Conference on Computational Logic
- Publication date
1996-02-27
- Fields of study
Sociology, Linguistics, Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-14 of 14 references · Page 1 of 1