Log in account

Features of Сohen’s kappa coefficient as a measure of expert opinion agreement

УДК 004.89

ISSN 2709-4707

Category: Information and communication technologies

Modern metrics for evaluating agreement coefficients between the experimental results and expert opinion are compared, and the possibility of using these metrics in experimental research in automatic text processing by machine learning methods is assessed. The choice of Cohen’s kappa coefficient as a measure of expert opinion agreement in the NLP and Text Mining problems is justified. An example of using Cohen’s kappa coefficient for evaluating the level of agreement between the opinion of an expert and the results of ML classification and the measure of agreement of expert opinions in the alignment of sentences of the Kazakh-Russian parallel corpus is given. Based on this analysis, it is proved that Cohen’s kappa coefficient is one of the best statistical methods for determining the level of agreement in experimental studies due to its ease of use, computing simplicity, and high accuracy of the results.

Key words: Text Mining, NLP, Cohen’s kappa statistic, agreement statistic, text classification with machine learning, parallel corpus.