Metrics to evaluate classification models with R codes: Confusion Matrix, Sensitivity, Specificity, Cohen's Kappa Value, Mcnemar's Test - Data Science Vidhya
Importance of Mathews Correlation Coefficient & Cohen's Kappa for Imbalanced Classes | by Sarit Maitra | Medium
Inter-Annotator Agreement (IAA). Pair-wise Cohen kappa and group Fleiss'… | by Louis de Bruijn | Towards Data Science
How does Cohen's Kappa view perfect percent agreement for two raters? Running into a division by 0 problem... : r/AskStatistics
K is for Cohen's Kappa | R-bloggers
Cohens Kappa in R berechnen - Björn Walther
Cohen's Kappa Score. The Kappa Coefficient, commonly… | by Mohammad Badhruddouza Khan | Bootcamp