![K. Gwet's Inter-Rater Reliability Blog : Benchmarking Agreement CoefficientsInter-rater reliability: Cohen kappa, Gwet AC1/AC2, Krippendorff Alpha K. Gwet's Inter-Rater Reliability Blog : Benchmarking Agreement CoefficientsInter-rater reliability: Cohen kappa, Gwet AC1/AC2, Krippendorff Alpha](http://1.bp.blogspot.com/-8lLMKISEeRo/VP2kWbXou8I/AAAAAAAAIFY/8kbySM4sPPM/s1600/altman_benchmark_scale.jpg)
K. Gwet's Inter-Rater Reliability Blog : Benchmarking Agreement CoefficientsInter-rater reliability: Cohen kappa, Gwet AC1/AC2, Krippendorff Alpha
![Agreement plot > Method comparison / Agreement > Statistical Reference Guide | Analyse-it® 6.15 documentation Agreement plot > Method comparison / Agreement > Statistical Reference Guide | Analyse-it® 6.15 documentation](https://analyse-it.com/docs/user-guide/method-comparison/images/agreementplot.png)
Agreement plot > Method comparison / Agreement > Statistical Reference Guide | Analyse-it® 6.15 documentation
![Understanding the calculation of the kappa statistic: A measure of inter-observer reliability | Semantic Scholar Understanding the calculation of the kappa statistic: A measure of inter-observer reliability | Semantic Scholar](https://d3i71xaburhd42.cloudfront.net/352d009ea266e6771ca6c699ab9869d8eba1bb24/3-Table5-1.png)