Skip to main content
. 2019 Jun 26;14:65. doi: 10.1186/s13000-019-0839-8

Table 2.

Inter-observer agreement in each observation method (Kappa coefficient)

Scanner A Scanner B Scanner C Scanner D Microscope
Observer 1 vs 2 0.677 0.770 0.814 0.815 0.735
Observer 1 vs 3 0.775 0.790 0.872 0.745 0.833
Observer 1 vs 4 0.885 0.788 0.852 0.865 0.779
Observer 1 vs 5 0.864 0.799 0.879 0.820 0.781
Observer 2 vs 3 0.621 0.807 0.834 0.667 0.742
Observer 2 vs 4 0.693 0.763 0.840 0.791 0.723
Observer 2 vs 5 0.699 0.802 0.840 0.786 0.666
Observer 3 vs 4 0.773 0.850 0.818 0.819 0.831
Observer 3 vs 5 0.765 0.862 0.845 0.785 0.789
Observer 4 vs 5 0.886 0.833 0.864 0.905 0.743