Hello,
I`d love to ask you for a question regarding appropriate sample size of interobserver reliability study, I`ve working on a study that evaluate a reliability of a new classification of flatfoot between three raters in 92 patients. The classification is composed of numbers (0,1,2) and letters (A,B,C,D,E), which in turn make a combination of numbers and letters, e.g. (1A1B1C1D, 1A2B1C, 2E......) Each rater determines a class of each patient`s foot e.g (rater 1: 1A1B1C, rater 2: 1A2B, rater 3: 1B1C).
I used Fleiss`s kappa for interobserver reliability between multiple raters using SPSS which yielded Fleiss Kappa=0.561, p