106 units are all assessed by the same 5 readers. Units were judged either positive or negative (dichotomous outcome).
What is the best applied statistical test to look at interrater agreement?
1) Is ICC (two-way random effect model, single rater, agreement) usefull, or is that only to apply to continous or categorical data with >2 possible ratings?
2) Is Fleiss kappa the statistical test of choice, and if so,
a) is the correct stata command kappa pos neg (when data are organised in: Column 1; subject id, column 2; number of positive reads (pos), column 3; number of negative reads (neg))
b) Does this test allow that oner reader is the basis/gold standard (the one others should agree with)
Will be very grateful for your input
What is the best applied statistical test to look at interrater agreement?
1) Is ICC (two-way random effect model, single rater, agreement) usefull, or is that only to apply to continous or categorical data with >2 possible ratings?
2) Is Fleiss kappa the statistical test of choice, and if so,
a) is the correct stata command kappa pos neg (when data are organised in: Column 1; subject id, column 2; number of positive reads (pos), column 3; number of negative reads (neg))
b) Does this test allow that oner reader is the basis/gold standard (the one others should agree with)
Will be very grateful for your input
Comment