Announcement

Collapse
No announcement yet.
X
  • Filter
  • Time
  • Show
Clear All
new posts

  • Kappa or Weighted Kappa in ICC calculation (KAPPAETC)

    Hi
    This is my first times here in the forum and I'm doing my first validation study with almost 0 experience . I have a data of 8 raters assess 31 subject. i used kappaetc code, but I'm a little bet confuse if i have to use wgt(quadratic) or not. I run both codes and get 2 different results, which one is correct.

    kappaetc Rater1_1 Rater2_1 Rater3_1 Rater4_1 Rater5_1 Rater6_1 Rater7_1 Rater8_1

    Interrater agreement Number of subjects = 31
    Ratings per subject = 8
    Number of rating categories = 5
    ------------------------------------------------------------------------------
    | Coef. Std. Err. t P>|t| [95% Conf. Interval]
    ---------------------+--------------------------------------------------------
    Percent Agreement | 0.5334 0.0424 12.59 0.000 0.4469 0.6199
    Brennan and Prediger | 0.4168 0.0529 7.87 0.000 0.3086 0.5249
    Cohen/Conger's Kappa | 0.4073 0.0604 6.75 0.000 0.2840 0.5305
    Scott/Fleiss' Kappa | 0.4042 0.0613 6.60 0.000 0.2791 0.5294
    Gwet's AC | 0.4198 0.0513 8.19 0.000 0.3151 0.5245
    Krippendorff's Alpha | 0.4066 0.0613 6.64 0.000 0.2815 0.5318
    ------------------------------------------------------------------------------

    . kappaetc Rater1_1 Rater2_1 Rater3_1 Rater4_1 Rater5_1 Rater6_1 Rater7_1 Rater8_1, wgt(quadratic)

    Interrater agreement Number of subjects = 31
    (weighted analysis) Ratings per subject = 8
    Number of rating categories = 5
    ------------------------------------------------------------------------------
    | Coef. Std. Err. t P>|t| [95% Conf. Interval]
    ---------------------+--------------------------------------------------------
    Percent Agreement | 0.9505 0.0068 140.00 0.000 0.9366 0.9643
    Brennan and Prediger | 0.8018 0.0272 29.53 0.000 0.7464 0.8573
    Cohen/Conger's Kappa |. 0.7541 0.0586 12.86 0.000 0.6344 0.8738
    Scott/Fleiss' Kappa | 0.7529 0.0594 12.67 0.000 0.6315 0.8742
    Gwet's AC | 0.8136 0.0260 31.30 0.000 0.7605 0.8667
    Krippendorff's Alpha | 0.7538 0.0594 12.68 0.000 0.6325 0.8752
    ------------------------------------------------------------------------------

    Thank you
    Best regards
    Hayder

  • #2
    Originally posted by Hayder Alhusseinawi View Post
    I run both codes and get 2 different results, which one is correct.
    There is seldom one correct answer in statistics. Whether the weighted agreement is plausible depends on the nature of the rating categories. If the 5 categories are ordered, then weighted coefficients might be better than unweighted coefficients. Which weights you use (quadratic, linear, ordinal, etc.) is somewhat of a judgment call. Often, it is useful to report both the unweighted and the weighted coefficients; that is especially true for the raw (percent agreement).

    Comment


    • #3
      Thank you very much Daniel.
      Originally posted by daniel klein View Post

      If the 5 categories are ordered, then weighted coefficients might be better than unweighted coefficients.
      My 5 categories is scale from 1-5 , where 1 meaning very bad, 5 is optimal. Does that mean it's ordered?. What do you think about using of ICC instead of kappa to assess agreement?

      Regards
      Hayder

      Comment


      • #4
        Using ICC is equivalent to Cohens kappa with quadratic weights, so the choice is moot there. Conceptually, linear weights assume a common penalty for any discrepancy, regardless of how different ratings are, and this is usually not desirable if the idea is that ratings should be the same, at least in expectation. Quadratic weights penalize differences by the square of the difference, so a difference of 1 category is less bad than a difference of two categories. Adopting quadratic weights is probably reasonable, but there isn’t enough information to say for sure.

        Comment


        • #5
          Originally posted by Leonardo Guizzetti View Post
          Using ICC is equivalent to Cohens kappa with quadratic weights, so the choice is moot there. Conceptually, linear weights assume a common penalty for any discrepancy, regardless of how different ratings are, and this is usually not desirable if the idea is that ratings should be the same, at least in expectation. Quadratic weights penalize differences by the square of the difference, so a difference of 1 category is less bad than a difference of two categories. Adopting quadratic weights is probably reasonable, but there isn’t enough information to say for sure.
          Thank you very much. I run both ICC and quadratic kappa and get the same results.

          Comment

          Working...
          X