Inter-rater reliability

Inter-rater reliability

In statistics, inter-rater reliability, inter-rater agreement, or concordance is the degree of agreement among raters. It gives a score of how much homogeneity, or consensus, there is in the ratings given by judges. It is useful in refining the tools given to human judges, for example by determining if a particular scale is appropriate for measuring a particular variable. If various raters do not agree, either the scale is defective or the raters need to be re-trained.

There are a number of statistics which can be used to determine inter-rater reliability. Different statistics are appropriate for different types of measurement. Some options are: joint-probability of agreement, Cohen's kappa and the related Fleiss' kappa, inter-rater correlation, concordance correlation coefficient and intra-class correlation.

Contents

Sources of inter-rater disagreement

Different raters can disagree about measurement results from the same object by e.g. variations in the procedures of carrying out the experiment, interpreting the results and, subsequently, presenting them. All these stages may be affected by experimenter's bias, that is, a tendency to deviate towards what is expected by the rater. When interpreting and presenting the results, there may be inter-rater variations in digit preference, that is, preferences differ whether to round off a value to a lower one or a higher one.

The philosophy of inter-rater agreement

There are several operational definitions [1] of "inter-rater reliability" in use by Examination Boards, reflecting different viewpoints about what is reliable agreement between raters.

There are three operational definitions of agreement:

1. Reliable raters agree with the "official" rating of a performance.

2. Reliable raters agree with each other about the exact ratings to be awarded.

3. Reliable raters agree about which performance is better and which is worse.

These combine with two operational definitions of behavior:

A. Reliable raters are automatons, behaving like "rating machines". This category includes rating of essays by computer [2]. This behavior can be evaluated by Generalizability theory.

B. Reliable raters behave like independent witnesses. They demonstrate their independence by disagreeing slightly. This behavior can be evaluated by the Rasch model.

Joint probability of agreement

The joint-probability of agreement is probably the most simple and least robust measure. It is the number of times each rating (e.g. 1, 2, ... 5) is assigned by each rater divided by the total number of ratings. It assumes that the data are entirely nominal. It does not take into account that agreement may happen solely based on chance. Some question, though, whether there is a need to 'correct' for chance agreement; and suggest that, in any case, any such adjustment should be based on an explicit model of how chance and error affect raters' decisions.[3]

When the number of categories being used is small (e.g. 2 or 3), the likelihood for 2 raters to agree by pure chance increases dramatically. This is because both raters must confine themselves to the limited number of options available, which impacts the overall agreement rate, and not necessarily their propensity for "intrinsic" agreement (is considered "intrinsic" agreement, an agreement not due to chance). Therefore, the joint probability of agreement will remain high even in the absence of any "intrinsic" agreement among raters[4]. A useful inter-rater reliability coefficient is expected to (1) be close to 0, when there is no "intrinsic" agreement, and (2) to increase as the "intrinsic" agreement rate improves. Most chance-corrected agreement coefficients achieve the first objective[5]. However, the second objective is not achieved by many known chance-corrected measures[6].

Kappa statistics

Main articles: Cohen's kappa, Fleiss' kappa

Cohen's kappa[7], which works for two raters, and Fleiss' kappa[8], an adaptation that works for any fixed number of raters, improve upon the joint probability in that they take into account the amount of agreement that could be expected to occur through chance. They suffer from the same problem as the joint-probability in that they treat the data as nominal and assume the ratings have no natural ordering. If the data do have an order, the information in the measurements is not fully taken advantage of.

Correlation coefficients

Main articles: Pearson product-moment correlation coefficient, Spearman's rank correlation coefficient

Either Pearson's r or Spearman's ρ can be used to measure pairwise correlation among raters using a scale that is ordered. Pearson assumes the rating scale is continuous; Spearman assumes only that it is ordinal. If more than two raters are observed, an average level of agreement for the group can be calculated as the mean of the r (or ρ) values from each possible pair of raters.

Both the Pearson and Spearman coefficients consider only relative position. For example, (1, 2, 1, 3) is considered perfectly correlated with (2, 3, 2, 4).

Intra-class correlation coefficient

Another way of performing reliability testing is to use the intra-class correlation coefficient (ICC) [9]. There are several types of this and one is defined as, "the proportion of variance of an observation due to between-subject variability in the true scores".[10] The range of the ICC may be between 0.0 and 1.0 (an early definition of ICC could be between −1 and +1). The ICC will be high when there is little variation between the scores given to each item by the raters, e.g. if all raters give the same, or similar scores to each of the items. The ICC is an improvement over Pearson's r and Spearman's ρ, as it takes into account of the differences in ratings for individual segments, along with the correlation between raters.

Limits of agreement

Bland–Altman plot

Another approach to agreement (useful when there are only two raters) is to calculate the mean of the differences between the two raters. The confidence limits around the mean provide insight into how much random variation may be influencing the ratings. If the raters tend to agree, the mean will be near zero. If one rater is usually higher than the other by a consistent amount, the mean will be far from zero, but the confidence interval will be narrow. If the raters tend to disagree, but without a consistent pattern of one rating higher than the other, the mean will be near zero but the confidence interval will be wide.

Bland and Altman [11] have expanded on this idea by graphing the difference of each point, the mean difference, and the confidence limits on the vertical against the average of the two ratings on the horizontal. The resulting Bland–Altman plot demonstrates not only the overall degree of agreement, but also whether the agreement is related to the underlying value of the item. For instance, two raters might agree closely in estimating the size of small items, but disagree about larger items.

Krippendorff’s Alpha

Krippendorff's alpha[12] is a versatile and general statistical measure for assessing the agreement achieved when multiple raters describe a set of objects of analysis in terms of the values of a variable. Alpha emerged in content analysis where textual units are categorized by trained coders and is used in counseling and survey research where experts code open-ended interview data into analyzable terms, in psychometrics where individual attributes are tested by multiple methods, or in observational studies where unstructured happenings are recorded for subsequent analysis.

Notes

  1. ^ Saal, F.E., Downey, R.G. and Lahey, M.A (1980) "Rating the Ratings: Assessing the Psychometric Quality of Rating Data" in Psychological Bulletin. Vol. 88, No. 2, pp. 413–428
  2. ^ Page, E. B, and Petersen, N. S. (1995) "The Computer Moves into Essay Grading: Updating the Ancient Test" in Phi Delta Kappan. Vol. 76, No. 7, pp. 561–565.
  3. ^ Uebersax, John S. (1987). "Diversity of decision making models and the measurement of interrater agreement" in Psychological Bulletin. Vol 101, pp. 140–146.
  4. ^ Cohen, J. (1960) "A coefficient for agreement for nominal scales" in Education and Psychological Measurement. Vol. 20, pp. 37–46
  5. ^ Fleiss, J. L. (1971) "Measuring nominal scale agreement among many raters" in Psychological Bulletin. Vol. 76, No. 5, pp. 378–382
  6. ^ Gwet, K. L. (2010a) "Correcting Inter-Rater Reliability for Chance Agreement: Why?"
  7. ^ Gwet, K. L. (2010) "Handbook of Inter-Rater Reliability (2nd Edition)"
  8. ^ Shrout, P. and Fleiss, J. L. (1979) "Intraclass correlation: uses in assessing rater reliability" in Psychological Bulletin. Vol. 86, No. 2, pp. 420–428
  9. ^ Everitt, B. (1996) Making Sense of Statistics in Psychology (Oxford : Oxford University Press) ISBN 0-19-852366-1
  10. ^ Bland, J. M., and Altman, D. G. (1986). Statistical methods for assessing agreement between two methods of clinical measurement. Lancet i, pp. 307–310.
  11. ^ Krippendorff, K. (2004). Content analysis: An introduction to its methodology. Thousand Oaks, CA: Sage. pp. 219–250.
  12. ^ Hayes, A. F. & Krippendorff, K. (2007). Answering the call for a standard reliability measure for coding data. Communication Methods and Measures, 1, 77-89.

Further reading

External links


Wikimedia Foundation. 2010.

Игры ⚽ Поможем написать реферат

Look at other dictionaries:

  • Reliability (statistics) — In statistics, reliability is the consistency of a set of measurements or measuring instrument, often used to describe a test. This can either be whether the measurements of the same instrument give or are likely to give the same measurement… …   Wikipedia

  • Ошибки оценщика (rater errors) — О. о. (эксперта, судьи и т. д.) наз. искажения, допускаемые наблюдателями при использовании оценочных шкал для характеристики деятельности индивидуума. Уровень компетентности оценщика, как и его пол, соц. статус, раса, религиозная принадлежность… …   Психологическая энциклопедия

  • Cohen's kappa — coefficient is a statistical measure of inter rater agreement or inter annotator agreement[1] for qualitative (categorical) items. It is generally thought to be a more robust measure than simple percent agreement calculation since κ takes into… …   Wikipedia

  • Rorschach inkblot test — The Rorschach inkblot test (pronounced IPA2|ʁoɐʃax) is a method of psychological evaluation. Psychologists use this test to try to examine the personality characteristics and emotional functioning of their patients. The Rorschach is currently the …   Wikipedia

  • Fleiss' kappa — is a statistical measure for assessing the reliability of agreement between a fixed number of raters when assigning categorical ratings to a number of items or classifying items. This contrasts with other kappas such as Cohen s kappa, which only… …   Wikipedia

  • Concordance correlation coefficient — In statistics, the concordance correlation coefficient measures the agreement between two variables, e.g., to evaluate reproducibility or for inter rater reliability. Definition Lawrence Lin has the form of the concordance correlation coefficient …   Wikipedia

  • Classification of mental disorders — Main article: Mental disorder Psychology …   Wikipedia

  • Diagnostic and Statistical Manual of Mental Disorders — The Diagnostic and Statistical Manual of Mental Disorders (DSM) is published by the American Psychiatric Association and provides a common language and standard criteria for the classification of mental disorders. It is used in the United States… …   Wikipedia

  • List of psychology topics — This page aims to list all topics related to psychology. This is so that those interested in the subject can monitor changes to the pages by clicking on Related changes in the sidebar. It is also to see the gaps in Wikipedia s coverage of the… …   Wikipedia

  • List of statistics topics — Please add any Wikipedia articles related to statistics that are not already on this list.The Related changes link in the margin of this page (below search) leads to a list of the most recent changes to the articles listed below. To see the most… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”