Interrater Agreement Bedeutung

Since it is used as a measure of compliance, only positive values are expected in most situations; Negative values would indicate systematic differences of opinion. Kappa can only reach very high values if both convergences are good and the rate of the target condition is close to 50% (because it includes the base rate in the calculation of common probabilities). Several authorities have proposed “ground rules” for interpreting the degree of agreement, many of which fundamentally agree, although the terms are not identical. [8] [9] [10] [11] In statistics, the reliability of the inter-advisor (also referred to by different similar names such as the Inter-Rater agreement, inter-rater concordance, inter-observer reliability, etc.) is the degree of consistency between evaluators. It is an assessment of homogeneity or consensus in the assessments of different judges. If the number of categories used is small (for example.B. 2 or 3), the probability that 2 evaluators agree by chance increases dramatically. This is because both evaluators must limit themselves to the limited number of options available, which affects the overall rate of the agreement, and not necessarily their propensity to enter into an “intrinsic” agreement (an agreement is considered “intrinsic” if it is not due to chance). The common probability of an agreement is the simplest and least robust measure. It is estimated as a percentage of the time during which evaluators agree in a nominal or categorical evaluation system. It does not take into account the fact that an agreement can be concluded solely on the basis of chance.


Please direct requests and inquiries to Rhiannon Richards - Sr. PR & Communications Manager, 22837 Ventura Blvd, 3rd Floor, Woodland Hills, CA 91364 or call 818-225-5100 ext 249 or

sex gifs