interrater reliability


in·ter·judge re·li·a·bil·i·ty

in psychology, the consistency of measurement obtained when different judges or examiners independently administer the same test to the same subject. Synonym(s): interrater reliability

interrater reliability

The extent to which two independent parties, each using the same tool or examining the same data, arrive at matching conclusions. Many health care investigators analyze graduated data, not binary data. In an analysis of anxiety, for example, a graduated scale may rate research subjects as “very anxious, ” “somewhat anxious, ” “mildly anxious, ” or “not at all anxious, ” whereas a binary method of rating anxiety might include just the two categories “anxious” and “not anxious.” If the study is carried out and coded by more than one psychologist, the coders may not agree on the implementation of the graduated scale: some may interview a patient and find him or her “somewhat” anxious; another might assess the patient as being ”very anxious.” The congruence in the application of the rating scale by more than one psychologist constitutes its interrater reliability. Synonym: interobserver reliabilitySee also: reliability