Chapter Measures of interrater agreement when each target is evaluated by a different group of raters

Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which...

Full description

Saved in:
Bibliographic Details
Main Author: BOVE, Giuseppe (auth)
Format: Electronic Book Chapter
Language:English
Published: Florence Firenze University Press, Genova University Press 2023
Series:Proceedings e report 134
Subjects:
Online Access:OAPEN Library: download the publication
OAPEN Library: description of the publication
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000naaaa2200000uu 4500
001 oapen_2024_20_500_12657_74900
005 20230803
003 oapen
006 m o d
007 cr|mn|---annan
008 20230803s2023 xx |||||o ||| 0|eng d
020 |a 979-12-215-0106-3.28 
020 |a 9791221501063 
040 |a oapen  |c oapen 
024 7 |a 10.36253/979-12-215-0106-3.28  |c doi 
041 0 |a eng 
042 |a dc 
072 7 |a J  |2 bicssc 
100 1 |a BOVE, Giuseppe  |4 auth 
245 1 0 |a Chapter Measures of interrater agreement when each target is evaluated by a different group of raters 
260 |a Florence  |b Firenze University Press, Genova University Press  |c 2023 
300 |a 1 electronic resource (6 p.) 
336 |a text  |b txt  |2 rdacontent 
337 |a computer  |b c  |2 rdamedia 
338 |a online resource  |b cr  |2 rdacarrier 
490 1 |a Proceedings e report  |v 134 
506 0 |a Open Access  |2 star  |f Unrestricted online access 
520 |a Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which agreement between ratings regards a group of targets where each target is evaluated by a different group of raters, like for instance when teachers in a school are evaluated by a questionnaire administered to all the pupils (students) in the classroom. In these situations, a first approach is to evaluate the level of agreement for the whole group of targets by the ANOVA one-way random model. A second approach is to apply subject-specific indices of interrater agreement like rWG, which represents the observed variance in ratings compared to the variance of a theoretical distribution representing no agreement (i.e., the null distribution). Both these approaches are not appropriate for ordinal or nominal scales. In this paper, an index is proposed to evaluate the agreement between raters for each single target (subject or object) on an ordinal scale, and to obtain also a global measure of the interrater agreement for the whole group of cases evaluated. The index is not affected by the possible concentration of ratings on a very small number of levels of the scale, like it happens for the measures based on the ANOVA approach, and it does not depend on the definition of a null distributions like rWG. The main features of the proposal will be illustrated in a study for the assessment of learning teacher behavior in classroom collected in a research conducted in 2018 at Roma Tre University. 
540 |a Creative Commons  |f https://creativecommons.org/licenses/by/4.0/  |2 cc  |4 https://creativecommons.org/licenses/by/4.0/ 
546 |a English 
650 7 |a Society & social sciences  |2 bicssc 
653 |a Interrater agreement 
653 |a Ordinal data 
653 |a Teacher evaluation 
773 1 0 |t ASA 2022 Data-Driven Decision Making  |7 nnaa  |o OAPEN Library UUID: 863aa499-dbee-4191-9a14-3b5d5ef9e635 
856 4 0 |a www.oapen.org  |u https://library.oapen.org/bitstream/id/35a68d36-1e53-4aea-92da-d89970a93ce6/9791221501063-28.pdf  |7 0  |z OAPEN Library: download the publication 
856 4 0 |a www.oapen.org  |u https://library.oapen.org/handle/20.500.12657/74900  |7 0  |z OAPEN Library: description of the publication