Chapter Measures of interrater agreement when each target is evaluated by a different group of raters

Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which...

Full description

Saved in:
Bibliographic Details
Main Author: BOVE, Giuseppe (auth)
Format: Electronic Book Chapter
Language:English
Published: Florence Firenze University Press, Genova University Press 2023
Series:Proceedings e report
Subjects:
Online Access:DOAB: download the publication
DOAB: description of the publication
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000naaaa2200000uu 4500
001 doab_20_500_12854_112079
005 20230808
003 oapen
006 m o d
007 cr|mn|---annan
008 20230808s2023 xx |||||o ||| 0|eng d
020 |a 979-12-215-0106-3.28 
020 |a 9791221501063 
040 |a oapen  |c oapen 
024 7 |a 10.36253/979-12-215-0106-3.28  |c doi 
041 0 |a eng 
042 |a dc 
072 7 |a J  |2 bicssc 
100 1 |a BOVE, Giuseppe  |4 auth 
245 1 0 |a Chapter Measures of interrater agreement when each target is evaluated by a different group of raters 
260 |a Florence  |b Firenze University Press, Genova University Press  |c 2023 
300 |a 1 electronic resource (6 p.) 
336 |a text  |b txt  |2 rdacontent 
337 |a computer  |b c  |2 rdamedia 
338 |a online resource  |b cr  |2 rdacarrier 
490 1 |a Proceedings e report 
506 0 |a Open Access  |2 star  |f Unrestricted online access 
520 |a Most measures of interrater agreement are defined for ratings regarding a group of targets, each rated by the same group of raters (e.g., the agreement of raters who assess on a rating scale the language proficiency of a corpus of argumentative written texts). However, there are situations in which agreement between ratings regards a group of targets where each target is evaluated by a different group of raters, like for instance when teachers in a school are evaluated by a questionnaire administered to all the pupils (students) in the classroom. In these situations, a first approach is to evaluate the level of agreement for the whole group of targets by the ANOVA one-way random model. A second approach is to apply subject-specific indices of interrater agreement like rWG, which represents the observed variance in ratings compared to the variance of a theoretical distribution representing no agreement (i.e., the null distribution). Both these approaches are not appropriate for ordinal or nominal scales. In this paper, an index is proposed to evaluate the agreement between raters for each single target (subject or object) on an ordinal scale, and to obtain also a global measure of the interrater agreement for the whole group of cases evaluated. The index is not affected by the possible concentration of ratings on a very small number of levels of the scale, like it happens for the measures based on the ANOVA approach, and it does not depend on the definition of a null distributions like rWG. The main features of the proposal will be illustrated in a study for the assessment of learning teacher behavior in classroom collected in a research conducted in 2018 at Roma Tre University. 
540 |a Creative Commons  |f https://creativecommons.org/licenses/by/4.0/  |2 cc  |4 https://creativecommons.org/licenses/by/4.0/ 
546 |a English 
650 7 |a Society & social sciences  |2 bicssc 
653 |a Interrater agreement 
653 |a Ordinal data 
653 |a Teacher evaluation 
773 1 0 |7 nnaa  |o OAPEN Library UUID: ASA 2022 Data-Driven Decision Making 
856 4 0 |a www.oapen.org  |u https://library.oapen.org/bitstream/20.500.12657/74900/1/9791221501063-28.pdf  |7 0  |z DOAB: download the publication 
856 4 0 |a www.oapen.org  |u https://directory.doabooks.org/handle/20.500.12854/112079  |7 0  |z DOAB: description of the publication