A method to predict interrater agreement in writing assessment based on raters' individual differences

Roberto Jaunez, Jorge Villalon, Gonzalo Munoz, Gabriela Baez

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Writing has been increasingly assessed within educational scenarios. However, assessing writing is a highly subjective task, making the grading of such tests a costly process that demands several graders for calculating interrater agreement to ensure reliability. A drawback of this approach is that interrater agreement does not explain the source of markers' differences; therefore, when agreement is low, the marking process must be repeated, and markers retrained. This article proposes a method to predict agreement, based on markers' individual differences using item response theory. Results showed that the method accurately predicts agreement even with partial evidence.

Original languageEnglish
Title of host publicationProceedings - IEEE 19th International Conference on Advanced Learning Technologies, ICALT 2019
EditorsMaiga Chang, Demetrios G Sampson, Ronghuai Huang, Alex Sandro Gomes, Nian-Shing Chen, Ig Ibert Bittencourt, Kinshuk Kinshuk, Diego Dermeval, Ibsen Mateus Bittencourt
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages255-257
Number of pages3
ISBN (Electronic)9781728134857
DOIs
StatePublished - Jul 2019
Event19th IEEE International Conference on Advanced Learning Technologies, ICALT 2019 - Maceio, Brazil
Duration: 15 Jul 201918 Jul 2019

Publication series

NameProceedings - IEEE 19th International Conference on Advanced Learning Technologies, ICALT 2019

Conference

Conference19th IEEE International Conference on Advanced Learning Technologies, ICALT 2019
Country/TerritoryBrazil
CityMaceio
Period15/07/1918/07/19

Keywords

  • Agreement prediction
  • Interrater agreement
  • Item response theory
  • Writing assessment

Fingerprint

Dive into the research topics of 'A method to predict interrater agreement in writing assessment based on raters' individual differences'. Together they form a unique fingerprint.

Cite this