Automated scoring of teachers’ pedagogical content knowledge : a comparison between human and machine scoring
Date issued
Editors
Journal Title
Journal ISSN
Volume Title
Publisher
License
Abstract
To validly assess teachers’ pedagogical content knowledge (PCK), performance-based tasks with open-response formats are required. Automated scoring is considered an appropriate approach to reduce the resource-intensity of human scoring and to achieve more consistent scoring results than human raters. The focus is on the comparability of human and automated scoring of PCK for economics teachers. The answers of (prospective) teachers (N = 852) to six open-response tasks from a standardized and validated test were scored by two trained human raters and the engine “Educational SCoRIng Toolkit” (ESCRITO). The average agreement between human and computer ratings, κw = 0.66, suggests a convergent validity of the scoring results. The results of the single-sector variance analysis show a significant influence of the answers for each homogeneous subgroup (students = 460, trainees = 230, in-service teachers = 162) on the automated scoring. Findings are discussed in terms of implications for the use of automated scoring in educational assessment and its potentials and limitations.
Description
Keywords
Citation
Published in
Frontiers in education, 5, Frontiers Media, Lausanne, 2020, https://doi.org/10.3389/feduc.2020.00149