Reference : Validation and Psychometric Analysis of 32 cognitive item models spanning Grades 1 to...
Scientific congresses, symposiums and conference proceedings : Unpublished conference
Social & behavioral sciences, psychology : Theoretical & cognitive psychology
http://hdl.handle.net/10993/52791
Validation and Psychometric Analysis of 32 cognitive item models spanning Grades 1 to 7 in the mathematical domain of numbers & operations
English
Michels, Michael Andreas mailto [University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > LUCET >]
Hornung, Caroline mailto [University of Luxembourg > Faculty of Language and Literature, Humanities, Arts and Education (FLSHASE) > Luxembourg Centre for Educational Testing (LUCET) >]
Gamo, Sylvie mailto [University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > LUCET >]
Roeder, Michel []
Gierl, Mark []
Cardoso-Leite, Pedro mailto [University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > Department of Behavioural and Cognitive Sciences (DBCS) >]
Fischbach, Antoine mailto [University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > Department of Education and Social Work (DESW) >]
Sonnleitner, Philipp mailto [University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > LUCET >]
Nov-2022
Yes
International
Luxembourg Educational Research Association Conference 2022
from 09-11-2022 to 10-11-2022
Esch-sur-Alzette
Luxembourg
[en] item development ; Automatic Item Generation ; cognitive models ; mathematics ; numbers & operation
[en] Today’s educational field has a tremendous hunger for valid and psychometrically sound items to reliably track and model students’ learning processes. Educational large-scale assessments, formative classroom assessment, and lately, digital learning platforms require a constant stream of high-quality, and unbiased items. However, traditional development of test items ties up a significant amount of time from subject matter experts, pedagogues and psychometricians and might not be suited anymore to nowadays demands. Salvation is sought in automatic item generation (AIG) which provides the possibility of generating multiple items within a short period of time based on the development of cognitively sound item templates by using algorithms (Gierl & Haladyna, 2013; Gierl et al., 2015).
The present study psychometrically analyses 35 cognitive item models that were developed by a team of national subject matter experts and psychometricians and then used for algorithmically producing items for the mathematical domain of numbers & shapes for Grades 1, 3, 5, and 7 of the Luxembourgish school system. Each item model was administered in 6 experimentally varied versions to investigate the impact of a) the context the mathematical problem was presented in, and b) problem characteristics which cognitive psychology identified to influence the problem solving process. Based on samples from Grade 1 (n = 5963), Grade 3 (n = 5527), Grade 5 (n = 5291), and Grade 7 (n = 3018) collected within the annual Épreuves standardisées, this design allows for evaluating whether psychometric characteristics of produced items per model are a) stable, b) can be predicted by problem characteristics, and c) are unbiased towards subgroups of students (known to be disadvantaged in the Luxembourgish school system).
After item calibration using the 1-PL model, each cognitive model was analyzed in-depth by descriptive comparisons of resulting IRT parameters, and the estimation of manipulated problem characteristics’ impact on item difficulty by using the linear logistic test model (LLTM, Fischer, 1972). Results are truly promising and show negligible effects of different problem contexts on item difficulty and reasonably stable effects of altered problem characteristics. Thus, the majority of developed cognitive models could be used to generate a huge number of items (> 10.000.000) for the domain of numbers & operations with known psychometric properties without the need for expensive field-trials. We end with discussing lessons learned from item difficulty prediction per model and highlighting differences between the Grades.

References:
Fischer, G. H. (1973). The linear logistic test model as an instrument in educational research. Acta Psychologica, 36, 359-374.
Gierl, M. J., & Haladyna, T. M. (Eds.). (2013). Automatic item generation: Theory and practice. New York, NY: Routledge.
Gierl, M. J., Lai, H., Hogan, J., & Matovinovic, D. (2015). A Method for Generating Educational Test Items That Are Aligned to the Common Core State Standards. Journal of Applied Testing Technology, 16(1), 1–18.
Researchers ; Professionals ; Students ; General public
http://hdl.handle.net/10993/52791
FnR ; FNR13650128 > Philipp Sonnleitner > FAIR-ITEMS > Fairness Of Latest Innovations In Item And Test Development In Mathematics > 01/09/2020 > 31/08/2023 > 2019

There is no file associated with this reference.

Bookmark and Share SFX Query

All documents in ORBilu are protected by a user license.