Article (Scientific journals)
Evaluating AI-generated vs. human-written reading comprehension passages: an expert SWOT analysis and comparative study for an educational large-scale assessment
Ripoll Y Schmitz, Lisa Marie; SONNLEITNER, Philipp
2025In Large-Scale Assessments in Education, 13 (1)
Peer Reviewed verified by ORBi
 

Files


Full Text
Ripoll Y Schmitz & Sonnleitner_2025_Evaluating AI-generated text passages.pdf
Publisher postprint (1.43 MB)
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
ChatGPT; Educational large-scale assessment; Generative artificial intelligence; Large language models; Prompt engineering; Reading comprehension; Text analysis cognitive model; Education
Abstract :
[en] Background: The increasing capabilities of generative artificial intelligence (AI), exemplified by OpenAI’s transformer-based language model GPT-4 (ChatGPT), have drawn attention to its application in educational contexts. This study evaluates the potential of such models in generating German reading comprehension texts for educational large-scale assessments, within the multilingual context of Luxembourg. Addressing the challenges faced by item developers in sourcing or manually developing numerous suitable texts, the study aims to determine if ChatGPT can assist text creation while maintaining high-quality standards. Methods: The study employed a mixed-methods approach. In a qualitative focus group discussion, experts identified the strengths, weaknesses, opportunities and threats (SWOT) of using GPT-4 for text generation. These insights informed the construction of a Text Analysis Cognitive Model (TACM), which served as theoretical foundation. Narrative and informative reading comprehension texts were then generated using two distinct prompt engineering techniques, derived from original passages and TACM specifications. In a blinded online review, N = 89 participants evaluated human-written and AI-generated texts with regard to their readability, correctness, coherence, engagement and adequacy for reading assessment. Results: All administered texts were of similarly high quality, with reviewers being unable to consistently identify authorship origins. Quantitative evaluations indicated that one-shot prompts are effective for creating high-quality informative texts, whereas human-written texts remain superior for narratives. Zero-shot prompts offer considerable flexibility and creativity, but still require human refinement. Conclusion: These findings offer promising first insights into GPT-4’s capacity to emulate human-written texts which can be used in the large-scale assessment context. The considerable potential of using generative AI-models as a flexible and efficacious assistant in the creation of reading comprehension texts is highlighted. Still, the necessity of human oversight is emphasized through an augmented intelligence-driven perspective. Given the jurisdictional framework of the European Union, an effective implementation of ChatGPT in the test development process remains hypothetical at this time but is likely to change.
Disciplines :
Education & instruction
Author, co-author :
Ripoll Y Schmitz, Lisa Marie;  University of Luxembourg, Esch-Sur-Alzette, Luxembourg
SONNLEITNER, Philipp  ;  University of Luxembourg > Faculty of Humanities, Education and Social Sciences (FHSE) > LUCET
External co-authors :
no
Language :
English
Title :
Evaluating AI-generated vs. human-written reading comprehension passages: an expert SWOT analysis and comparative study for an educational large-scale assessment
Publication date :
July 2025
Journal title :
Large-Scale Assessments in Education
eISSN :
2196-0739
Publisher :
Springer
Volume :
13
Issue :
1
Peer reviewed :
Peer Reviewed verified by ORBi
Available on ORBilu :
since 29 July 2025

Statistics


Number of views
70 (12 by Unilu)
Number of downloads
40 (1 by Unilu)

Scopus citations®
 
0
Scopus citations®
without self-citations
0
OpenCitations
 
0
OpenAlex citations
 
0
WoS citations
 
0

Bibliography


Similar publications



Contact ORBilu