PIRLS Category-specific Question Generation for Reading Comprehension
dc.contributor.author | Poon, Yin | |
dc.contributor.author | Wang, Qiong | |
dc.contributor.author | Lee, John S. Y. | |
dc.contributor.author | Lam, Yu Yan | |
dc.contributor.author | Kai Wah Chu, Samuel | |
dc.contributor.editor | Muñoz Sánchez, Ricardo | |
dc.contributor.editor | Alfter, David | |
dc.contributor.editor | Volodina, Elena | |
dc.contributor.editor | Kallas, Jelena | |
dc.coverage.spatial | Tallinn, Estonia | |
dc.date.accessioned | 2025-02-17T10:44:15Z | |
dc.date.available | 2025-02-17T10:44:15Z | |
dc.date.issued | 2025-03 | |
dc.description.abstract | According to the internationally recognized PIRLS (Progress in International Reading Literacy Study) assessment standards, reading comprehension questions should encompass all four comprehension processes: retrieval, inferencing, integrating and evaluation. This paper investigates whether Large Language Models can produce high-quality questions for each of these categories. Human assessment on a Chinese dataset shows that GPT-4o can generate usable and category-specific questions, ranging from 74% to 90% accuracy depending on the category. | |
dc.identifier.uri | https://hdl.handle.net/10062/107171 | |
dc.language.iso | en | |
dc.publisher | University of Tartu Library | |
dc.rights | Attribution-NonCommercial-NoDerivatives 4.0 International | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.title | PIRLS Category-specific Question Generation for Reading Comprehension | |
dc.type | Article |
Failid
Originaal pakett
1 - 1 1