Exploring an intelligent tutoring system as a conversation-based assessment tool for reading comprehension
Abstract
Reading comprehension is often assessed by having students read passages and administering a test that assesses their understanding of the text. Shorter assessments may fail to give a full picture of comprehension ability while more thorough ones can be time consuming and costly. This study used data from a conversational intelligent tutoring system (AutoTutor) to assess reading comprehension ability in 52 low-literacy adults who interacted with the system. We analyzed participants’ accuracy and time spent answering questions in conversations in lessons that targeted four theoretical components of comprehension: Word, Textbase, Situation Model, and Rhetorical Structure. Accuracy and answer response time were analyzed to track adults’ proficiency for comprehension components, and we analyzed whether the four components predicted reading grade level. We discuss the results with respect to the advantages that a conversational intelligent tutoring system assessment may provide over traditional assessment tools and the linking of theory to practice in adult literacy.
Publication Title
Behaviormetrika
Recommended Citation
Shi, G., Lippert, A., Shubeck, K., Fang, Y., Chen, S., Pavlik, P., Greenberg, D., & Graesser, A. (2018). Exploring an intelligent tutoring system as a conversation-based assessment tool for reading comprehension. Behaviormetrika, 45 (2), 615-633. https://doi.org/10.1007/s41237-018-0065-9