NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 7 results Save | Export
Peer reviewed Peer reviewed
Andreea Dutulescu; Stefan Ruseti; Mihai Dascalu; Danielle S. McNamara – Grantee Submission, 2024
Assessing the difficulty of reading comprehension questions is crucial to educational methodologies and language understanding technologies. Traditional methods of assessing question difficulty rely frequently on human judgments or shallow metrics, often failing to accurately capture the intricate cognitive demands of answering a question. This…
Descriptors: Difficulty Level, Reading Tests, Test Items, Reading Comprehension
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Malik, Ali; Wu, Mike; Vasavada, Vrinda; Song, Jinpeng; Coots, Madison; Mitchell, John; Goodman, Noah; Piech, Chris – International Educational Data Mining Society, 2021
Access to high-quality education at scale is limited by the difficulty of providing student feedback on open-ended assignments in structured domains like programming, graphics, and short response questions. This problem has proven to be exceptionally difficult: for humans, it requires large amounts of manual work, and for computers, until…
Descriptors: Grading, Accuracy, Computer Assisted Testing, Automation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lu, Chang; Cutumisu, Maria – International Educational Data Mining Society, 2021
Digitalization and automation of test administration, score reporting, and feedback provision have the potential to benefit large-scale and formative assessments. Many studies on automated essay scoring (AES) and feedback generation systems were published in the last decade, but few connected AES and feedback generation within a unified framework.…
Descriptors: Learning Processes, Automation, Computer Assisted Testing, Scoring
Price, Beth; Steinle, Vicki; Stacey, Kaye; Gvozdenko, Eugene – Mathematics Education Research Group of Australasia, 2014
This study reports on the use of formative, diagnostic online assessments for the topic percentages. Two new item formats (drag-drop and slider) are described. About one-third of the school students (Years 7 to 9) could, using a slider, estimate "80% more than" a given length, in contrast with over two-thirds who could estimate "90%…
Descriptors: Computation, Mathematical Concepts, Formative Evaluation, Diagnostic Tests
Wang, Shudong; Jiao, Hong – Online Submission, 2011
For decades, researchers and practitioners have made a great deal of effort to study a variety of methods to increase parameter accuracy, but only recently can researchers start focusing on improving parameter estimations by using a joint model that could incorporate RT and students information as CI. Given that many tests are currently…
Descriptors: Reaction Time, Item Response Theory, Computer Assisted Testing, Computation
Wang, Shudong; Jiao, Hong; He, Wei – Online Submission, 2011
The ability estimation procedure is one of the most important components in a computerized adaptive testing (CAT) system. Currently, all CATs that provide K-12 student scores are based on the item response theory (IRT) model(s); while such application directly violates the assumption of independent sample of a person in IRT models because ability…
Descriptors: Accuracy, Computation, Computer Assisted Testing, Adaptive Testing
He, Wei; Reckase, Mark – Online Submission, 2008
Test security has been a concern for computerized adaptive tests (CAT) due to the nature of continuous testing. This concern becomes unprecedentedly severe with increasingly easy access to the World-Wide-Web where some examinees post on the internet their recollections of items they are administered, leaving future examinees with opportunities to…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Items, Item Banks