Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 7 |
Descriptor
Computer Assisted Testing | 7 |
Correlation | 7 |
Foreign Countries | 4 |
Test Items | 3 |
Achievement Tests | 2 |
Automation | 2 |
High Stakes Tests | 2 |
Measurement | 2 |
Scoring | 2 |
Secondary School Students | 2 |
Statistical Analysis | 2 |
More ▼ |
Source
Applied Measurement in… | 7 |
Author
Ainley, John | 1 |
Ben-Simon, Anat | 1 |
Boulais, André-Philippe | 1 |
Breland, Hunter | 1 |
Cohen, Yoav | 1 |
De Champlain, André | 1 |
Fraillon, Julian | 1 |
Gebhardt, Eveline | 1 |
Gierl, Mark J. | 1 |
Greiff, Samuel | 1 |
Herde, Christoph Nils | 1 |
More ▼ |
Publication Type
Journal Articles | 7 |
Reports - Research | 5 |
Reports - Evaluative | 2 |
Information Analyses | 1 |
Education Level
Secondary Education | 3 |
Higher Education | 2 |
Postsecondary Education | 2 |
Elementary Education | 1 |
Grade 8 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Cohen, Yoav; Levi, Effi; Ben-Simon, Anat – Applied Measurement in Education, 2018
In the current study, two pools of 250 essays, all written as a response to the same prompt, were rated by two groups of raters (14 or 15 raters per group), thereby providing an approximation to the essay's true score. An automated essay scoring (AES) system was trained on the datasets and then scored the essays using a cross-validation scheme. By…
Descriptors: Test Validity, Automation, Scoring, Computer Assisted Testing
Lottridge, Susan; Wood, Scott; Shaw, Dan – Applied Measurement in Education, 2018
This study sought to provide a framework for evaluating machine score-ability of items using a new score-ability rating scale, and to determine the extent to which ratings were predictive of observed automated scoring performance. The study listed and described a set of factors that are thought to influence machine score-ability; these factors…
Descriptors: Program Effectiveness, Computer Assisted Testing, Test Scoring Machines, Scoring
Gierl, Mark J.; Lai, Hollis; Pugh, Debra; Touchie, Claire; Boulais, André-Philippe; De Champlain, André – Applied Measurement in Education, 2016
Item development is a time- and resource-intensive process. Automatic item generation integrates cognitive modeling with computer technology to systematically generate test items. To date, however, items generated using cognitive modeling procedures have received limited use in operational testing situations. As a result, the psychometric…
Descriptors: Psychometrics, Multiple Choice Tests, Test Items, Item Analysis
Herde, Christoph Nils; Wüstenberg, Sascha; Greiff, Samuel – Applied Measurement in Education, 2016
Complex Problem Solving (CPS) is seen as a cross-curricular 21st century skill that has attracted interest in large-scale-assessments. In the Programme for International Student Assessment (PISA) 2012, CPS was assessed all over the world to gain information on students' skills to acquire and apply knowledge while dealing with nontransparent…
Descriptors: Problem Solving, Achievement Tests, Foreign Countries, International Assessment
Ainley, John; Fraillon, Julian; Schulz, Wolfram; Gebhardt, Eveline – Applied Measurement in Education, 2016
The development of information technologies has transformed the environment in which young people access, create, and share information. Many countries, having recognized the imperative of digital technology, acknowledge the need to educate young people in the use of these technologies so as to underpin economic and social benefits. This article…
Descriptors: Cross Cultural Studies, Information Literacy, Computer Literacy, Grade 8
Wise, Steven L.; Pastor, Dena A.; Kong, Xiaojing J. – Applied Measurement in Education, 2009
Previous research has shown that rapid-guessing behavior can degrade the validity of test scores from low-stakes proficiency tests. This study examined, using hierarchical generalized linear modeling, examinee and item characteristics for predicting rapid-guessing behavior. Several item characteristics were found significant; items with more text…
Descriptors: Guessing (Tests), Achievement Tests, Correlation, Test Items
Breland, Hunter; Lee, Yong-Won – Applied Measurement in Education, 2007
The objective of the present investigation was to examine the comparability of writing prompts for different gender groups in the context of the computer-based Test of English as a Foreign Language[TM] (TOEFL[R]-CBT). A total of 87 prompts administered from July 1998 through March 2000 were analyzed. An extended version of logistic regression for…
Descriptors: Learning Theories, Writing Evaluation, Writing Tests, Second Language Learning