Publication Date
| In 2026 | 0 |
| Since 2025 | 215 |
| Since 2022 (last 5 years) | 1084 |
| Since 2017 (last 10 years) | 2594 |
| Since 2007 (last 20 years) | 4955 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Marta Siedlecka; Piotr Litwin; Paulina Szyszka; Boryslaw Paulewicz – European Journal of Psychology of Education, 2025
Students change their responses during tests, and these revisions are often correct. Some studies have suggested that decisions regarding revisions are informed by metacognitive monitoring. We investigated whether assessing and reporting response confidence increases the accuracy of revisions and the final test score, and whether confidence in a…
Descriptors: Student Evaluation, Decision Making, Responses, Achievement Tests
Ikkyu Choi; Jiyun Zu – Language Testing, 2025
Today's language models can produce syntactically accurate and semantically coherent texts. This capability presents new opportunities for generating content for language assessments, which have traditionally required intensive expert resources. However, these models are also known to generate biased texts, leading to representational harms.…
Descriptors: Artificial Intelligence, Language Tests, Test Bias, Test Construction
Linh Thi Thao Le; Nam Thi Phuong Ho; Nguyen Huynh Trang; Hung Tan Ha – SAGE Open, 2025
The International English Language Testing System (IELTS) has served as one of the most reliable proofs of people's English language proficiency. There have been rumors about the discrepancy in difficulty between the two modules of IELTS, namely Academic (AC) and General Training (GT); however, there is little empirical evidence to confirm such a…
Descriptors: English (Second Language), Language Tests, Second Language Learning, Reading Tests
Abdulla Alzarouni; R. J. De Ayala – Practical Assessment, Research & Evaluation, 2025
The assessment of model fit in latent trait modeling is an integral part of correctly applying the model. Still the assessment of model fit has been less utilized for ideal point models such as the Generalized Graded Unfolding Models (GGUM). The current study assesses the performance of the relative fit indices "AIC" and "BIC,"…
Descriptors: Goodness of Fit, Models, Statistical Analysis, Sample Size
Jianbin Fu; Xuan Tan; Patrick C. Kyllonen – Journal of Educational Measurement, 2024
This paper presents the item and test information functions of the Rank two-parameter logistic models (Rank-2PLM) for items with two (pair) and three (triplet) statements in forced-choice questionnaires. The Rank-2PLM model for pairs is the MUPP-2PLM (Multi-Unidimensional Pairwise Preference) and, for triplets, is the Triplet-2PLM. Fisher's…
Descriptors: Questionnaires, Test Items, Item Response Theory, Models
Anna Planas-Lladó; Xavier Úcar – American Journal of Evaluation, 2024
Empowerment is a concept that has become increasingly used over recent years. However, little research has been undertaken into how empowerment can be evaluated, particularly in the case of young people. The aim of this article is to present an inventory of dimensions and indicators of youth empowerment. The article describes the various phases in…
Descriptors: Youth, Empowerment, Test Construction, Test Validity
The Impact of Measurement Noninvariance across Time and Group in Longitudinal Item Response Modeling
In-Hee Choi – Asia Pacific Education Review, 2024
Longitudinal item response data often exhibit two types of measurement noninvariance: the noninvariance of item parameters between subject groups and that of item parameters across multiple time points. This study proposes a comprehensive approach to the simultaneous modeling of both types of measurement noninvariance in terms of longitudinal item…
Descriptors: Longitudinal Studies, Item Response Theory, Growth Models, Error of Measurement
Janet Mee; Ravi Pandian; Justin Wolczynski; Amy Morales; Miguel Paniagua; Polina Harik; Peter Baldwin; Brian E. Clauser – Advances in Health Sciences Education, 2024
Recent advances in automated scoring technology have made it practical to replace multiple-choice questions (MCQs) with short-answer questions (SAQs) in large-scale, high-stakes assessments. However, most previous research comparing these formats has used small examinee samples testing under low-stakes conditions. Additionally, previous studies…
Descriptors: Multiple Choice Tests, High Stakes Tests, Test Format, Test Items
Bilal Ghanem; Alona Fyshe – International Educational Data Mining Society, 2024
Multiple choice questions (MCQs) are a common way to assess reading comprehension. Every MCQ needs a set of distractor answers that are incorrect, but plausible enough to test student knowledge. However, good distractors are hard to create. Distractor generation (DG) models have been proposed, and their performance is typically evaluated using…
Descriptors: Multiple Choice Tests, Reading Comprehension, Test Items, Testing
Daniel Lewis; Melanie Graw; Michael Baker – Journal of Applied Testing Technology, 2024
Embedded Standard Setting (ESS; Lewis & Cook, 2020) transforms standard setting from a standalone workshop to an active part of the assessment development lifecycle. ESS purports to lower costs by eliminating the standard-setting workshop and enhance the validity argument by maintaining a consistent focus on the evidentiary relationship…
Descriptors: Standard Setting (Scoring), Test Items, Test Construction, Food Service
Kasli, Murat; Zopluoglu, Cengiz; Toton, Sarah L. – Journal of Educational Measurement, 2023
Response times (RTs) have recently attracted a significant amount of attention in the literature as they may provide meaningful information about item preknowledge. In this study, a new model, the Deterministic Gated Lognormal Response Time (DG-LNRT) model, is proposed to identify examinees with item preknowledge using RTs. The proposed model was…
Descriptors: Reaction Time, Test Items, Models, Familiarity
Neuert, Cornelia E.; Meitinger, Katharina; Behr, Dorothée – Sociological Methods & Research, 2023
The method of web probing integrates cognitive interviewing techniques into web surveys and is increasingly used to evaluate survey questions. In a usual web probing scenario, probes are administered immediately after the question to be tested (concurrent probing), typically as open-ended questions. A second possibility of administering probes is…
Descriptors: Internet, Online Surveys, Test Items, Evaluation
He, Yinhong; Qi, Yuanyuan – Journal of Educational Measurement, 2023
In multidimensional computerized adaptive testing (MCAT), item selection strategies are generally constructed based on responses, and they do not consider the response times required by items. This study constructed two new criteria (referred to as DT-inc and DT) for MCAT item selection by utilizing information from response times. The new designs…
Descriptors: Reaction Time, Adaptive Testing, Computer Assisted Testing, Test Items
Ben Babcock; Kim Brunnert – Journal of Applied Testing Technology, 2023
Automatic Item Generation (AIG) is an extremely useful tool to construct many high-quality exam items more efficiently than traditional item writing methods. A large pool of items, however, presents challenges like identifying a particular item to meet a specific need. For example, when making a fixed form exam, best practices forbid item stems…
Descriptors: Test Items, Automation, Algorithms, Artificial Intelligence
Aiman Mohammad Freihat; Omar Saleh Bani Yassin – Educational Process: International Journal, 2025
Background/purpose: This study aimed to reveal the accuracy of estimation of multiple-choice test items parameters following the models of the item-response theory in measurement. Materials/methods: The researchers depended on the measurement accuracy indicators, which express the absolute difference between the estimated and actual values of the…
Descriptors: Accuracy, Computation, Multiple Choice Tests, Test Items

Peer reviewed
Direct link
