Publication Date
In 2025 | 5 |
Since 2024 | 14 |
Since 2021 (last 5 years) | 60 |
Since 2016 (last 10 years) | 145 |
Since 2006 (last 20 years) | 224 |
Descriptor
Difficulty Level | 271 |
Multiple Choice Tests | 271 |
Test Items | 188 |
Foreign Countries | 115 |
Test Construction | 74 |
Item Analysis | 66 |
Item Response Theory | 60 |
Test Reliability | 57 |
Test Format | 52 |
Test Validity | 50 |
Statistical Analysis | 41 |
More ▼ |
Source
Author
Andrich, David | 3 |
Atalmis, Erkan Hasan | 3 |
Cizek, Gregory J. | 3 |
Fischer, Martin R. | 3 |
Marais, Ida | 3 |
Albanese, Mark A. | 2 |
Bauer, Daniel | 2 |
Bolt, Daniel M. | 2 |
Bucak, S. Deniz | 2 |
Bulut, Okan | 2 |
Crisp, Victoria | 2 |
More ▼ |
Publication Type
Journal Articles | 271 |
Reports - Research | 236 |
Reports - Evaluative | 25 |
Tests/Questionnaires | 19 |
Reports - Descriptive | 9 |
Information Analyses | 6 |
Speeches/Meeting Papers | 2 |
Guides - Non-Classroom | 1 |
Education Level
Higher Education | 97 |
Postsecondary Education | 77 |
Secondary Education | 53 |
Elementary Education | 35 |
Middle Schools | 22 |
High Schools | 19 |
Intermediate Grades | 14 |
Junior High Schools | 14 |
Grade 6 | 10 |
Grade 7 | 10 |
Grade 8 | 9 |
More ▼ |
Audience
Teachers | 2 |
Administrators | 1 |
Practitioners | 1 |
Researchers | 1 |
Location
Turkey | 14 |
Indonesia | 8 |
Australia | 7 |
Canada | 7 |
Germany | 7 |
Nigeria | 7 |
Jordan | 5 |
Taiwan | 5 |
Malaysia | 4 |
Thailand | 4 |
United Kingdom | 4 |
More ▼ |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
What Works Clearinghouse Rating
Aiman Mohammad Freihat; Omar Saleh Bani Yassin – Educational Process: International Journal, 2025
Background/purpose: This study aimed to reveal the accuracy of estimation of multiple-choice test items parameters following the models of the item-response theory in measurement. Materials/methods: The researchers depended on the measurement accuracy indicators, which express the absolute difference between the estimated and actual values of the…
Descriptors: Accuracy, Computation, Multiple Choice Tests, Test Items
Berenbon, Rebecca F.; McHugh, Bridget C. – Educational Measurement: Issues and Practice, 2023
To assemble a high-quality test, psychometricians rely on subject matter experts (SMEs) to write high-quality items. However, SMEs are not typically given the opportunity to provide input on which content standards are most suitable for multiple-choice questions (MCQs). In the present study, we explored the relationship between perceived MCQ…
Descriptors: Test Items, Multiple Choice Tests, Standards, Difficulty Level
Chen, Yun-Zu; Yang, Kai-Lin – Applied Cognitive Psychology, 2023
This study investigated whether the three variables of task form, squares carried, and figural complexity, for designing cube folding tasks, affect sixth graders' cube folding performance. Two task forms were used to develop two versions of "cube folding test." Each version was designed based on two levels of squares carried and three…
Descriptors: Elementary School Students, Grade 6, Geometric Concepts, Task Analysis
Lang, Joseph B. – Journal of Educational and Behavioral Statistics, 2023
This article is concerned with the statistical detection of copying on multiple-choice exams. As an alternative to existing permutation- and model-based copy-detection approaches, a simple randomization p-value (RP) test is proposed. The RP test, which is based on an intuitive match-score statistic, makes no assumptions about the distribution of…
Descriptors: Identification, Cheating, Multiple Choice Tests, Item Response Theory
Ludewig, Ulrich; Schwerter, Jakob; McElvany, Nele – Journal of Psychoeducational Assessment, 2023
A better understanding of how distractor features influence the plausibility of distractors is essential for an efficient multiple-choice (MC) item construction in educational assessment. The plausibility of distractors has a major influence on the psychometric characteristics of MC items. Our analysis utilizes the nominal categories model to…
Descriptors: Vocabulary, Language Tests, German, Grade 4
Emily K. Toutkoushian; Huaping Sun; Mark T. Keegan; Ann E. Harman – Measurement: Interdisciplinary Research and Perspectives, 2024
Linear logistic test models (LLTMs), leveraging item response theory and linear regression, offer an elegant method for learning about item characteristics in complex content areas. This study used LLTMs to model single-best-answer, multiple-choice-question response data from two medical subspecialty certification examinations in multiple years…
Descriptors: Licensing Examinations (Professions), Certification, Medical Students, Test Items
Lae Lae Shwe; Sureena Matayong; Suntorn Witosurapot – Education and Information Technologies, 2024
Multiple Choice Questions (MCQs) are an important evaluation technique for both examinations and learning activities. However, the manual creation of questions is time-consuming and challenging for teachers. Hence, there is a notable demand for an Automatic Question Generation (AQG) system. Several systems have been created for this aim, but the…
Descriptors: Difficulty Level, Computer Assisted Testing, Adaptive Testing, Multiple Choice Tests
Douglas-Morris, Jan; Ritchie, Helen; Willis, Catherine; Reed, Darren – Anatomical Sciences Education, 2021
Multiple-choice (MC) anatomy "spot-tests" (identification-based assessments on tagged cadaveric specimens) offer a practical alternative to traditional free-response (FR) spot-tests. Conversion of the two spot-tests in an upper limb musculoskeletal anatomy unit of study from FR to a novel MC format, where one of five tagged structures on…
Descriptors: Multiple Choice Tests, Anatomy, Test Reliability, Difficulty Level
Jin, Kuan-Yu; Siu, Wai-Lok; Huang, Xiaoting – Journal of Educational Measurement, 2022
Multiple-choice (MC) items are widely used in educational tests. Distractor analysis, an important procedure for checking the utility of response options within an MC item, can be readily implemented in the framework of item response theory (IRT). Although random guessing is a popular behavior of test-takers when answering MC items, none of the…
Descriptors: Guessing (Tests), Multiple Choice Tests, Item Response Theory, Attention
Thayaamol Upapong; Apantee Poonputta – Educational Process: International Journal, 2025
Background/purpose: The purposes of this research are to develop a reliable and valid assessment tool for measuring systems thinking skills in upper primary students in Thailand and to establish a normative criterion for evaluating their systems thinking abilities based on educational standards. Materials/methods: The study followed a three-phase…
Descriptors: Thinking Skills, Elementary School Students, Measures (Individuals), Foreign Countries
Eka Febri Zulissetiana; Muhammad Irfannuddin; Siti Sarahdeaz Fazzaura Putri; Syifa Alkaf; Susilawati Susilawati; Jihan Marshanda; Ra Fadila Septiany; Hasyimiah Az-Zahra; Robert G. Carroll – Advances in Physiology Education, 2024
Complex subjects such as physiology can be challenging for students to learn. These challenges are not uncommon in implementing the learning process in physiology and affect learning outcomes. Dramatization is an interactive and effective method to improve learning outcomes. In a project designed by senior medical students, junior medical students…
Descriptors: Drama, Teaching Methods, Physiology, Science Instruction
Kuo-Zheng Feng – Language Testing in Asia, 2024
This study addressed a gap in existing research on Multiple-Choice (MC) cloze tests by focusing on the learners' perspective, specifically examining the difficulties faced by vocational high school students (VHSs). A nationwide sample of 293 VHSs participated, providing both quantitative and qualitative data through a self-developed questionnaire.…
Descriptors: Language Tests, Multiple Choice Tests, Cloze Procedure, Student Attitudes
van den Broek, Gesa S. E.; Gerritsen, Suzanne L.; Oomen, Iris T. J.; Velthoven, Eva; van Boxtel, Femke H. J.; Kester, Liesbeth; van Gog, Tamara – Journal of Educational Psychology, 2023
Multiple-choice questions (MCQs) are popular in vocabulary software because they can be scored automatically and are compatible with many input devices (e.g., touchscreens). Answering MCQs is beneficial for learning, especially when learners retrieve knowledge from memory to evaluate plausible answer alternatives. However, such retrieval may not…
Descriptors: Multiple Choice Tests, Vocabulary Development, Test Format, Cues
Roger Young; Emily Courtney; Alexander Kah; Mariah Wilkerson; Yi-Hsin Chen – Teaching of Psychology, 2025
Background: Multiple-choice item (MCI) assessments are burdensome for instructors to develop. Artificial intelligence (AI, e.g., ChatGPT) can streamline the process without sacrificing quality. The quality of AI-generated MCIs and human experts is comparable. However, whether the quality of AI-generated MCIs is equally good across various domain-…
Descriptors: Item Response Theory, Multiple Choice Tests, Psychology, Textbooks
Liu, Chunyan; Jurich, Daniel; Morrison, Carol; Grabovsky, Irina – Applied Measurement in Education, 2021
The existence of outliers in the anchor items can be detrimental to the estimation of examinee ability and undermine the validity of score interpretation across forms. However, in practice, anchor item performance can become distorted due to various reasons. This study compares the performance of modified "INFIT" and "OUTFIT"…
Descriptors: Equated Scores, Test Items, Item Response Theory, Difficulty Level