Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 4 |
Descriptor
Difficulty Level | 8 |
Multiple Choice Tests | 8 |
Test Items | 7 |
Test Construction | 3 |
Test Reliability | 3 |
Equated Scores | 2 |
High School Students | 2 |
Higher Education | 2 |
Ability | 1 |
Academic Achievement | 1 |
Academic Standards | 1 |
More ▼ |
Source
Applied Measurement in… | 8 |
Author
Ascalon, M. Evelina | 1 |
Beddow, Peter A. | 1 |
Bolt, Daniel M. | 1 |
Davis, Bruce W. | 1 |
Elliott, Stephen N. | 1 |
Feldt, Leonard S. | 1 |
Frary, Robert B. | 1 |
Grabovsky, Irina | 1 |
Jurich, Daniel | 1 |
Kettler, Ryan J. | 1 |
Kurz, Alexander | 1 |
More ▼ |
Publication Type
Journal Articles | 8 |
Reports - Research | 7 |
Reports - Evaluative | 1 |
Education Level
Elementary Secondary Education | 1 |
Grade 8 | 1 |
High Schools | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Secondary Education | 1 |
Audience
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Liu, Chunyan; Jurich, Daniel; Morrison, Carol; Grabovsky, Irina – Applied Measurement in Education, 2021
The existence of outliers in the anchor items can be detrimental to the estimation of examinee ability and undermine the validity of score interpretation across forms. However, in practice, anchor item performance can become distorted due to various reasons. This study compares the performance of modified "INFIT" and "OUTFIT"…
Descriptors: Equated Scores, Test Items, Item Response Theory, Difficulty Level
Papenberg, Martin; Musch, Jochen – Applied Measurement in Education, 2017
In multiple-choice tests, the quality of distractors may be more important than their number. We therefore examined the joint influence of distractor quality and quantity on test functioning by providing a sample of 5,793 participants with five parallel test sets consisting of items that differed in the number and quality of distractors.…
Descriptors: Multiple Choice Tests, Test Items, Test Validity, Test Reliability
Kettler, Ryan J.; Rodriguez, Michael C.; Bolt, Daniel M.; Elliott, Stephen N.; Beddow, Peter A.; Kurz, Alexander – Applied Measurement in Education, 2011
Federal policy on alternate assessment based on modified academic achievement standards (AA-MAS) inspired this research. Specifically, an experimental study was conducted to determine whether tests composed of modified items would have the same level of reliability as tests composed of original items, and whether these modified items helped reduce…
Descriptors: Multiple Choice Tests, Test Items, Alternative Assessment, Test Reliability
Ascalon, M. Evelina; Meyers, Lawrence S.; Davis, Bruce W.; Smits, Niels – Applied Measurement in Education, 2007
This article examined two item-writing guidelines: the format of the item stem and homogeneity of the answer set. Answering the call of Haladyna, Downing, and Rodriguez (2002) for empirical tests of item writing guidelines and extending the work of Smith and Smith (1988) on differential use of item characteristics, a mock multiple-choice driver's…
Descriptors: Guidelines, Difficulty Level, Standard Setting, Driver Education

Newman, Dianna L.; And Others – Applied Measurement in Education, 1988
The effect of using statistical and cognitive item difficulty to determine item order on multiple-choice tests was examined, using 120 undergraduate students. Students performed better when items were ordered by increasing cognitive difficulty rather than decreasing difficulty. The statistical ordering of difficulty had little effect on…
Descriptors: Cognitive Tests, Difficulty Level, Higher Education, Multiple Choice Tests

Wang, Xiang-bo; And Others – Applied Measurement in Education, 1995
An experiment is reported in which 225 high school students were asked to choose among several multiple-choice items but then were required to answer them all. It is concluded that allowing choice while having fair tests is only possible when choice is irrelevant in terms of difficulty. (SLD)
Descriptors: Adaptive Testing, Difficulty Level, Equated Scores, High School Students

Feldt, Leonard S. – Applied Measurement in Education, 1993
The recommendation that the reliability of multiple-choice tests will be enhanced if the distribution of item difficulties is concentrated at approximately 0.50 is reinforced and extended in this article by viewing the 0/1 item scoring as a dichotomization of an underlying normally distributed ability score. (SLD)
Descriptors: Ability, Difficulty Level, Guessing (Tests), Mathematical Models

Frary, Robert B. – Applied Measurement in Education, 1991
The use of the "none-of-the-above" option (NOTA) in 20 college-level multiple-choice tests was evaluated for classes with 100 or more students. Eight academic disciplines were represented, and 295 NOTA and 724 regular test items were used. It appears that the NOTA can be compatible with good classroom measurement. (TJH)
Descriptors: College Students, Comparative Testing, Difficulty Level, Discriminant Analysis