Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 4 |
Descriptor
Difficulty Level | 5 |
Statistical Analysis | 5 |
Test Items | 4 |
Item Response Theory | 3 |
Computation | 2 |
Models | 2 |
Academic Standards | 1 |
Achievement Tests | 1 |
Bayesian Statistics | 1 |
Cognitive Tests | 1 |
Content Analysis | 1 |
More ▼ |
Source
Applied Measurement in… | 5 |
Author
Chen, Haiqin | 1 |
Karabatsos, George | 1 |
Lee, Won-Chan | 1 |
Lim, Euijin | 1 |
Lozano, José H. | 1 |
Newman, Dianna L. | 1 |
Revuelta, Javier | 1 |
Tang, Xiaodan | 1 |
Traynor, Anne | 1 |
Publication Type
Journal Articles | 5 |
Reports - Research | 5 |
Education Level
Elementary Education | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Tang, Xiaodan; Karabatsos, George; Chen, Haiqin – Applied Measurement in Education, 2020
In applications of item response theory (IRT) models, it is known that empirical violations of the local independence (LI) assumption can significantly bias parameter estimates. To address this issue, we propose a threshold-autoregressive item response theory (TAR-IRT) model that additionally accounts for order dependence among the item responses…
Descriptors: Item Response Theory, Test Items, Models, Computation
Lozano, José H.; Revuelta, Javier – Applied Measurement in Education, 2021
The present study proposes a Bayesian approach for estimating and testing the operation-specific learning model, a variant of the linear logistic test model that allows for the measurement of the learning that occurs during a test as a result of the repeated use of the operations involved in the items. The advantages of using a Bayesian framework…
Descriptors: Bayesian Statistics, Computation, Learning, Testing
Lim, Euijin; Lee, Won-Chan – Applied Measurement in Education, 2020
The purpose of this study is to address the necessity of subscore equating and to evaluate the performance of various equating methods for subtests. Assuming the random groups design and number-correct scoring, this paper analyzed real data and simulated data with four study factors including test dimensionality, subtest length, form difference in…
Descriptors: Equated Scores, Test Length, Test Format, Difficulty Level
Traynor, Anne – Applied Measurement in Education, 2017
It has long been argued that U.S. states' differential performance on nationwide assessments may reflect differences in students' opportunity to learn the tested content that is primarily due to variation in curricular content standards, rather than in instructional quality or educational investment. To quantify the effect of differences in…
Descriptors: Test Items, Difficulty Level, State Standards, Academic Standards

Newman, Dianna L.; And Others – Applied Measurement in Education, 1988
The effect of using statistical and cognitive item difficulty to determine item order on multiple-choice tests was examined, using 120 undergraduate students. Students performed better when items were ordered by increasing cognitive difficulty rather than decreasing difficulty. The statistical ordering of difficulty had little effect on…
Descriptors: Cognitive Tests, Difficulty Level, Higher Education, Multiple Choice Tests