Publication Date
In 2025 | 1 |
Since 2024 | 3 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 14 |
Since 2006 (last 20 years) | 44 |
Descriptor
Models | 57 |
Test Items | 20 |
Item Response Theory | 18 |
Psychometrics | 17 |
Foreign Countries | 13 |
Comparative Analysis | 12 |
Evaluation Methods | 10 |
Scores | 10 |
Statistical Analysis | 10 |
Testing | 10 |
Goodness of Fit | 9 |
More ▼ |
Source
International Journal of… | 57 |
Author
Bartram, Dave | 3 |
Rupp, Andre A. | 3 |
Baghaei, Purya | 2 |
Bradshaw, Laine P. | 2 |
Chen, Yi-Hsin | 2 |
Gorin, Joanna S. | 2 |
Tatsuoka, Kikumi K. | 2 |
Tay, Louis | 2 |
Ackerman, Terry | 1 |
Arce-Ferrer, Alvaro J. | 1 |
Aryadoust, Vahid | 1 |
More ▼ |
Publication Type
Education Level
Higher Education | 8 |
Grade 4 | 6 |
Elementary Education | 5 |
Postsecondary Education | 5 |
Secondary Education | 4 |
High Schools | 3 |
Intermediate Grades | 3 |
Elementary Secondary Education | 2 |
Adult Education | 1 |
Grade 3 | 1 |
Audience
Practitioners | 1 |
Researchers | 1 |
Location
Canada | 3 |
United States | 3 |
Argentina | 1 |
Arizona | 1 |
Australia | 1 |
Belgium | 1 |
China | 1 |
France | 1 |
Germany | 1 |
Iran | 1 |
Malaysia | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Xiaowen Liu – International Journal of Testing, 2024
Differential item functioning (DIF) often arises from multiple sources. Within the context of multidimensional item response theory, this study examined DIF items with varying secondary dimensions using the three DIF methods: SIBTEST, Mantel-Haenszel, and logistic regression. The effect of the number of secondary dimensions on DIF detection rates…
Descriptors: Item Analysis, Test Items, Item Response Theory, Correlation
Fu, Yanyan; Strachan, Tyler; Ip, Edward H.; Willse, John T.; Chen, Shyh-Huei; Ackerman, Terry – International Journal of Testing, 2020
This research examined correlation estimates between latent abilities when using the two-dimensional and three-dimensional compensatory and noncompensatory item response theory models. Simulation study results showed that the recovery of the latent correlation was best when the test contained 100% of simple structure items for all models and…
Descriptors: Item Response Theory, Models, Test Items, Simulation
Finkelman, Matthew D.; de la Torre, Jimmy; Karp, Jeremy A. – International Journal of Testing, 2020
Cognitive diagnosis models (CDMs) have been studied as a means of providing detailed diagnostic information about the skills that have been mastered, and the skills that have not, by examinees. Prior research has examined the use of automated test assembly (ATA) alongside CDMs; however, no previous study has investigated how to perform ATA when a…
Descriptors: Cognitive Measurement, Models, Automation, Test Construction
The Choice between Cognitive Diagnosis and Item Response Theory: A Case Study from Medical Education
Youn Seon Lim; Catherine Bangeranye – International Journal of Testing, 2024
Feedback is a powerful instructional tool for motivating learning. But effective feedback, requires that instructors have accurate information about their students' current knowledge status and their learning progress. In modern educational measurement, two major theoretical perspectives on student ability and proficiency can be distinguished.…
Descriptors: Cognitive Measurement, Diagnostic Tests, Item Response Theory, Case Studies
Kim, Kyung Yong; Lim, Euijin; Lee, Won-Chan – International Journal of Testing, 2019
For passage-based tests, items that belong to a common passage often violate the local independence assumption of unidimensional item response theory (UIRT). In this case, ignoring local item dependence (LID) and estimating item parameters using a UIRT model could be problematic because doing so might result in inaccurate parameter estimates,…
Descriptors: Item Response Theory, Equated Scores, Test Items, Models
George, Ann Cathrice; Robitzsch, Alexander – International Journal of Testing, 2021
Modern large-scale studies such as the Progress in International Reading Literacy Study (PIRLS) do not only report reading competence of students on a global reading scale but also report reading on the level of reading subskills. However, the number of and the dependencies between the subskills are frequently discussed. In this study, different…
Descriptors: Foreign Countries, Grade 4, Achievement Tests, International Assessment
Cao, Mengyang; Song, Q. Chelsea; Tay, Louis – International Journal of Testing, 2018
There is a growing use of noncognitive assessments around the world, and recent research has posited an ideal point response process underlying such measures. A critical issue is whether the typical use of dominance approaches (e.g., average scores, factor analysis, and the Samejima's graded response model) in scoring such measures is adequate.…
Descriptors: Comparative Analysis, Item Response Theory, Factor Analysis, Models
Ravand, Hamdollah; Baghaei, Purya – International Journal of Testing, 2020
More than three decades after their introduction, diagnostic classification models (DCM) do not seem to have been implemented in educational systems for the purposes they were devised. Most DCM research is either methodological for model development and refinement or retrofitting to existing nondiagnostic tests and, in the latter case, basically…
Descriptors: Classification, Models, Diagnostic Tests, Test Construction
Hu, Jinxiang; Miller, M. David; Huggins-Manley, Anne Corinne; Chen, Yi-Hsin – International Journal of Testing, 2016
Cognitive diagnosis models (CDMs) estimate student ability profiles using latent attributes. Model fit to the data needs to be ascertained in order to determine whether inferences from CDMs are valid. This study investigated the usefulness of some popular model fit statistics to detect CDM fit including relative fit indices (AIC, BIC, and CAIC),…
Descriptors: Models, Goodness of Fit, Psychometrics, Ability
Bradshaw, Laine P.; Madison, Matthew J. – International Journal of Testing, 2016
In item response theory (IRT), the invariance property states that item parameter estimates are independent of the examinee sample, and examinee ability estimates are independent of the test items. While this property has long been established and understood by the measurement community for IRT models, the same cannot be said for diagnostic…
Descriptors: Classification, Models, Simulation, Psychometrics
Wang, Ting; Li, Min; Thummaphan, Phonraphee; Ruiz-Primo, Maria Araceli – International Journal of Testing, 2017
Contextualized items have been widely used in science testing. Despite common use of item contexts, how the influence of a chosen context on the reliability and validity of the score inferences remains unclear. We focused on sequential cues of contextual information, referring to the order of events or descriptions presented in item contexts. We…
Descriptors: Science Tests, Cues, Difficulty Level, Test Items
Tay, Louis; Vermunt, Jeroen K.; Wang, Chun – International Journal of Testing, 2013
We evaluate the item response theory with covariates (IRT-C) procedure for assessing differential item functioning (DIF) without preknowledge of anchor items (Tay, Newman, & Vermunt, 2011). This procedure begins with a fully constrained baseline model, and candidate items are tested for uniform and/or nonuniform DIF using the Wald statistic.…
Descriptors: Item Response Theory, Test Bias, Models, Statistical Analysis
Sessoms, John; Finney, Sara J. – International Journal of Testing, 2015
Because schools worldwide use low-stakes tests to make important decisions, value-added indices computed from test scores must accurately reflect student learning, which requires equal test-taking effort across testing occasions. Evaluating change in effort assumes effort is measured equivalently across occasions. We evaluated the longitudinal…
Descriptors: Testing, Measurement, Models, Change
Kunina-Habenicht, Olga; Rupp, André A.; Wilhelm, Oliver – International Journal of Testing, 2017
Diagnostic classification models (DCMs) hold great potential for applications in summative and formative assessment by providing discrete multivariate proficiency scores that yield statistically driven classifications of students. Using data from a newly developed diagnostic arithmetic assessment that was administered to 2032 fourth-grade students…
Descriptors: Grade 4, Foreign Countries, Classification, Mathematics Tests