NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 23 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wenchao Ma; Miguel A. Sorrel; Xiaoming Zhai; Yuan Ge – Journal of Educational Measurement, 2024
Most existing diagnostic models are developed to detect whether students have mastered a set of skills of interest, but few have focused on identifying what scientific misconceptions students possess. This article developed a general dual-purpose model for simultaneously estimating students' overall ability and the presence and absence of…
Descriptors: Models, Misconceptions, Diagnostic Tests, Ability
Peer reviewed Peer reviewed
Direct linkDirect link
Ranger, Jochen; Kuhn, Jörg-Tobias; Wolgast, Anett – Journal of Educational Measurement, 2021
Van der Linden's hierarchical model for responses and response times can be used in order to infer the ability and mental speed of test takers from their responses and response times in an educational test. A standard approach for this is maximum likelihood estimation. In real-world applications, the data of some test takers might be partly…
Descriptors: Models, Reaction Time, Item Response Theory, Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Bolt, Daniel M.; Kim, Jee-Seon – Journal of Educational Measurement, 2018
Cognitive diagnosis models (CDMs) typically assume skill attributes with discrete (often binary) levels of skill mastery, making the existence of skill continuity an anticipated form of model misspecification. In this article, misspecification due to skill continuity is argued to be of particular concern for several CDM applications due to the…
Descriptors: Cognitive Measurement, Models, Mastery Learning, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Cheng, Ying; Liu, Cheng – Journal of Educational Measurement, 2016
For a certification, licensure, or placement exam, allowing examinees to take multiple attempts at the test could effectively change the pass rate. Change in the pass rate can occur without any change in the underlying latent trait, and can be an artifact of multiple attempts and imperfect reliability of the test. By deriving formulae to compute…
Descriptors: Testing, Computation, Change, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Lathrop, Quinn N.; Cheng, Ying – Journal of Educational Measurement, 2014
When cut scores for classifications occur on the total score scale, popular methods for estimating classification accuracy (CA) and classification consistency (CC) require assumptions about a parametric form of the test scores or about a parametric response model, such as item response theory (IRT). This article develops an approach to estimate CA…
Descriptors: Cutting Scores, Classification, Computation, Nonparametric Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Liang, Tie; Wells, Craig S.; Hambleton, Ronald K. – Journal of Educational Measurement, 2014
As item response theory has been more widely applied, investigating the fit of a parametric model becomes an important part of the measurement process. There is a lack of promising solutions to the detection of model misfit in IRT. Douglas and Cohen introduced a general nonparametric approach, RISE (Root Integrated Squared Error), for detecting…
Descriptors: Item Response Theory, Measurement Techniques, Nonparametric Statistics, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Huang, Hung-Yu; Wang, Wen-Chung – Journal of Educational Measurement, 2014
The DINA (deterministic input, noisy, and gate) model has been widely used in cognitive diagnosis tests and in the process of test development. The outcomes known as slip and guess are included in the DINA model function representing the responses to the items. This study aimed to extend the DINA model by using the random-effect approach to allow…
Descriptors: Models, Guessing (Tests), Probability, Ability
Peer reviewed Peer reviewed
Nandakumar, Ratna; Yu, Feng – Journal of Educational Measurement, 1996
DIMTEST is a nonparametric statistical test procedure for assessing unidimensionality of binary item response data that uses the T-statistic of W. F. Stout (1987). This study investigates the performance of the T-statistic with respect to different shapes of ability distributions and confirms its nonparametric nature. (SLD)
Descriptors: Ability, Nonparametric Statistics, Statistical Distributions, Validity
Peer reviewed Peer reviewed
Oshima, T. C. – Journal of Educational Measurement, 1994
The effect of violating the assumption of nonspeededness on ability and item parameter estimates in item response theory was studied through simulation under three speededness conditions. Results indicate that ability estimation was least affected by speededness but that substantial effects on item parameter estimates were found. (SLD)
Descriptors: Ability, Computer Simulation, Estimation (Mathematics), Item Response Theory
Peer reviewed Peer reviewed
Camilli, Gregory; Penfield, Douglas A. – Journal of Educational Measurement, 1997
The simultaneous assessment of differential item functioning (DIF) for a collection of test items through an index that measures the variance of DIF on a test as an indicator of the degree to which different items show DIF in different directions is proposed and evaluated through simulations. (SLD)
Descriptors: Ability, Estimation (Mathematics), Item Bias, Item Response Theory
Peer reviewed Peer reviewed
Caruso, John C.; Witkiewitz, Katie – Journal of Educational Measurement, 2002
As an alternative to equally weighted difference scores, examined an orthogonal reliable component analysis (RCA) solution and an oblique principal components analysis (PCA) solution for the standardization sample of the Kaufman Assessment Battery for Children (KABC; A. Kaufman and N. Kaufman, 1983). Discusses the practical implications of the…
Descriptors: Ability, Academic Achievement, Children, Factor Analysis
Peer reviewed Peer reviewed
Vispoel, Walter P.; Rocklin, Thomas R.; Wang, Tianyou; Bleiler, Timothy – Journal of Educational Measurement, 1999
Investigated the effectiveness of H. Wainer's (1993) strategy for obtaining positively biased ability estimates when examinees can review and change answers on computerized adaptive tests. Results, based on simulation and testing data from 87 college students, show that the Wainer strategy sometimes yields inflated ability estimates and sometimes…
Descriptors: Ability, College Students, Computer Assisted Testing, Higher Education
Peer reviewed Peer reviewed
Mazor, Kathleen M.; And Others – Journal of Educational Measurement, 1995
This article describes a method of incorporating two ability estimates into a single regression analysis so that substantially fewer items exhibit differential item functioning (DIF). Comparative analysis with the Mantel-Haenszel procedure also shows more accurate matching of examinees, preventing identifying multidimensional item impact as DIF.…
Descriptors: Ability, Comparative Analysis, Estimation (Mathematics), Identification
Peer reviewed Peer reviewed
Wang, Tianyou; Vispoel, Walter P. – Journal of Educational Measurement, 1998
Used simulations of computerized adaptive tests to evaluate results yielded by four commonly used ability estimation methods: maximum likelihood estimation (MLE) and three Bayesian approaches. Results show clear distinctions between MLE and Bayesian methods. (SLD)
Descriptors: Ability, Adaptive Testing, Bayesian Statistics, Computer Assisted Testing
Peer reviewed Peer reviewed
Zwick, Rebecca; And Others – Journal of Educational Measurement, 1995
In a simulation study of ability and estimation of differential item functioning (DIF) in computerized adaptive tests, Rasch-based DIF statistics were highly correlated with generating DIF, but DIF statistics tended to be slightly smaller than in the three-parameter logistic model analyses. (SLD)
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Computer Simulation
Previous Page | Next Page »
Pages: 1  |  2