NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 24 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Lei, Pui-Wa; Li, Hongli – Applied Psychological Measurement, 2013
Minimum sample sizes of about 200 to 250 per group are often recommended for differential item functioning (DIF) analyses. However, there are times when sample sizes for one or both groups of interest are smaller than 200 due to practical constraints. This study attempts to examine the performance of Simultaneous Item Bias Test (SIBTEST),…
Descriptors: Sample Size, Test Bias, Computation, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Fukuhara, Hirotaka; Kamata, Akihito – Applied Psychological Measurement, 2011
A differential item functioning (DIF) detection method for testlet-based data was proposed and evaluated in this study. The proposed DIF model is an extension of a bifactor multidimensional item response theory (MIRT) model for testlets. Unlike traditional item response theory (IRT) DIF models, the proposed model takes testlet effects into…
Descriptors: Item Response Theory, Test Bias, Test Items, Bayesian Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Jinming – Applied Psychological Measurement, 2012
It is common to assume during a statistical analysis of a multiscale assessment that the assessment is composed of several unidimensional subtests or that it has simple structure. Under this assumption, the unidimensional and multidimensional approaches can be used to estimate item parameters. These two approaches are equivalent in parameter…
Descriptors: Simulation, Computation, Models, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Gu, Fei; Skorupski, William P.; Hoyle, Larry; Kingston, Neal M. – Applied Psychological Measurement, 2011
Ramsay-curve item response theory (RC-IRT) is a nonparametric procedure that estimates the latent trait using splines, and no distributional assumption about the latent trait is required. For item parameters of the two-parameter logistic (2-PL), three-parameter logistic (3-PL), and polytomous IRT models, RC-IRT can provide more accurate estimates…
Descriptors: Intervals, Item Response Theory, Models, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Green, Bert F. – Applied Psychological Measurement, 2011
This article refutes a recent claim that computer-based tests produce biased scores for very proficient test takers who make mistakes on one or two initial items and that the "bias" can be reduced by using a four-parameter IRT model. Because the same effect occurs with pattern scores on nonadaptive tests, the effect results from IRT scoring, not…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Bias, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
de la Torre, Jimmy; Hong, Yuan – Applied Psychological Measurement, 2010
Sample size ranks as one of the most important factors that affect the item calibration task. However, due to practical concerns (e.g., item exposure) items are typically calibrated with much smaller samples than what is desired. To address the need for a more flexible framework that can be used in small sample item calibration, this article…
Descriptors: Sample Size, Markov Processes, Tests, Data Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
de la Torre, Jimmy; Song, Hao – Applied Psychological Measurement, 2009
Assessments consisting of different domains (e.g., content areas, objectives) are typically multidimensional in nature but are commonly assumed to be unidimensional for estimation purposes. The different domains of these assessments are further treated as multi-unidimensional tests for the purpose of obtaining diagnostic information. However, when…
Descriptors: Ability, Tests, Item Response Theory, Data Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Cheng-Te; Wang, Wen-Chung – Applied Psychological Measurement, 2007
This study explores the effects of ignoring item interaction on item parameter estimation and the efficiency of using the local dependence index Q[subscript 3] and the SAS NLMIXED procedure to detect item interaction under the three-parameter logistic model and the generalized partial credit model. Through simulations, it was found that ignoring…
Descriptors: Models, Item Response Theory, Simulation, Generalization
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Bo; Walker, Cindy M. – Applied Psychological Measurement, 2008
The purpose of this research was to examine the effects of missing data on person-model fit and person trait estimation in tests with dichotomous items. Under the missing-completely-at-random framework, four missing data treatment techniques were investigated including pairwise deletion, coding missing responses as incorrect, hotdeck imputation,…
Descriptors: Item Response Theory, Computation, Goodness of Fit, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Cohen, Jon; Chan, Tsze; Jiang, Tao; Seburn, Mary – Applied Psychological Measurement, 2008
U.S. state educational testing programs administer tests to track student progress and hold schools accountable for educational outcomes. Methods from item response theory, especially Rasch models, are usually used to equate different forms of a test. The most popular method for estimating Rasch models yields inconsistent estimates and relies on…
Descriptors: Testing Programs, Educational Testing, Item Response Theory, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
de la Torre, Jimmy – Applied Psychological Measurement, 2008
Recent work has shown that multidimensionally scoring responses from different tests can provide better ability estimates. For educational assessment data, applications of this approach have been limited to binary scores. Of the different variants, the de la Torre and Patz model is considered more general because implementing the scoring procedure…
Descriptors: Markov Processes, Scoring, Data Analysis, Item Response Theory
Peer reviewed Peer reviewed
Ackerman, Terry – Applied Psychological Measurement, 1996
This paper illustrates how graphical analyses can enhance the interpretation and understanding of multidimensional item response theory (IRT) analyses. Examples of different graphical representations are presented, including item response surfaces, information vectors, and centroid plots of trait distributions. All graphs are intended to…
Descriptors: Data Analysis, Educational Testing, Item Response Theory
Peer reviewed Peer reviewed
Blashfield, Roger K.; Aldenderfer, Mark S. – Applied Psychological Measurement, 1979
Eight programs which perform iterative partitioning cluster analysis were analyzed in terms of versatility of options, accuracy, and cost. It was not possible to recommend any one program because of the variability in programs and the lack of knowledge about iterative partitioning methods. (Author/MH)
Descriptors: Cluster Analysis, Computer Programs, Cost Effectiveness, Data Analysis
Peer reviewed Peer reviewed
Overall, John E.; Woodward, – Applied Psychological Measurement, 1977
A method for studying relationships among groups in terms of categorical data patterns is described. The procedure yields a dimensional representation of configural relationships among multiple groups and a quantitative scaling of categorical data patterns for use in subsequent assignment of new individuals to the groups. (Author/CTM)
Descriptors: Clinical Diagnosis, Data Analysis, Discriminant Analysis, Mental Disorders
Peer reviewed Peer reviewed
Ackerman, Terry – Applied Psychological Measurement, 1996
This special issue is devoted to current developments in multidimensional item response theory (MIRT). The six papers included in this issue (and two for the March issue) present applications of MIRT to practical testing problems and demonstrate innovative MIRT techniques for assessment. (SLD)
Descriptors: Data Analysis, Educational Assessment, Educational Testing, Item Response Theory
Previous Page | Next Page ยป
Pages: 1  |  2