NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 17 results Save | Export
Thompson, Bruce; Melancon, Janet G. – 1996
This study investigated the benefits of creating item "testlets" or "parcels" in the context of structural equation modeling confirmatory factor analysis (CFA). Testlets are defined as groups of items related to a single content area that is developed as a unit. The strategy is illustrated using data from the administration of…
Descriptors: Statistical Distributions, Structural Equation Models, Test Construction
Matthews-Lopez, Joy L.; Hombo, Catherine M. – 2001
The purpose of this study was to examine the recovery of item parameters in simulated Automatic Item Generation (AIG) conditions, using Markov chain Monte Carlo (MCMC) estimation methods to attempt to recover the generating distributions. To do this, variability in item and ability parameters was manipulated. Realistic AIG conditions were…
Descriptors: Estimation (Mathematics), Monte Carlo Methods, Statistical Distributions, Test Construction
Peer reviewed Peer reviewed
de Gruijter, Dato N. M. – Journal of Educational Measurement, 1997
K. May and W. A. Nicewander recently concluded (1994) that percentile ranks are inferior or raw scores as indicators of latent ability. It is argued that their conclusions are incorrect, and an error in their derivation is identified. The incorrect equation results in an incorrect conclusion, as work by F. M. Lord (1980) also indicates.…
Descriptors: Equations (Mathematics), Estimation (Mathematics), Raw Scores, Statistical Distributions
Peer reviewed Peer reviewed
Enders, Craig K.; Bandalos, Deborah L. – Applied Measurement in Education, 1999
Examined the degree to which coefficient alpha is affected by including items with different distribution shapes within a unidimensional scale. Computer simulation results indicate that reliability does not increase dramatically as a result of using differentially shaped items within a scale. Discusses implications for test construction. (SLD)
Descriptors: Computer Simulation, Reliability, Scaling, Statistical Distributions
Peer reviewed Peer reviewed
Piedmont, Ralph L.; Hyland, Michael E. – Educational and Psychological Measurement, 1993
The use of mean inter-item correlation as a technique for examining homogeneity is proposed as a descriptive tool that can orient researchers to salient aspects of their scales. A study of 341 undergraduates who completed the NEO Personality Inventory illustrates the technique. (SLD)
Descriptors: Correlation, Evaluation Methods, Higher Education, Personality Measures
Peer reviewed Peer reviewed
Chan, Wai; And Others – Multivariate Behavioral Research, 1995
It is suggested that using an unbiased estimate of the weight matrix may eliminate the small or intermediate sample size bias of the asymptotically distribution-free (ADF) test statistic. Results of simulations show that test statistics based on the biased estimator or the unbiased estimate are highly similar. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Matrices, Sample Size
Peer reviewed Peer reviewed
Berger, Martijn P. F. – Applied Psychological Measurement, 1994
This paper focuses on similarities of optimal design of fixed-form tests, adaptive tests, and testlets within the framework of the general theory of optimal designs. A sequential design procedure is proposed that uses these similarities to obtain consistent estimates for the trait level distribution. (SLD)
Descriptors: Achievement Tests, Adaptive Testing, Algorithms, Estimation (Mathematics)
Peer reviewed Peer reviewed
Rogers, H. Jane; Swaminathan, Hariharan – Applied Psychological Measurement, 1993
Performance of the logistic regression (LR) procedure was compared to that of the Mantel Haenszel (MH) procedure in the detection of uniform and nonuniform differential item functioning on a simulation examining distributional properties of the LR and MH test statistics and the relative power of the two procedures. (SLD)
Descriptors: Comparative Analysis, Computer Simulation, Item Bias, Mathematical Models
Peer reviewed Peer reviewed
van der Linden, Wim J.; Luecht, Richard M. – Psychometrika, 1998
Derives a set of linear conditions of item-response functions that guarantees identical observed-score distributions on two test forms. The conditions can be added as constraints to a linear programming model for test assembly. An example illustrates the use of the model for an item pool from the Law School Admissions Test (LSAT). (SLD)
Descriptors: Equated Scores, Item Banks, Item Response Theory, Linear Programming
Peer reviewed Peer reviewed
Bramley, Tom – Evaluation & Research in Education, 2001
Analyzed data from a session of the General Certificate of Secondary Education (GCSE) mathematics examination to identify items displaying a bi-modal expected score distribution, try to explain the bi-modality, rescore the items to remove under-used middle categories, and determine the effect on test reliability of rescoring the data. Discusses…
Descriptors: Foreign Countries, Mathematics Tests, Reliability, Scores
van der Linden, Wim J.; Luecht, Richard M. – 1994
An optimization model is presented that allows test assemblers to control the shape of the observed-score distribution on a test for a population with a known ability distribution. An obvious application is for item response theory-based test assembly in programs where observed scores are reported and operational test forms are required to produce…
Descriptors: Ability, Foreign Countries, Heuristics, Item Response Theory
Peer reviewed Peer reviewed
Feldt, Leonard S. – Applied Measurement in Education, 1993
The recommendation that the reliability of multiple-choice tests will be enhanced if the distribution of item difficulties is concentrated at approximately 0.50 is reinforced and extended in this article by viewing the 0/1 item scoring as a dichotomization of an underlying normally distributed ability score. (SLD)
Descriptors: Ability, Difficulty Level, Guessing (Tests), Mathematical Models
Peer reviewed Peer reviewed
Luecht, Richard M.; Hirsch, Thomas M. – Applied Psychological Measurement, 1992
Derivations of several item selection algorithms for use in fitting test items to target information functions (IFs) are described. These algorithms, which use an average growth approximation of target IFs, were tested by generating six test forms and were found to provide reliable fit. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Equations (Mathematics), Goodness of Fit
Peer reviewed Peer reviewed
Wainer, Howard; And Others – Journal of Educational Measurement, 1992
Computer simulations were run to measure the relationship between testlet validity and factors of item pool size and testlet length for both adaptive and linearly constructed testlets. Making a testlet adaptive yields only modest increases in aggregate validity because of the peakedness of the typical proficiency distribution. (Author/SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Peer reviewed Peer reviewed
Schiel, Jeffrey L.; Shaw, Dale G. – Applied Measurement in Education, 1992
Changes in information retention resulting from changes in reliability and number of intervals in scale construction were studied to provide quantitative information to help in decisions about choosing intervals. Information retention reached a maximum when the number of intervals was about 8 or more and reliability was near 1.0. (SLD)
Descriptors: Decision Making, Knowledge Level, Mathematical Models, Monte Carlo Methods
Previous Page | Next Page ยป
Pages: 1  |  2