NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 9 results Save | Export
Peer reviewed Peer reviewed
Huynh, Huynh – Journal of Educational Statistics, 1986
Under the assumptions of classical measurement theory and the condition of normality, a formula is derived for the reliability of composite scores. The formula represents an extension of the Spearman-Brown formula to the case of truncated data. (Author/JAZ)
Descriptors: Computer Simulation, Error of Measurement, Expectancy Tables, Scoring Formulas
Peer reviewed Peer reviewed
Harris, Deborah J.; Subkoviak, Michael J. – Educational and Psychological Measurement, 1986
This study examined three statistical methods for selecting items for mastery tests: (1) pretest-posttest; (2) latent trait; and (3) agreement statistics. The correlation between the latent trait method and agreement statistics, proposed here as an alternative, was substantial. Results for the pretest-posttest method confirmed its reputed…
Descriptors: Computer Simulation, Correlation, Item Analysis, Latent Trait Theory
Morrison, Carol A.; Fitzpatrick, Steven J. – 1992
An attempt was made to determine which item response theory (IRT) equating method results in the least amount of equating error or "scale drift" when equating scores across one or more test forms. An internal anchor test design was employed with five different test forms, each consisting of 30 items, 10 in common with the base test and 5…
Descriptors: Comparative Analysis, Computer Simulation, Equated Scores, Error of Measurement
van der Linden, Wim J.; Adema, Jos J. – 1988
Two optimalization models for the construction of tests with a maximal value of coefficient alpha are given. Both models have a linear form and can be solved by using a branch-and-bound algorithm. The first model assumes an item bank calibrated under the Rasch model and can be used, for instance, when classical test theory has to serve as an…
Descriptors: Algorithms, Computer Simulation, Estimation (Mathematics), Foreign Countries
Weitzman, R. A. – 1982
The goal of this research was to predict from a recruit's responses to the Armed Services Vocational Aptitude Battery (ASVAB) items whether the recruit would pass the Armed Forces Qualification Test (AFQT). The data consisted of the responses (correct/incorrect) of 1,020 Navy recruits to 200 items of the ASVAB together with the scores of these…
Descriptors: Adults, Armed Forces, Computer Oriented Programs, Computer Simulation
Peer reviewed Peer reviewed
Harrison, David A. – Journal of Educational Statistics, 1986
Multidimensional item response data were created. The strength of a general factor, the number of common factors, the distribution of items loadingon common factors, and the number of items in simulated tests were manipulated. LOGIST effectively recovered both item and trait parameters in nearly all of the experimental conditions. (Author/JAZ)
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Simulation, Correlation
Ackerman, Terry A. – 1987
The purpose of this study was to investigate the effect of using multidimensional items in a computer adaptive test (CAT) setting which assumes a unidimensional item response theory (IRT) framework. Previous research has suggested that the composite of multidimensional abilities being estimated by a unidimensional IRT model is not constant…
Descriptors: Adaptive Testing, College Entrance Examinations, Computer Assisted Testing, Computer Simulation
Vale, C. David; Gialluca, Kathleen A. – 1985
ASCAL is a microcomputer-based program for calibrating items according to the three-parameter logistic model of item response theory. It uses a modified multivariate Newton-Raphson procedure for estimating item parameters. This study evaluated this procedure using Monte Carlo Simulation Techniques. The current version of ASCAL was then compared to…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Computer Simulation
Levine, Michael V.; Drasgow, Fritz – 1984
Some examinees' test-taking behavior may be so idiosyncratic that their scores are not comparable to the scores of more typical examinees. Appropriateness indices, which provide quantitative measures of response-pattern atypicality, can be viewed as statistics for testing a null hypothesis of normal test-taking behavior against an alternative…
Descriptors: Cheating, College Entrance Examinations, Computer Simulation, Estimation (Mathematics)