NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 511 to 525 of 639 results Save | Export
Davey, Tim; Pommerich, Mary; Thompson, Tony D. – 1999
In computerized adaptive testing (CAT), new or experimental items are frequently administered alongside operational tests to gather the pretest data needed to replenish and replace item pools. The two basic strategies used to combine pretest and operational items are embedding and appending. Variable-length CATs are preferred because of the…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Measurement Techniques
Peer reviewed Peer reviewed
Wainer, Howard; And Others – Journal of Educational Measurement, 1992
Computer simulations were run to measure the relationship between testlet validity and factors of item pool size and testlet length for both adaptive and linearly constructed testlets. Making a testlet adaptive yields only modest increases in aggregate validity because of the peakedness of the typical proficiency distribution. (Author/SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Shu-Ying; Ankenman, Robert D. – Journal of Educational Measurement, 2004
The purpose of this study was to compare the effects of four item selection rules--(1) Fisher information (F), (2) Fisher information with a posterior distribution (FP), (3) Kullback-Leibler information with a posterior distribution (KP), and (4) completely randomized item selection (RN)--with respect to the precision of trait estimation and the…
Descriptors: Test Length, Adaptive Testing, Computer Assisted Testing, Test Selection
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, Holmes – Applied Psychological Measurement, 2005
This study compares the ability of the multiple indicators, multiple causes (MIMIC) confirmatory factor analysis model to correctly identify cases of differential item functioning (DIF) with more established methods. Although the MIMIC model might have application in identifying DIF for multiple grouping variables, there has been little…
Descriptors: Identification, Factor Analysis, Test Bias, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Su, Ya-Hui – Applied Psychological Measurement, 2004
Eight independent variables (differential item functioning [DIF] detection method, purification procedure, item response model, mean latent trait difference between groups, test length, DIF pattern, magnitude of DIF, and percentage of DIF items) were manipulated, and two dependent variables (Type I error and power) were assessed through…
Descriptors: Test Length, Test Bias, Simulation, Item Response Theory
Rodriguez-Aragon, Graciela; And Others – 1993
The predictive power of the Split-Half version of the Wechsler Intelligence Scale for Children--Revised (WISC-R) Object Assembly (OA) subtest was compared to that of the full administration of the OA subtest. A cohort of 218 male and 49 female adolescent offenders detained in a Texas juvenile detention facility between 1990 and 1992 was used. The…
Descriptors: Adolescents, Cohort Analysis, Comparative Testing, Correlation
Epstein, Kenneth I.; Steinheiser, Frederick H., Jr. – 1978
A multiparameter, programmable model was developed to examine the interactive influence of certain parameters on the probability of deciding that an examinee had attained a specified degree of mastery. It was applied within the simulated context of performance testing of military trainees. These parameters included: (1) the number of assumed…
Descriptors: Academic Ability, Bayesian Statistics, Cutting Scores, Hypothesis Testing
Steinheiser, Frederick H., Jr. – 1976
A computer simulation of Bayes' Theorem was conducted in order to determine the probability that an examinee was a master conditional upon his test score. The inputs were: number of mastery states assumed, test length, prior expectation of masters in the examinee population, and conditional probability of a master getting a randomly selected test…
Descriptors: Bayesian Statistics, Classification, Computer Programs, Criterion Referenced Tests
Peer reviewed Peer reviewed
Huynh, Huynh; Casteel, Jim – Journal of Experimental Education, 1987
In the context of pass/fail decisions, using the Bock multi-nominal latent trait model for moderate-length tests does not produce decisions that differ substantially from those based on the raw scores. The Bock decisions appear to relate less strongly to outside criteria than those based on the raw scores. (Author/JAZ)
Descriptors: Cutting Scores, Error Patterns, Grade 6, Intermediate Grades
Peer reviewed Peer reviewed
Harrington, Robert G.; Jennings, Valerie – Contemporary Educational Psychology, 1986
Three short forms of the McCarthy Scales of Children's Abilities (MSCA) have been developed to screen the cognitive skills of young children suspected of learning disorders and developmental delays. Correlations were obtained between scores on the full form of the MSCA and the Kaufman, Taylor, and McCarthy Screening Test short forms. (Author/LMO)
Descriptors: Cognitive Tests, Comparative Testing, Correlation, Early Childhood Education
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Chen, Cheng-Te – Educational and Psychological Measurement, 2005
This study investigates item parameter recovery, standard error estimates, and fit statistics yielded by the WINSTEPS program under the Rasch model and the rating scale model through Monte Carlo simulations. The independent variables were item response model, test length, and sample size. WINSTEPS yielded practically unbiased estimates for the…
Descriptors: Statistics, Test Length, Rating Scales, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Chen, Hsueh-Chu – Educational and Psychological Measurement, 2004
As item response theory (IRT) becomes popular in educational and psychological testing, there is a need of reporting IRT-based effect size measures. In this study, we show how the standardized mean difference can be generalized into such a measure. A disattenuation procedure based on the IRT test reliability is proposed to correct the attenuation…
Descriptors: Test Reliability, Rating Scales, Sample Size, Error of Measurement
Bergstrom, Betty A.; Gershon, Richard – 1992
The most useful method of item selection for making pass-fail decisions with a Computerized Adaptive Test (CAT) was studied. Medical technology students (n=86) took a computer adaptive test in which items were targeted to the ability of the examinee. The adaptive algorithm that selected items and estimated person measures used the Rasch model and…
Descriptors: Adaptive Testing, Algorithms, Comparative Analysis, Computer Assisted Testing
Parr, Phyllis; And Others – 1996
In order to establish a tertiary educational sector that is able to provide equal opportunities to all students, it is vital that institutions diversify curriculum delivery and course assessment strategies. Although it may appear impractical to develop a course that takes into consideration the various learning preferences of all individuals,…
Descriptors: College Faculty, Disabilities, Educational Assessment, Equal Education
Cohen, Allan S.; Kim, Seock-Ho – 1993
Equating tests from different calibrations under item response theory (IRT) requires calculation of the slope and intercept of the appropriate linear transformation. Two methods have been proposed recently for equating graded response items under IRT, a test characteristic curve method and a minimum chi-square method. These two methods are…
Descriptors: Chi Square, Comparative Analysis, Computer Simulation, Equated Scores
Pages: 1  |  ...  |  31  |  32  |  33  |  34  |  35  |  36  |  37  |  38  |  39  |  ...  |  43