Publication Date
| In 2026 | 0 |
| Since 2025 | 12 |
| Since 2022 (last 5 years) | 56 |
| Since 2017 (last 10 years) | 108 |
| Since 2007 (last 20 years) | 255 |
Descriptor
| Sample Size | 404 |
| Simulation | 404 |
| Item Response Theory | 113 |
| Statistical Analysis | 92 |
| Error of Measurement | 86 |
| Models | 84 |
| Test Items | 80 |
| Comparative Analysis | 77 |
| Monte Carlo Methods | 75 |
| Correlation | 67 |
| Evaluation Methods | 64 |
| More ▼ | |
Source
Author
| Fan, Xitao | 7 |
| Beretvas, S. Natasha | 5 |
| Algina, James | 4 |
| Chan, Wai | 4 |
| Cohen, Allan S. | 4 |
| De Champlain, Andre | 4 |
| Finch, W. Holmes | 4 |
| French, Brian F. | 4 |
| Kim, Seock-Ho | 4 |
| Kromrey, Jeffrey D. | 4 |
| Paek, Insu | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Teachers | 4 |
| Researchers | 3 |
Location
| North Carolina | 2 |
| Armenia | 1 |
| Austria | 1 |
| Canada | 1 |
| Florida (Miami) | 1 |
| Hong Kong | 1 |
| Indiana | 1 |
| Iran | 1 |
| Montana | 1 |
| New York (New York) | 1 |
| Norway | 1 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedBroodbooks, Wendy J.; Elmore, Patricia B. – Educational and Psychological Measurement, 1987
The effects of sample size, number of variables, and population value of the congruence coefficient on the sampling distribution of the congruence coefficient were examined. Sample data were generated on the basis of the common factor model, and principal axes factor analyses were performed. (Author/LMO)
Descriptors: Factor Analysis, Mathematical Models, Monte Carlo Methods, Predictor Variables
Tang, K. Linda; And Others – 1993
This study compared the performance of the LOGIST and BILOG computer programs on item response theory (IRT) based scaling and equating for the Test of English as a Foreign Language (TOEFL) using real and simulated data and two calibration structures. Applications of IRT for the TOEFL program are based on the three-parameter logistic (3PL) model.…
Descriptors: Comparative Analysis, Computer Simulation, Equated Scores, Estimation (Mathematics)
Kromrey, Jeffrey D.; Rendina-Gobioff, Gianna – Educational and Psychological Measurement, 2006
The performance of methods for detecting publication bias in meta-analysis was evaluated using Monte Carlo methods. Four methods of bias detection were investigated: Begg's rank correlation, Egger's regression, funnel plot regression, and trim and fill. Five factors were included in the simulation design: number of primary studies in each…
Descriptors: Comparative Analysis, Meta Analysis, Monte Carlo Methods, Correlation
Seco, Guillermo Vallejo; Izquierdo, Marcelino Cuesta; Garcia, M. Paula Fernandez; Diez, F. Javier Herrero – Educational and Psychological Measurement, 2006
The authors compare the operating characteristics of the bootstrap-F approach, a direct extension of the work of Berkovits, Hancock, and Nevitt, with Huynh's improved general approximation (IGA) and the Brown-Forsythe (BF) multivariate approach in a mixed repeated measures design when normality and multisample sphericity assumptions do not hold.…
Descriptors: Sample Size, Comparative Analysis, Simulation, Multivariate Analysis
Kim, Seock-Ho; Cohen, Allan S. – 1991
Studies of differential item functioning (DIF) under item response theory require that item parameter estimates be placed on the same metric before comparisons can be made. Evidence that methods for linking metrics may be influenced by the presence of differentially functioning items has been inconsistent. The effects of three methods for linking…
Descriptors: Chi Square, Comparative Analysis, Equations (Mathematics), Estimation (Mathematics)
Peer reviewedThomas, Hoben – Journal of Educational Statistics, 1986
This paper is concerned with the construction of effect size standard errors in situations where the effect sizes are independent but the data have likely been sampled from non-normal distributions, and possibly for different studies, from different families of non-normal distributions. Asymptotic distribution-free estimators are provided for two…
Descriptors: Control Groups, Effect Size, Equations (Mathematics), Error of Measurement
Bonett, Douglas G. – Applied Psychological Measurement, 2006
Comparing variability of test scores across alternate forms, test conditions, or subpopulations is a fundamental problem in psychometrics. A confidence interval for a ratio of standard deviations is proposed that performs as well as the classic method with normal distributions and performs dramatically better with nonnormal distributions. A simple…
Descriptors: Intervals, Mathematical Concepts, Comparative Analysis, Psychometrics
Oil and Gas on Indian Reservations: Statistical Methods Help to Establish Value for Royalty Purposes
Fowler, Mary S.; Kadane, Joseph B. – Journal of Statistics Education, 2006
Part of the history of oil and gas development on Indian reservations concerns potential underpayment of royalties due to under-valuation of production by oil companies. This paper discusses a model used by the Shoshone and Arapaho tribes in a lawsuit against the Federal government, claiming the Government failed to collect adequate royalties.…
Descriptors: Fuels, American Indian Education, Federal Government, Probability
De Champlain, Andre; Gessaroli, Marc E. – 1996
The use of indices and statistics based on nonlinear factor analysis (NLFA) has become increasingly popular as a means of assessing the dimensionality of an item response matrix. Although the indices and statistics currently available to the practitioner have been shown to be useful and accurate in many testing situations, few studies have…
Descriptors: Adaptive Testing, Chi Square, Computer Assisted Testing, Factor Analysis
Schumacker, Randall E.; And Others – 1994
Rasch between and total weighted and unweighted fit statistics were compared using varying test lengths and sample sizes. Two test lengths (20 and 50 items) and three sample sizes (150, 500, and 1,000 were crossed. Each of the six combinations were replicated 100 times. In addition, power comparisons were made. Results indicated that there were no…
Descriptors: Comparative Analysis, Goodness of Fit, Item Response Theory, Power (Statistics)
Reckase, Mark D. – 1978
Five comparisons were made relative to the quality of estimates of ability parameters and item calibrations obtained from the one-parameter and three-parameter logistic models. The results indicate: (1) The three-parameter model fit the test data better in all cases than did the one-parameter model. For simulation data sets, multi-factor data were…
Descriptors: Comparative Analysis, Goodness of Fit, Item Analysis, Mathematical Models
Boldt, R. F. – 1994
The comparison of item response theory models for the Test of English as a Foreign Language (TOEFL) was extended to an equating context as simulation trials were used to "equate the test to itself." Equating sample data were generated from administration of identical item sets. Equatings that used procedures based on each model (simple…
Descriptors: Comparative Analysis, Cutting Scores, English (Second Language), Equated Scores
Olejnik, Stephen F.; Algina, James – 1986
Sampling distributions for ten tests for comparing population variances in a two group design were generated for several combinations of equal and unequal sample sizes, population means, and group variances when distributional forms differed. The ten procedures included: (1) O'Brien's (OB); (2) O'Brien's with adjusted degrees of freedom; (3)…
Descriptors: Error of Measurement, Evaluation Methods, Measurement Techniques, Nonparametric Statistics
Ree, Malcom James; Jensen, Harald E. – 1980
By means of computer simulation of test responses, the reliability of item analysis data and the accuracy of equating were examined for hypothetical samples of 250, 500, 1000, and 2000 subjects for two tests with 20 equating items plus 60 additional items on the same scale. Birnbaum's three-parameter logistic model was used for the simulation. The…
Descriptors: Computer Assisted Testing, Equated Scores, Error of Measurement, Item Analysis
Gierl, Mark J.; Leighton, Jacqueline P.; Tan, Xuan – Journal of Educational Measurement, 2006
DETECT, the acronym for Dimensionality Evaluation To Enumerate Contributing Traits, is an innovative and relatively new nonparametric dimensionality assessment procedure used to identify mutually exclusive, dimensionally homogeneous clusters of items using a genetic algorithm ( Zhang & Stout, 1999). Because the clusters of items are mutually…
Descriptors: Program Evaluation, Cluster Grouping, Evaluation Methods, Multivariate Analysis

Direct link
