Descriptor
Mathematical Models | 13 |
Statistical Distributions | 13 |
Test Items | 13 |
Equations (Mathematics) | 7 |
Item Response Theory | 6 |
Estimation (Mathematics) | 5 |
Computer Simulation | 4 |
Goodness of Fit | 4 |
Item Bias | 4 |
Scores | 4 |
Ability | 3 |
More ▼ |
Source
Applied Psychological… | 3 |
Educational and Psychological… | 3 |
Journal of Educational… | 2 |
Applied Measurement in… | 1 |
Psychometrika | 1 |
Author
Publication Type
Journal Articles | 10 |
Reports - Evaluative | 8 |
Reports - Research | 5 |
Speeches/Meeting Papers | 3 |
Reports - Descriptive | 1 |
Education Level
Audience
Researchers | 2 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Monte Carlo Based Null Distribution for an Alternative Goodness-of-Fit Test Statistic in IRT Models.

Stone, Clement A. – Journal of Educational Measurement, 2000
Describes a goodness-of-fit statistic that considers the imprecision with which ability is estimated and involves constructing item fit tables based on each examinee's posterior distribution of ability, given the likelihood of the response pattern and an assumed marginal ability distribution. Also describes a Monte Carlo resampling procedure to…
Descriptors: Goodness of Fit, Item Response Theory, Mathematical Models, Monte Carlo Methods

Smith, Richard M. – Educational and Psychological Measurement, 1991
This study reports results of an investigation based on simulated data of the distributional properties of the item fit statistics that are commonly used in the Rasch model calibration programs as indices of the fit of responses to individual items to the measurement model. (SLD)
Descriptors: Computer Simulation, Equations (Mathematics), Goodness of Fit, Item Response Theory

Feldt, Leonard S. – Applied Measurement in Education, 1993
The recommendation that the reliability of multiple-choice tests will be enhanced if the distribution of item difficulties is concentrated at approximately 0.50 is reinforced and extended in this article by viewing the 0/1 item scoring as a dichotomization of an underlying normally distributed ability score. (SLD)
Descriptors: Ability, Difficulty Level, Guessing (Tests), Mathematical Models

Seong, Tae-Je – Applied Psychological Measurement, 1990
The sensitivity of marginal maximum likelihood estimation of item and ability (theta) parameters was examined when prior ability distributions were not matched to underlying ability distributions. Thirty sets of 45-item test data were generated. Conditions affecting the accuracy of estimation are discussed. (SLD)
Descriptors: Ability, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)

Mislevy, Robert J. – Psychometrika, 1984
Assuming vectors of item responses depend on ability through a fully specified item response model, this paper presents maximum likelihood equations for estimating the population parameters without estimating an ability parameter for each subject. Asymptotic standard errors, tests of fit, computing approximations, and details of four special cases…
Descriptors: Bayesian Statistics, Estimation (Mathematics), Goodness of Fit, Latent Trait Theory

Luecht, Richard M.; Hirsch, Thomas M. – Applied Psychological Measurement, 1992
Derivations of several item selection algorithms for use in fitting test items to target information functions (IFs) are described. These algorithms, which use an average growth approximation of target IFs, were tested by generating six test forms and were found to provide reliable fit. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Equations (Mathematics), Goodness of Fit
Reckase, Mark D.; And Others – 1985
Factor analysis is the traditional method for studying the dimensionality of test data. However, under common conditions, the factor analysis of tetrachoric correlations does not recover the underlying structure of dichotomous data. The purpose of this paper is to demonstrate that the factor analyses of tetrachoric correlations is unlikely to…
Descriptors: Correlation, Difficulty Level, Factor Analysis, Item Analysis

Dodd, Barbara G.; And Others – Educational and Psychological Measurement, 1993
Effects of the following variables on performance of computerized adaptive testing (CAT) procedures for the partial credit model (PCM) were studied: (1) stopping rule for terminating CAT; (2) item pool size; and (3) distribution of item difficulties. Implications of findings for CAT systems based on the PCM are discussed. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Simulation, Difficulty Level
Thompson, Bruce; Borrello, Gloria M. – 1987
Attitude measures frequently produce distributions of item scores that attenuate interitem correlations and thus also distort findings regarding the factor structure underlying the items. An actual data set involving 260 adult subjects' responses to 55 items on the Love Relationships Scale is employed to illustrate empirical methods for…
Descriptors: Adults, Analysis of Covariance, Attitude Measures, Correlation

Fowler, Robert L.; Clingman, Joy M. – Educational and Psychological Measurement, 1992
Monte Carlo techniques are used to examine the power of the "B" statistic of R. L. Brennan (1972) to detect negatively discriminating items drawn from a variety of nonnormal population distributions. A simplified procedure is offered for conducting an item-discrimination analysis on typical classroom objective tests. (SLD)
Descriptors: Classroom Techniques, Elementary Secondary Education, Equations (Mathematics), Item Analysis

Camilli, Gregory – Applied Psychological Measurement, 1992
A mathematical model is proposed to describe how group differences in distributions of abilities, which are distinct from the target ability, influence the probability of a correct item response. In the multidimensional approach, differential item functioning is considered a function of the educational histories of the examinees. (SLD)
Descriptors: Ability, Comparative Analysis, Equations (Mathematics), Factor Analysis
Spray, Judith A.; Miller, Timothy R. – 1992
A popular method of analyzing test items for differential item functioning (DIF) is to compute a statistic that conditions samples of examinees from different populations on an estimate of ability. This conditioning or matching by ability is intended to produce an appropriate statistic that is sensitive to true differences in item functioning,…
Descriptors: Blacks, College Entrance Examinations, Comparative Testing, Computer Simulation

Cohen, Allan S.; And Others – Journal of Educational Measurement, 1991
Detecting differential item functioning (DIF) on test items constructed to favor 1 group over another was investigated on parameter estimates from 2 item response theory-based computer programs--BILOG and LOGIST--using data for 1,000 White and 1,000 Black college students. Use of prior distributions and marginal-maximum a posteriori estimation is…
Descriptors: Black Students, College Students, Computer Assisted Testing, Equations (Mathematics)