Descriptor
Difficulty Level | 15 |
Mathematical Models | 15 |
Test Theory | 15 |
Test Items | 13 |
Latent Trait Theory | 10 |
Achievement Tests | 5 |
Item Analysis | 5 |
Statistical Studies | 5 |
Test Construction | 5 |
Estimation (Mathematics) | 4 |
Item Banks | 4 |
More ▼ |
Source
Contemporary Educational… | 1 |
Educational and Psychological… | 1 |
Journal of Educational… | 1 |
Psychometrika | 1 |
Author
Publication Type
Reports - Research | 13 |
Speeches/Meeting Papers | 6 |
Journal Articles | 4 |
Opinion Papers | 1 |
Reports - Evaluative | 1 |
Education Level
Audience
Researchers | 5 |
Location
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 1 |
What Works Clearinghouse Rating
Reckase, Mark D.; McKinley, Robert L. – 1984
The purpose of this paper is to present a generalization of the concept of item difficulty to test items that measure more than one dimension. Three common definitions of item difficulty were considered: the proportion of correct responses for a group of individuals; the probability of a correct response to an item for a specific person; and the…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Mathematical Models

Feldt, Leonard S. – Educational and Psychological Measurement, 1984
The binomial error model includes form-to-form difficulty differences as error variance and leads to Ruder-Richardson formula 21 as an estimate of reliability. If the form-to-form component is removed from the estimate of error variance, the binomial model leads to KR 20 as the reliability estimate. (Author/BW)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Mathematical Formulas

Hutchinson, T. P. – Contemporary Educational Psychology, 1986
Qualitative evidence for the operation of partial knowledge is given by two findings. First, performance when second and subsequent choices are made is above the chance level. Second, it is positively related to first choice performance. A number of theories incorporating partial knowledge are compared quantitatively. (Author/LMO)
Descriptors: Difficulty Level, Feedback, Goodness of Fit, Mathematical Models

Ramsay, James O. – Psychometrika, 1989
An alternative to the Rasch model is introduced. It characterizes strength of response according to the ratio of ability and difficulty parameters rather than their difference. Joint estimation and marginal estimation models are applied to two test data sets. (SLD)
Descriptors: Ability, Bayesian Statistics, College Entrance Examinations, Comparative Analysis
Engelhard, George, Jr. – 1980
The Rasch model is described as a latent trait model which meets the five criteria that characterize reasonable and objective measurements of an individual's ability independent of the test items used. The criteria are: (1) calibration of test items must be independent of particular norming groups; (2) measurement of individuals must be…
Descriptors: Achievement Tests, Difficulty Level, Elementary Secondary Education, Equated Scores

Jansen, Margo G. H. – Journal of Educational Statistics, 1986
In this paper a Bayesian procedure is developed for the simultaneous estimation of the reading ability and difficulty parameters which are assumed to be factors in reading errors by the multiplicative Poisson Model. According to several criteria, the Bayesian estimates are better than comparable maximum likelihood estimates. (Author/JAZ)
Descriptors: Achievement Tests, Bayesian Statistics, Comparative Analysis, Difficulty Level
Livingston, Samuel A. – 1986
This paper deals with test fairness regarding a test consisting of two parts: (1) a "common" section, taken by all students; and (2) a "variable" section, in which some students may answer a different set of questions from other students. For example, a test taken by several thousand students each year contains a common multiple-choice portion and…
Descriptors: Difficulty Level, Error of Measurement, Essay Tests, Mathematical Models
Zwick, Rebecca – 1986
Although perfectly scalable items rarely occur in practice, Guttman's concept of a scale has proved to be valuable to the development of measurement theory. If the score distribution is uniform and there is an equal number of items at each difficulty level, both the elements and the eigenvalues of the Pearson correlation matrix of dichotomous…
Descriptors: Correlation, Difficulty Level, Item Analysis, Latent Trait Theory

Cook, Linda L.; Hambleton, Ronald K. – 1978
Latent trait models may offer considerable potential for the improvement of educational measurement practices, but until recently, they have received only limited attention from measurement specialists. This paper provides a brief introduction to latent trait models, and provides test practitioners with a non-technical introduction to the…
Descriptors: Career Development, Criterion Referenced Tests, Difficulty Level, Item Analysis
Forster, Fred – 1987
Studies carried out over a 12-year period addressed fundamental questions on the use of Rasch-based item banks. Large field tests administered in grades 3-8 of reading, mathematics, and science items, as well as standardized test results were used to explore the possible effects of many factors on item calibrations. In general, the results…
Descriptors: Achievement Tests, Difficulty Level, Elementary Education, Item Analysis
Curry, Allen R.; And Others – 1978
The efficacy of employing subsets of items from a calibrated item pool to estimate the Rasch model person parameters was investigated. Specifically, the degree of invariance of Rasch model ability-parameter estimates was examined across differing collections of simulated items. The ability-parameter estimates were obtained from a simulation of…
Descriptors: Career Development, Difficulty Level, Equated Scores, Error of Measurement
Ackerman, Terry A. – 1987
The purpose of this study was to investigate the effect of using multidimensional items in a computer adaptive test (CAT) setting which assumes a unidimensional item response theory (IRT) framework. Previous research has suggested that the composite of multidimensional abilities being estimated by a unidimensional IRT model is not constant…
Descriptors: Adaptive Testing, College Entrance Examinations, Computer Assisted Testing, Computer Simulation
Thomas, Gregory P. – 1986
This paper argues that no single measurement strategy serves all purposes and that applying methods and techniques which allow a variety of data elements to be retrieved and juxtaposed may be an investment in the future. Item response theory, Rasch model, and latent trait theory are all approaches to a single conceptual topic. An abbreviated look…
Descriptors: Achievement Tests, Adaptive Testing, Criterion Referenced Tests, Data Collection
Wainer, Howard; Kiely, Gerard L. – 1986
Recent experience with the Computerized Adaptive Test (CAT) has raised a number of concerns about its practical applications. The concerns are principally involved with the concept of having the computer construct the test from a precalibrated item pool, and substituting statistical characteristics for the test developer's skills. Problems with…
Descriptors: Adaptive Testing, Algorithms, Computer Assisted Testing, Construct Validity
Hambleton, Ronald K.; Rogers, H. Jane – 1986
This report was designed to respond to two major methodological shortcomings in the item bias literature: (1) misfitting test models; and (2) the use of significance tests. Specifically, the goals of the research were to describe a newly developed method known as the "plot method" for identifying potentially biased test items and to…
Descriptors: Criterion Referenced Tests, Culture Fair Tests, Difficulty Level, Estimation (Mathematics)