Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Mathematical Models | 42 |
Test Items | 42 |
Test Theory | 42 |
Latent Trait Theory | 24 |
Item Analysis | 16 |
Test Construction | 15 |
Difficulty Level | 13 |
Achievement Tests | 11 |
Statistical Studies | 10 |
Criterion Referenced Tests | 9 |
Error of Measurement | 9 |
More ▼ |
Source
Author
Publication Type
Reports - Research | 33 |
Speeches/Meeting Papers | 14 |
Journal Articles | 12 |
Reports - Evaluative | 6 |
Opinion Papers | 3 |
Information Analyses | 2 |
Collected Works - General | 1 |
Collected Works - Proceedings | 1 |
Education Level
Audience
Researchers | 14 |
Location
United Kingdom (England) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 2 |
California Achievement Tests | 1 |
Comprehensive Tests of Basic… | 1 |
SAT (College Admission Test) | 1 |
Stanford Achievement Tests | 1 |
Students Evaluation of… | 1 |
What Works Clearinghouse Rating

Woodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores
Graham, James M. – Educational and Psychological Measurement, 2006
Coefficient alpha, the most commonly used estimate of internal consistency, is often considered a lower bound estimate of reliability, though the extent of its underestimation is not typically known. Many researchers are unaware that coefficient alpha is based on the essentially tau-equivalent measurement model. It is the violation of the…
Descriptors: Models, Test Theory, Reliability, Structural Equation Models
Reckase, Mark D.; McKinley, Robert L. – 1984
The purpose of this paper is to present a generalization of the concept of item difficulty to test items that measure more than one dimension. Three common definitions of item difficulty were considered: the proportion of correct responses for a group of individuals; the probability of a correct response to an item for a specific person; and the…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Mathematical Models

Reuterberg, Sven-Eric; Gustafsson, Jan-Eric – Educational and Psychological Measurement, 1992
The use of confirmatory factor analysis by the LISREL program is demonstrated as an assumption-testing method when computing reliability coefficients under different model assumptions. Results indicate that reliability estimates are robust against departure from the assumption of parallelism of test items. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Mathematical Models, Robustness (Statistics)

Jannarone, Robert J. – Psychometrika, 1986
Conjunctive item response models are introduced such that: (1) sufficient statistics for latent traits are not necessarily additive in item scores; (2) items are not necessarily locally independent; and (3) existing compensatory (additive) item response models including the binomial, Rasch, logistic, and general locally independent model are…
Descriptors: Cognitive Processes, Hypothesis Testing, Latent Trait Theory, Mathematical Models

Reckase, Mark D. – Psychological Assessment, 1996
Summarizes the current state of the art in test construction and contrasts it with previous conceptual models, some of which are wrong or misleading. New methodologies for item selection and review are presented, with current thinking on the specification of technical characteristics of tests. (Author/SLD)
Descriptors: Mathematical Models, Psychological Testing, Selection, State of the Art Reviews

Armstrong, Ronald D.; Jones, Douglas H. – Applied Psychological Measurement, 1992
Polynomial algorithms are presented that are used to solve selected problems in test theory, and computational results from sample problems with several hundred decision variables are provided that demonstrate the benefits of these algorithms. The algorithms are based on optimization theory in networks (graphs). (SLD)
Descriptors: Algorithms, Decision Making, Equations (Mathematics), Mathematical Models

Divgi, D. R. – Journal of Educational Measurement, 1986
This paper discusses various issues involved in using the Rasch Model with multiple-choice tests and questions the suitability of this model for multiple-choice items. Results of some past studies supporting the model are shown to be irrelevant. The effects of the model's misfit on test equating are demonstrated. (Author JAZ)
Descriptors: Equated Scores, Goodness of Fit, Latent Trait Theory, Mathematical Models

Feldt, Leonard S. – Educational and Psychological Measurement, 1984
The binomial error model includes form-to-form difficulty differences as error variance and leads to Ruder-Richardson formula 21 as an estimate of reliability. If the form-to-form component is removed from the estimate of error variance, the binomial model leads to KR 20 as the reliability estimate. (Author/BW)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Mathematical Formulas
Mellenbergh, Gideon J.; van der Linden, Wim J. – Evaluation in Education: International Progress, 1982
Three item selection methods for criterion-referenced tests are examined: the classical theory of item difficulty and item-test correlation; the latent trait theory of item characteristic curves; and a decision-theoretic approach for optimal item selection. Item contribution to the standardized expected utility of mastery testing is discussed. (CM)
Descriptors: Criterion Referenced Tests, Educational Testing, Item Analysis, Latent Trait Theory
Hutchinson, T. P. – 1984
One means of learning about the processes operating in a multiple choice test is to include some test items, called nonsense items, which have no correct answer. This paper compares two versions of a mathematical model of test performance to interpret test data that includes both genuine and nonsense items. One formula is based on the usual…
Descriptors: Foreign Countries, Guessing (Tests), Mathematical Models, Multiple Choice Tests
Levine, Michael V. – 1982
Significant to a latent trait or item response theory analysis of a mental test is the determination of exactly what is being quantified. The following are practical problems to be considered in the formulation of a good theory: (1) deciding whether two tests measure the same trait or traits; (2) analyzing the relative contributions of a pair of…
Descriptors: Item Analysis, Latent Trait Theory, Mathematical Models, Measurement Techniques

Hutchinson, T. P. – Contemporary Educational Psychology, 1986
Qualitative evidence for the operation of partial knowledge is given by two findings. First, performance when second and subsequent choices are made is above the chance level. Second, it is positively related to first choice performance. A number of theories incorporating partial knowledge are compared quantitatively. (Author/LMO)
Descriptors: Difficulty Level, Feedback, Goodness of Fit, Mathematical Models
Ackerman, Terry A. – 1987
One of the important underlying assumptions of all item response theory (IRT) models is that of local independence. This assumption requires that the response to an item on a test not be influenced by the response to any other items. This assumption is often taken for granted, with little or no scrutiny of the response process required to answer…
Descriptors: Computer Software, Correlation, Estimation (Mathematics), Latent Trait Theory
Wilcox, Rand R. – 1978
Two fundamental problems in mental test theory are to estimate true score and to estimate the amount of error when testing an examinee. In this report, three probability models which characterize a single test item in terms of a population of examinees are described. How these models may be modified to characterize a single examinee in terms of an…
Descriptors: Achievement Tests, Comparative Analysis, Error of Measurement, Mathematical Models