Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 1 |
Descriptor
| Mathematical Models | 29 |
| Test Reliability | 29 |
| Test Theory | 29 |
| Career Development | 10 |
| Error of Measurement | 10 |
| Item Analysis | 8 |
| Statistical Analysis | 8 |
| Latent Trait Theory | 7 |
| Test Items | 7 |
| Criterion Referenced Tests | 6 |
| Test Construction | 6 |
| More ▼ | |
Source
| Educational and Psychological… | 7 |
| Psychometrika | 3 |
| Journal of Experimental… | 2 |
| Assessment & Evaluation in… | 1 |
| Journal of Educational… | 1 |
| Journal of Educational… | 1 |
Author
| Zimmerman, Donald W. | 3 |
| Schulman, Robert S. | 2 |
| Wilcox, Rand R. | 2 |
| Budescu, David | 1 |
| Burton, Richard F. | 1 |
| Cahan, Sorel | 1 |
| Cason, Gerald J. | 1 |
| Cliff, Norman | 1 |
| Cohen, Allan S., Comp. | 1 |
| Cook, Linda L. | 1 |
| Ecob, Russell | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 20 |
| Journal Articles | 13 |
| Speeches/Meeting Papers | 5 |
| Reports - Evaluative | 3 |
| Reference Materials -… | 2 |
| Reports - Descriptive | 2 |
| Collected Works - General | 1 |
| Opinion Papers | 1 |
Education Level
Audience
| Researchers | 2 |
Location
| United Kingdom (Great Britain) | 1 |
Laws, Policies, & Programs
| Elementary and Secondary… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Zumbo, Bruno D.; Kroc, Edward – Educational and Psychological Measurement, 2019
Chalmers recently published a critique of the use of ordinal a[alpha] proposed in Zumbo et al. as a measure of test reliability in certain research settings. In this response, we take up the task of refuting Chalmers' critique. We identify three broad misconceptions that characterize Chalmers' criticisms: (1) confusing assumptions with…
Descriptors: Test Reliability, Statistical Analysis, Misconceptions, Mathematical Models
Peer reviewedSchulman, Robert S.; Haden, Richard L. – Psychometrika, 1975
A model is proposed for the description of ordinal test scores based on the definition of true score as expected rank; its deviations are compared with results from classical test theory. An unbiased estimator of population true score from sample data is calculated. Score variance and population reliability are examined. (Author/BJG)
Descriptors: Career Development, Mathematical Models, Test Reliability, Test Theory
Peer reviewedWilcox, Rand R. – Journal of Educational Statistics, 1981
Both the binomial and beta-binomial models are applied to various problems occurring in mental test theory. The paper reviews and critiques these models. The emphasis is on the extensions of the models that have been proposed in recent years, and that might not be familiar to many educators. (Author)
Descriptors: Error of Measurement, Item Analysis, Mathematical Models, Test Reliability
Peer reviewedYarnold, Paul R. – Educational and Psychological Measurement, 1984
Unreliable profiles impose the difficulty that ordinal and interval relations among the individual's scores become uncertain or unstable. A profile reliability coefficient is derived to estimate the relative expected extent of this ordinal and interval "inversion" for any profile of K measures. (Author/DWH)
Descriptors: Error of Measurement, Mathematical Models, Profiles, Test Reliability
Peer reviewedWilliams, Richard H.; Zimmerman, Donald W. – Journal of Experimental Education, 1982
A mathematical link between test reliability and test validity is derived, taking into account the correlation between error scores on a test and error scores on a criterion measure. When this correlation is positive, the "paradoxical" nonmonotonic relation between test reliability and test validity occurs universally. (Author/BW)
Descriptors: Correlation, Error of Measurement, Mathematical Models, Test Reliability
Peer reviewedKraemer, Helena Chmura – Psychometrika, 1981
Limitations and extensions of Feldt's approach to testing the equality of Cronbach's alpha coefficients in independent and matched samples are discussed. In particular, this approach is used to test equality of intraclass correlation coefficients. (Author)
Descriptors: Analysis of Variance, Correlation, Hypothesis Testing, Mathematical Models
Peer reviewedJones, W. Paul – Educational and Psychological Measurement, 1991
A Bayesian alternative to interpretations based on classical reliability theory is presented. Procedures are detailed for calculation of a posterior score and credible interval with joint consideration of item sample and occasion error. (Author/SLD)
Descriptors: Bayesian Statistics, Equations (Mathematics), Mathematical Models, Statistical Inference
Peer reviewedZimmerman, Donald W.; And Others – Journal of Experimental Education, 1981
Reliability coefficients of linear combinations of observed scores have anomalous properties which have led to difficulties in the investigation of difference scores and gain scores in test theory. Discrepancies between classical results and correct results obtained from more general formulas, which allow for correlated errors, are examined…
Descriptors: Error of Measurement, Mathematical Formulas, Mathematical Models, Scores
Peer reviewedZimmerman, Donald W. – Educational and Psychological Measurement, 1976
Using the concepts of conditional probability, conditional expectation, and conditional independence, the main results of the classical test theory model can be derived in a very few steps with minimal assumptions. The present effort explores the possibility that present classical test theories can be further condensed. (Author/RC)
Descriptors: Career Development, Correlation, Mathematical Models, Measurement
Peer reviewedReuterberg, Sven-Eric; Gustafsson, Jan-Eric – Educational and Psychological Measurement, 1992
The use of confirmatory factor analysis by the LISREL program is demonstrated as an assumption-testing method when computing reliability coefficients under different model assumptions. Results indicate that reliability estimates are robust against departure from the assumption of parallelism of test items. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Mathematical Models, Robustness (Statistics)
Peer reviewedBurton, Richard F. – Assessment & Evaluation in Higher Education, 2001
Describes four measures of test unreliability that quantify effects of question selection and guessing, both separately and together--three chosen for immediacy and one for greater mathematical elegance. Quantifies their dependence on test length and number of answer options per question. Concludes that many multiple choice tests are unreliable…
Descriptors: Guessing (Tests), Mathematical Models, Multiple Choice Tests, Objective Tests
Peer reviewedSchulman, Robert S. – Psychometrika, 1979
An alternative to the uniform probability distribution model for ordinal data is considered. Implications for statistics and for test theory are discussed. (JKS)
Descriptors: Career Development, Correlation, Mathematical Models, Nonparametric Statistics
Peer reviewedBudescu, David – Journal of Educational Measurement, 1985
An important determinant of equating process efficiency is the correlation between the anchor test and components of each form. Use of some monotonic function of this correlation as a measure of equating efficiency is suggested. A model relating anchor test length and test reliability to this measure of efficiency is presented. (Author/DWH)
Descriptors: Correlation, Equated Scores, Mathematical Models, Standardized Tests
Peer reviewedFeldt, Leonard S. – Educational and Psychological Measurement, 1984
The binomial error model includes form-to-form difficulty differences as error variance and leads to Ruder-Richardson formula 21 as an estimate of reliability. If the form-to-form component is removed from the estimate of error variance, the binomial model leads to KR 20 as the reliability estimate. (Author/BW)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Mathematical Formulas
PDF pending restorationLovett, Hubert T. – 1975
The reliability of a criterion referenced test was defined as a measure of the degree to which the test discriminates between an individual's level of performance and a predetermined criterion level. The variances of observed and true scores were defined as the squared deviation of the score from the criterion. Based on these definitions and the…
Descriptors: Career Development, Comparative Analysis, Criterion Referenced Tests, Mathematical Models
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
