Descriptor
Mathematical Models | 35 |
Test Items | 35 |
Test Reliability | 35 |
Item Analysis | 17 |
Difficulty Level | 16 |
Test Construction | 13 |
Error of Measurement | 12 |
Latent Trait Theory | 11 |
Criterion Referenced Tests | 10 |
Statistical Analysis | 10 |
Comparative Analysis | 8 |
More ▼ |
Source
Educational and Psychological… | 3 |
Applied Measurement in… | 1 |
Applied Psychological… | 1 |
Journal of Educational… | 1 |
Multivariate Behavioral… | 1 |
Psychometrika | 1 |
Author
Reckase, Mark D. | 3 |
Benson, Jeri | 2 |
Feldt, Leonard S. | 2 |
Patience, Wayne M. | 2 |
Wilcox, Rand R. | 2 |
Ackerman, Terry A. | 1 |
Algina, James | 1 |
Bejar, Issac I. | 1 |
Bernknopf, Stan | 1 |
Cliff, Norman | 1 |
Cobern, William W. | 1 |
More ▼ |
Publication Type
Reports - Research | 24 |
Speeches/Meeting Papers | 11 |
Journal Articles | 7 |
Reports - Evaluative | 7 |
Collected Works - General | 1 |
Computer Programs | 1 |
Guides - General | 1 |
Reports - Descriptive | 1 |
Reports - General | 1 |
Education Level
Audience
Researchers | 3 |
Location
Florida | 1 |
Georgia | 1 |
South Carolina | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Comprehensive Tests of Basic… | 1 |
Minnesota Multiphasic… | 1 |
SAT (College Admission Test) | 1 |
School and College Ability… | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating

Hsu, Louis M. – Multivariate Behavioral Research, 1992
D.V. Budescu and J.L. Rogers (1981) proposed a method of adjusting correlations of scales to eliminate spurious components resulting from the overlapping of scales. Three reliability correction formulas are derived in this article that are based on more tenable assumptions. (SLD)
Descriptors: Correlation, Equations (Mathematics), Mathematical Models, Personality Measures

Reuterberg, Sven-Eric; Gustafsson, Jan-Eric – Educational and Psychological Measurement, 1992
The use of confirmatory factor analysis by the LISREL program is demonstrated as an assumption-testing method when computing reliability coefficients under different model assumptions. Results indicate that reliability estimates are robust against departure from the assumption of parallelism of test items. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Mathematical Models, Robustness (Statistics)

Kane, Michael; Moloney, James – Applied Psychological Measurement, 1978
The answer-until-correct (AUC) procedure requires that examinees respond to a multi-choice item until they answer it correctly. Using a modified version of Horst's model for examinee behavior, this paper compares the effect of guessing on item reliability for the AUC procedure and the zero-one scoring procedure. (Author/CTM)
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests

Feldt, Leonard S. – Educational and Psychological Measurement, 1984
The binomial error model includes form-to-form difficulty differences as error variance and leads to Ruder-Richardson formula 21 as an estimate of reliability. If the form-to-form component is removed from the estimate of error variance, the binomial model leads to KR 20 as the reliability estimate. (Author/BW)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Mathematical Formulas

Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models
Samejima, Fumiko – 1990
Test validity is a concept that has often been ignored in the context of latent trait models and in modern test theory, particularly as it relates to computerized adaptive testing. Some considerations about the validity of a test and of a single item are proposed. This paper focuses on measures that are population-free and that will provide local…
Descriptors: Adaptive Testing, Computer Assisted Testing, Equations (Mathematics), Item Response Theory
Divgi, D. R. – 1978
One aim of criterion-referenced testing is to classify an examinee without reference to a norm group; therefore, any statements about the dependability of such classification ought to be group-independent also. A population-independent index is proposed in terms of the probability of incorrect classification near the cutoff true score. The…
Descriptors: Criterion Referenced Tests, Cutting Scores, Difficulty Level, Error of Measurement

Cobern, William W. – 1986
This computer program, written in BASIC, performs three different calculations of test reliability: (1) the Kuder-Richardson method; (2); the "common split-half" method; and (3) the Rulon-Guttman split-half method. The program reads sequential access data files for microcomputers that have been set up by statistical packages such as…
Descriptors: Computer Software, Difficulty Level, Educational Research, Equations (Mathematics)

Feldt, Leonard S. – Applied Measurement in Education, 1993
The recommendation that the reliability of multiple-choice tests will be enhanced if the distribution of item difficulties is concentrated at approximately 0.50 is reinforced and extended in this article by viewing the 0/1 item scoring as a dichotomization of an underlying normally distributed ability score. (SLD)
Descriptors: Ability, Difficulty Level, Guessing (Tests), Mathematical Models
Wilcox, Rand R. – 1981
These studies in test adequacy focus on two problems: procedures for estimating reliability, and techniques for identifying ineffective distractors. Fourteen papers are presented on recent advances in measuring achievement (a response to Molenaar); "an extension of the Dirichlet-multinomial model that allows true score and guessing to be…
Descriptors: Achievement Tests, Criterion Referenced Tests, Guessing (Tests), Mathematical Models
Douglass, James B. – 1979
A general process for testing the feasibility of applying alternative mathematical or statistical models to the solution of a practical problem is presented and flowcharted. The system is used to develop a plan to compare models for test equating. The five alternative models to be considered for equating are: (1) anchor test equating using…
Descriptors: Equated Scores, Error of Measurement, Latent Trait Theory, Mathematical Models
Benson, Jeri – 1979
Two methods of item selection were used to select sets of 40 items from a 50-item verbal analogies test, and the resulting item sets were compared for relative efficiency. The BICAL program was used to select the 40 items having the best mean square fit to the one parameter logistic (Rasch) model. The LOGIST program was used to select the 40 items…
Descriptors: Comparative Analysis, Computer Programs, Costs, Efficiency

Huck, Schuyler W.; And Others – Educational and Psychological Measurement, 1981
Believing that examinee-by-item interaction should be conceptualized as true score variability rather than as a result of errors of measurement, Lu proposed a modification of Hoyt's analysis of variance reliability procedure. Via a computer simulation study, it is shown that Lu's approach does not separate interaction from error. (Author/RL)
Descriptors: Analysis of Variance, Comparative Analysis, Computer Programs, Difficulty Level
Ackerman, Terry A.; Evans, John A. – 1992
The relationship between levels of reliability and the power of two bias and differential item functioning (DIF) detection methods is examined. Both methods, the Mantel-Haenszel (MH) procedure of P. W. Holland and D. T. Thayer (1988) and the Simultaneous Item Bias (SIB) procedure of R. Shealy and W. Stout (1991), use examinees' raw scores as a…
Descriptors: Comparative Analysis, Equations (Mathematics), Error of Measurement, Item Bias
Bejar, Issac I. – 1976
The concept of testing for partial knowledge is considered with the concept of tailored testing. Following the special usage of latent trait theory, the word valdity is used to mean the correlation of a test with the construct the test measures. The concept of a method factor in the test is also considered as a part of the validity. The possible…
Descriptors: Achievement Tests, Adaptive Testing, Computer Assisted Testing, Confidence Testing