Descriptor
Comparative Analysis | 12 |
Mathematical Models | 12 |
Test Theory | 12 |
Achievement Tests | 4 |
Criterion Referenced Tests | 4 |
Estimation (Mathematics) | 4 |
Statistical Analysis | 4 |
Bayesian Statistics | 3 |
Error of Measurement | 3 |
Higher Education | 3 |
Item Analysis | 3 |
More ▼ |
Author
Haladyna, Tom | 2 |
Roid, Gale | 2 |
Choppin, Bruce | 1 |
Cope, Ronald T. | 1 |
Everson, Howard | 1 |
Jansen, Margo G. H. | 1 |
Jarjoura, David | 1 |
Lovett, Hubert T. | 1 |
Medoff, Deborah R. | 1 |
Millsap, Roger E. | 1 |
O'Grady, Kevin E. | 1 |
More ▼ |
Publication Type
Reports - Research | 9 |
Journal Articles | 5 |
Reports - Evaluative | 4 |
Speeches/Meeting Papers | 3 |
Education Level
Audience
Researchers | 2 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating

Jarjoura, David – Journal of Educational Statistics, 1985
Issues regarding tolerance and confidence intervals are discussed within the context of educational measurement, and conceptual distinctions are drawn between these two types of intervals. Points are raised about the advantages of tolerance intervals when the focus is on a particular observed score rather than a particular examinee. (Author/BW)
Descriptors: Comparative Analysis, Error of Measurement, Mathematical Models, Test Interpretation

Millsap, Roger E.; Everson, Howard – Multivariate Behavioral Research, 1991
Use of confirmatory factor analysis (CFA) with nonzero latent means in testing six different measurement models from classical test theory is discussed. Implications of the six models for observed mean and covariance structures are described, and three examples of the use of CFA in testing the models are presented. (SLD)
Descriptors: Comparative Analysis, Equations (Mathematics), Goodness of Fit, Mathematical Models

Lovett, Hubert T. – 1975
The reliability of a criterion referenced test was defined as a measure of the degree to which the test discriminates between an individual's level of performance and a predetermined criterion level. The variances of observed and true scores were defined as the squared deviation of the score from the criterion. Based on these definitions and the…
Descriptors: Career Development, Comparative Analysis, Criterion Referenced Tests, Mathematical Models

Ramsay, James O. – Psychometrika, 1989
An alternative to the Rasch model is introduced. It characterizes strength of response according to the ratio of ability and difficulty parameters rather than their difference. Joint estimation and marginal estimation models are applied to two test data sets. (SLD)
Descriptors: Ability, Bayesian Statistics, College Entrance Examinations, Comparative Analysis

O'Grady, Kevin E.; Medoff, Deborah R. – Multivariate Behavioral Research, 1991
A procedure for evaluating a variety of rater reliability models is presented. A multivariate linear model is used to describe and assess a set of ratings. Parameters are represented in terms of a factor analytic model, and maximum likelihood methods test the model parameters. Illustrative examples are presented. (SLD)
Descriptors: Comparative Analysis, Correlation, Equations (Mathematics), Estimation (Mathematics)
Wilcox, Rand R. – 1978
Two fundamental problems in mental test theory are to estimate true score and to estimate the amount of error when testing an examinee. In this report, three probability models which characterize a single test item in terms of a population of examinees are described. How these models may be modified to characterize a single examinee in terms of an…
Descriptors: Achievement Tests, Comparative Analysis, Error of Measurement, Mathematical Models
Choppin, Bruce; And Others – 1982
A detailed description of five latent structure models of achievement measurement is presented. The first project paper, by David L. McArthur, analyzes the history of mental testing to show how conventional item analysis procedures were developed, and how dissatisfaction with them has led to fragmentation. The range of distinct conceptual and…
Descriptors: Academic Achievement, Achievement Tests, Comparative Analysis, Data Analysis
Cope, Ronald T. – 1986
Comparisons were made of three Angoff Design V linear equating methods (two forms equated to a common test, two forms predicted by a common test, or two forms used to predict a common test) and Tucker's and R. Levine's linear methods, under common item linear equating with non-equivalent populations. Forms of a professional certification test…
Descriptors: Certification, Comparative Analysis, Equated Scores, Higher Education

Jansen, Margo G. H. – Journal of Educational Statistics, 1986
In this paper a Bayesian procedure is developed for the simultaneous estimation of the reading ability and difficulty parameters which are assumed to be factors in reading errors by the multiplicative Poisson Model. According to several criteria, the Bayesian estimates are better than comparable maximum likelihood estimates. (Author/JAZ)
Descriptors: Achievement Tests, Bayesian Statistics, Comparative Analysis, Difficulty Level
Phillips, Gary W. – 1982
This paper presents an introduction to the use of latent trait models for the estimation of domain scores. It was shown that these models provided an advantage over classical test theory and binomial error models in that unbiased estimates of true domain scores could be obtained even when items were not randomly selected from a universe of items.…
Descriptors: Comparative Analysis, Criterion Referenced Tests, Estimation (Mathematics), Goodness of Fit
Haladyna, Tom; Roid, Gale – 1980
The problems associated with misclassifying students when pass-fail decisions are based on test scores are discussed. One protection against misclassification is to set a confidence interval around the cutting score. Those whose scores fall above the interval are passed; those whose scores fall below the interval are failed; and those whose scores…
Descriptors: Bayesian Statistics, Classification, Comparative Analysis, Criterion Referenced Tests
Haladyna, Tom; Roid, Gale – 1976
Three approaches to the construction of achievement tests are compared: construct, operational, and empirical. The construct approach is based upon classical test theory and measures an abstract representation of the instructional objectives. The operational approach specifies instructional intent through instructional objectives, facet design,…
Descriptors: Academic Achievement, Achievement Tests, Career Development, Comparative Analysis