Descriptor
Source
Journal of Educational… | 13 |
Author
Huynh, Huynh | 2 |
Jarjoura, David | 2 |
Wainer, Howard | 2 |
Armstrong, Ronald D. | 1 |
Casteel, Jim | 1 |
Hager, Willi | 1 |
Harrison, David A. | 1 |
Holland, Paul W. | 1 |
Jansen, Margo G. H. | 1 |
Kolen, Michael J. | 1 |
Morgan, Anne | 1 |
More ▼ |
Publication Type
Journal Articles | 13 |
Reports - Research | 8 |
Reports - Evaluative | 5 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 1 |
National Assessment of… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating

Wilcox, Rand R. – Journal of Educational Statistics, 1981
Both the binomial and beta-binomial models are applied to various problems occurring in mental test theory. The paper reviews and critiques these models. The emphasis is on the extensions of the models that have been proposed in recent years, and that might not be familiar to many educators. (Author)
Descriptors: Error of Measurement, Item Analysis, Mathematical Models, Test Reliability

Woodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores

Holland, Paul W.; Thayer, Dorothy T. – Journal of Educational Statistics, 1985
Section pre-equating (SPE) equates a new test to an old test prior to the actual use of a new test by making extensive use of experimental sections of a testing instrument. SPE theory is extended to allow for practice effects on both the old and new tests. (Author/BS)
Descriptors: Equated Scores, Mathematical Models, Statistical Studies, Test Construction

Jarjoura, David – Journal of Educational Statistics, 1985
Issues regarding tolerance and confidence intervals are discussed within the context of educational measurement, and conceptual distinctions are drawn between these two types of intervals. Points are raised about the advantages of tolerance intervals when the focus is on a particular observed score rather than a particular examinee. (Author/BW)
Descriptors: Comparative Analysis, Error of Measurement, Mathematical Models, Test Interpretation

Wainer, Howard; And Others – Journal of Educational Statistics, 1985
In this paper, scores from the Department of Education's table, "State Education Statistics," are examined to see if they can be used for state-by-state comparisons to aid in the evaluation of educational policies that vary across states. (Author/LMO)
Descriptors: Educational Assessment, Educational Indicators, Multivariate Analysis, National Norms

Westermann, Rainer; Hager, Willi – Journal of Educational Statistics, 1986
The well-known problem of cumulating error probabilities is reconsidered from a general epistemological perspective, namely, the concepts of severity and of fairness of tests. It is shown that not only Type 1 but also Type 2 errors can cumulate. A new adjustment strategy is proposed and applied. (Author/JAZ)
Descriptors: Educational Research, Error of Measurement, Hypothesis Testing, Measurement Techniques

Huynh, Huynh – Journal of Educational Statistics, 1986
Under the assumptions of classical measurement theory and the condition of normality, a formula is derived for the reliability of composite scores. The formula represents an extension of the Spearman-Brown formula to the case of truncated data. (Author/JAZ)
Descriptors: Computer Simulation, Error of Measurement, Expectancy Tables, Scoring Formulas

Armstrong, Ronald D.; And Others – Journal of Educational Statistics, 1994
A network-flow model is formulated for constructing parallel tests based on classical test theory while using test reliability as the criterion. Practitioners can specify a test-difficulty distribution for values of item difficulties as well as test-composition requirements. An empirical study illustrates the reliability of generated tests. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Difficulty Level, Item Banks

Morgan, Anne; Wainer, Howard – Journal of Educational Statistics, 1980
Two estimation procedures for the Rasch Model of test analysis are reviewed in detail, particularly with respect to new developments that make the more statistically rigorous conditional maximum likelihood estimation practical for use with longish tests. (Author/JKS)
Descriptors: Error of Measurement, Latent Trait Theory, Maximum Likelihood Statistics, Psychometrics

Huynh, Huynh; Casteel, Jim – Journal of Educational Statistics, 1985
Two approaches, the minimax approach and the Rasch procedure, are described for the simultaneous determination of passing scores for subtests when the passing score for the total test is known. (Author/LMO)
Descriptors: Cutting Scores, Educational Assessment, Elementary Secondary Education, Latent Trait Theory

Jarjoura, David; Kolen, Michael J. – Journal of Educational Statistics, 1985
An equating design in which two groups of examinees from slightly different populations are administered a different test form with a subset of common items is widely used. This paper presents standard errors and a simulation that verifies the equation for large samples for an equipercentile equating procedure for this design. (Author/BS)
Descriptors: Computer Simulation, Equated Scores, Error of Measurement, Estimation (Mathematics)

Jansen, Margo G. H. – Journal of Educational Statistics, 1986
In this paper a Bayesian procedure is developed for the simultaneous estimation of the reading ability and difficulty parameters which are assumed to be factors in reading errors by the multiplicative Poisson Model. According to several criteria, the Bayesian estimates are better than comparable maximum likelihood estimates. (Author/JAZ)
Descriptors: Achievement Tests, Bayesian Statistics, Comparative Analysis, Difficulty Level

Harrison, David A. – Journal of Educational Statistics, 1986
Multidimensional item response data were created. The strength of a general factor, the number of common factors, the distribution of items loadingon common factors, and the number of items in simulated tests were manipulated. LOGIST effectively recovered both item and trait parameters in nearly all of the experimental conditions. (Author/JAZ)
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Simulation, Correlation