Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 7 |
| Since 2017 (last 10 years) | 15 |
| Since 2007 (last 20 years) | 36 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 7 |
| Policymakers | 1 |
| Practitioners | 1 |
| Students | 1 |
Location
| Oregon | 2 |
| Pennsylvania | 2 |
| Turkey | 2 |
| United Kingdom (England) | 2 |
| Australia | 1 |
| Bosnia and Herzegovina | 1 |
| China | 1 |
| Croatia | 1 |
| Florida | 1 |
| Germany | 1 |
| Israel | 1 |
| More ▼ | |
Laws, Policies, & Programs
| Elementary and Secondary… | 1 |
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedRomberg, Thomas A.; Braswell, James – Journal for Research in Mathematics Education, 1973
Described is a method to assist developers of mathematics programs to make formative evaluations. The method involves periodic achievement testing via item sampling. Details of item construction, periodic testing, and profile construction are given in a discussion of how this method was applied to modify a sixth-grade mathematics program. (JP)
Descriptors: Curriculum, Curriculum Development, Evaluation, Formative Evaluation
Peer reviewedCliff, Norman; And Others – Applied Psychological Measurement, 1979
Monte Carlo research with TAILOR, a program using implied orders as a basis for tailored testing, is reported. TAILOR typically required about half the available items to estimate, for each simulated examinee, the responses on the remainder. (Author/CTM)
Descriptors: Adaptive Testing, Computer Programs, Item Sampling, Nonparametric Statistics
Peer reviewedHoste, R. – British Journal of Educational Psychology, 1981
In this paper, a proposal is made by which a content validity coefficient can be calculated. An example of the use of the coefficient is given, demonstrating that different question combinations in a CSE biology examination in which a choice of questions was given gave different levels of content validity. (Author)
Descriptors: Achievement Tests, Biology, Content Analysis, Item Sampling
Molina, Huberto; Shoemaker, David M. – 1973
The Michigan Oral Language program was installed in the Santa Paula School District for a 2-month tryout and student performance before and after the program was assessed through multiple matrix sampling. Student performance over seven content areas was determined for both the English and Spanish language components. In all seven content areas,…
Descriptors: Bilingual Education, Elementary School Students, English (Second Language), Item Sampling
PDF pending restorationHill, Richard K. – 1974
When norming tests, it may be preferable to use the matrix sampling technique. The results from the samples may be used to estimate what the distribution of scores would have been if each subject had taken all the items. This paper compares four methods for making these estimates. The sample size made it possible to compare the techniques in a…
Descriptors: Bayesian Statistics, Comparative Analysis, Data Analysis, Item Sampling
Barrows, Thomas S. – 1971
This speech discusses performance contracting as educational research, notes some evaluation problems, and proposes solutions to these problems. The term performance contracting, according to the report, denotes an administrative rather than an instructional innovation. The author observes that the understanding of instruction and learning derived…
Descriptors: Educational Experiments, Educational Innovation, Evaluation Criteria, Evaluation Methods
Peer reviewedReilly, Richard R.; Jackson, Rex – Journal of Educational Measurement, 1973
The present study suggests that although the reliability of an academic aptitude test given under formula-score condition can be increased substantially through empirical option weighting, much of the increase is due to the capitalization of the keying procedure on omitting tendencies which are reliable but not valid. (Author)
Descriptors: Aptitude Tests, Correlation, Factor Analysis, Item Sampling
Peer reviewedvan der Linden, Wim J. – Applied Psychological Measurement, 1979
The restrictions on item difficulties that must be met when binomial models are applied to domain-referenced testing are examined. Both a deterministic and a stochastic conception of item responses are discussed with respect to difficulty and Guttman-type items. (Author/BH)
Descriptors: Difficulty Level, Item Sampling, Latent Trait Theory, Mathematical Models
Revuelta, Javier – Psychometrika, 2004
Two psychometric models are presented for evaluating the difficulty of the distractors in multiple-choice items. They are based on the criterion of rising distractor selection ratios, which facilitates interpretation of the subject and item parameters. Statistical inferential tools are developed in a Bayesian framework: modal a posteriori…
Descriptors: Multiple Choice Tests, Psychometrics, Models, Difficulty Level
Whitley, Evangeline L. – 1988
The governance options matrix is provided to offer a way for state and university policymakers to examine the functioning environments of specific university-owned public teaching hospitals. With it, they can consider the benefits and problems involved with different options for governance. The issues related to the environmental factors affecting…
Descriptors: Administration, Allied Health Occupations Education, Evaluation Methods, Governance
Peer reviewedMessick, Samuel – American Psychologist, 1975
Argues that even for purposes of applied decision making, reliance upon criterion validity -- the degree to which measures correlate with specific criteria -- or content coverage is not enough, that the meaning of the measure must also be analyzed in order to evaluate responsibly the possible consequences of the proposed use, it is stated.…
Descriptors: Educational Diagnosis, Educational Objectives, Educational Programs, Item Sampling
Linn, Robert – 1978
A series of studies on conceptual and design problems in competency-based measurements are explained. The concept of validity within the context of criterion-referenced measurement is reviewed. The authors believe validation should be viewed as a process rather than an end product. It is the process of marshalling evidence to support…
Descriptors: Criterion Referenced Tests, Item Analysis, Item Sampling, Test Bias
Upp, Caroline M.; Barcikowski, Robert S. – 1981
Demands for more complete information on educational programs have emanated from national, state and local sources. Their focus is on the processes that are occurring in individual classrooms. The information that is collected to provide insight into educational programs is customarily summative in nature, answering, for example, questions…
Descriptors: Academic Achievement, Attitude Measures, Cognitive Measurement, Evaluation Methods
de Gruijter, Dato N. M. – 1980
In a situation where the population distribution of latent trait scores can be estimated, the ordinary maximum likelihood estimator of latent trait scores may be improved upon by taking the estimated population distribution into account. In this paper empirical Bayes estimators are compared with the liklihood estimator for three samples of 300…
Descriptors: Bayesian Statistics, Comparative Analysis, Goodness of Fit, Item Sampling
Cluxton, Sue Ellen; Mandeville, Garrett K. – 1979
A comparison was made between four different scoring procedures for the 45-item Reading Comprehension subtest, Level I, of the Comprehensive Test of Basic Skills, Form S, for a sample of 1,000 third grade students. These students were selected to have been among those who omitted from 3 to 22 of the 45 items. Another representative sample of 1,300…
Descriptors: Achievement Tests, Guessing (Tests), Item Sampling, Latent Trait Theory

Direct link
