Publication Date
In 2025 | 0 |
Since 2024 | 6 |
Since 2021 (last 5 years) | 9 |
Since 2016 (last 10 years) | 15 |
Since 2006 (last 20 years) | 39 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
Researchers | 7 |
Policymakers | 1 |
Practitioners | 1 |
Students | 1 |
Location
Oregon | 2 |
Pennsylvania | 2 |
Turkey | 2 |
United Kingdom (England) | 2 |
Australia | 1 |
Bosnia and Herzegovina | 1 |
China | 1 |
Croatia | 1 |
Florida | 1 |
Germany | 1 |
Israel | 1 |
More ▼ |
Laws, Policies, & Programs
Elementary and Secondary… | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating

Brandenburg, Dale C.; Forsyth, Robert A. – Journal of Educational and Psychological Measurement, 1974
Descriptors: Achievement Tests, Comparative Analysis, Item Sampling, Mathematical Models

Romberg, Thomas A.; Braswell, James – Journal for Research in Mathematics Education, 1973
Described is a method to assist developers of mathematics programs to make formative evaluations. The method involves periodic achievement testing via item sampling. Details of item construction, periodic testing, and profile construction are given in a discussion of how this method was applied to modify a sixth-grade mathematics program. (JP)
Descriptors: Curriculum, Curriculum Development, Evaluation, Formative Evaluation

Cliff, Norman; And Others – Applied Psychological Measurement, 1979
Monte Carlo research with TAILOR, a program using implied orders as a basis for tailored testing, is reported. TAILOR typically required about half the available items to estimate, for each simulated examinee, the responses on the remainder. (Author/CTM)
Descriptors: Adaptive Testing, Computer Programs, Item Sampling, Nonparametric Statistics

Hoste, R. – British Journal of Educational Psychology, 1981
In this paper, a proposal is made by which a content validity coefficient can be calculated. An example of the use of the coefficient is given, demonstrating that different question combinations in a CSE biology examination in which a choice of questions was given gave different levels of content validity. (Author)
Descriptors: Achievement Tests, Biology, Content Analysis, Item Sampling
Graham, John W.; Taylor, Bonnie J.; Olchowski, Allison E.; Cumsille, Patricio E. – Psychological Methods, 2006
The authors describe 2 efficiency (planned missing data) designs for measurement: the 3-form design and the 2-method measurement design. The 3-form design, a kind of matrix sampling, allows researchers to leverage limited resources to collect data for 33% more survey questions than can be answered by any 1 respondent. Power tables for estimating…
Descriptors: Cost Effectiveness, Structural Equation Models, Psychological Studies, Data Collection
Molina, Huberto; Shoemaker, David M. – 1973
The Michigan Oral Language program was installed in the Santa Paula School District for a 2-month tryout and student performance before and after the program was assessed through multiple matrix sampling. Student performance over seven content areas was determined for both the English and Spanish language components. In all seven content areas,…
Descriptors: Bilingual Education, Elementary School Students, English (Second Language), Item Sampling

Hill, Richard K. – 1974
When norming tests, it may be preferable to use the matrix sampling technique. The results from the samples may be used to estimate what the distribution of scores would have been if each subject had taken all the items. This paper compares four methods for making these estimates. The sample size made it possible to compare the techniques in a…
Descriptors: Bayesian Statistics, Comparative Analysis, Data Analysis, Item Sampling
Barrows, Thomas S. – 1971
This speech discusses performance contracting as educational research, notes some evaluation problems, and proposes solutions to these problems. The term performance contracting, according to the report, denotes an administrative rather than an instructional innovation. The author observes that the understanding of instruction and learning derived…
Descriptors: Educational Experiments, Educational Innovation, Evaluation Criteria, Evaluation Methods

Reilly, Richard R.; Jackson, Rex – Journal of Educational Measurement, 1973
The present study suggests that although the reliability of an academic aptitude test given under formula-score condition can be increased substantially through empirical option weighting, much of the increase is due to the capitalization of the keying procedure on omitting tendencies which are reliable but not valid. (Author)
Descriptors: Aptitude Tests, Correlation, Factor Analysis, Item Sampling

van der Linden, Wim J. – Applied Psychological Measurement, 1979
The restrictions on item difficulties that must be met when binomial models are applied to domain-referenced testing are examined. Both a deterministic and a stochastic conception of item responses are discussed with respect to difficulty and Guttman-type items. (Author/BH)
Descriptors: Difficulty Level, Item Sampling, Latent Trait Theory, Mathematical Models
Revuelta, Javier – Psychometrika, 2004
Two psychometric models are presented for evaluating the difficulty of the distractors in multiple-choice items. They are based on the criterion of rising distractor selection ratios, which facilitates interpretation of the subject and item parameters. Statistical inferential tools are developed in a Bayesian framework: modal a posteriori…
Descriptors: Multiple Choice Tests, Psychometrics, Models, Difficulty Level
Whitley, Evangeline L. – 1988
The governance options matrix is provided to offer a way for state and university policymakers to examine the functioning environments of specific university-owned public teaching hospitals. With it, they can consider the benefits and problems involved with different options for governance. The issues related to the environmental factors affecting…
Descriptors: Administration, Allied Health Occupations Education, Evaluation Methods, Governance

Messick, Samuel – American Psychologist, 1975
Argues that even for purposes of applied decision making, reliance upon criterion validity -- the degree to which measures correlate with specific criteria -- or content coverage is not enough, that the meaning of the measure must also be analyzed in order to evaluate responsibly the possible consequences of the proposed use, it is stated.…
Descriptors: Educational Diagnosis, Educational Objectives, Educational Programs, Item Sampling
Linn, Robert – 1978
A series of studies on conceptual and design problems in competency-based measurements are explained. The concept of validity within the context of criterion-referenced measurement is reviewed. The authors believe validation should be viewed as a process rather than an end product. It is the process of marshalling evidence to support…
Descriptors: Criterion Referenced Tests, Item Analysis, Item Sampling, Test Bias
Upp, Caroline M.; Barcikowski, Robert S. – 1981
Demands for more complete information on educational programs have emanated from national, state and local sources. Their focus is on the processes that are occurring in individual classrooms. The information that is collected to provide insight into educational programs is customarily summative in nature, answering, for example, questions…
Descriptors: Academic Achievement, Attitude Measures, Cognitive Measurement, Evaluation Methods