NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 11 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Camilli, Gregory – Educational Research and Evaluation, 2013
In the attempt to identify or prevent unfair tests, both quantitative analyses and logical evaluation are often used. For the most part, fairness evaluation is a pragmatic attempt at determining whether procedural or substantive due process has been accorded to either a group of test takers or an individual. In both the individual and comparative…
Descriptors: Alternative Assessment, Test Bias, Test Content, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
van Rijn, P. W.; Beguin, A. A.; Verstralen, H. H. F. M. – Assessment in Education: Principles, Policy & Practice, 2012
While measurement precision is relatively easy to establish for single tests and assessments, it is much more difficult to determine for decision making with multiple tests on different subjects. This latter is the situation in the system of final examinations for secondary education in the Netherlands and is used as an example in this paper. This…
Descriptors: Secondary Education, Tests, Foreign Countries, Decision Making
Hills, John R. – 1984
The literature on item bias, i.e., the question of whether some items in tests favor one cultural group over another cultural group due to irrelevant factors, is reviewed and evaluated. All known references through 1981 are described including a large number of unpublished reports. Each method is described and the criticisms that have appeared in…
Descriptors: Evaluation Methods, Item Analysis, Racial Differences, Test Bias
Miller, M. David; Burstein, Leigh – 1981
Two studies are presented in this report. The first is titled "Empirical Studies of Multilevel Approaches to Test Development and Interpretation: Measuring Between-Group Differences in Instruction." Because of a belief that schooling does affect student achievement, researchers have questioned the empirical and measurement techniques…
Descriptors: Error Patterns, Evaluation Methods, Item Analysis, Models
Klein, Stephen P.; Kosecoff, Jacqueline P. – 1975
A procedure for in-depth analysis of a limited number of tests being considered for selection by a school, district, project, or state personnel is described. This procedure involves listing the objectives that it would be desirable to measure determining the relative importance of each of these objectives, having "judges" match test items to…
Descriptors: Correlation, Educational Objectives, Evaluation, Evaluation Criteria
Peer reviewed Peer reviewed
Bhaskar, R.; Dillard, Jesse F. – Instructional Science, 1983
Description of an objective method for assigning weights to questions on examinations includes discussions of classical test theory, knowledge organization, and how task analysis can be used to identify knowledge elements required to solve specific problems, rank them, and assign objective weights to exam questions using a Pareto distribution (7…
Descriptors: Accounting, Epistemology, Evaluation Methods, Item Analysis
van der Linden, Wim J. – Evaluation in Education: International Progress, 1982
Instructional programs organized according to modern educational technology are discussed within the purposes of criterion-referenced measurements used. The problems of criterion-referenced measurements include scoring and score interpretation, item and test analysis, and mastery testing. An overview of solutions and approaches to the problems and…
Descriptors: Criterion Referenced Tests, Educational Testing, Evaluation Methods, Item Analysis
Constable, Elizabeth; Andrich, David – 1984
In circumstances where judges are required to make ratings of performance, it is usually required to have two or more raters who are trained to agree on independent ratings of the same performance. It is suggested that such a requirement may produce the paradox of attenuation associated with item analysis, in which too high a correlation between…
Descriptors: Elementary Secondary Education, Evaluation Methods, Interrater Reliability, Interviews
Hambleton, Ronald K.; Rogers, H. Jane – 1988
Issues in preparing a review form to detect item bias in tests are discussed and the first draft of an item bias review form is presented. While stereotyping is the consistent representation of a given group in a particular light, bias is the presence of some characteristic of an item that results in differential performance of two individuals of…
Descriptors: Content Analysis, Culture Fair Tests, Ethnic Stereotypes, Evaluation Methods
Jaeger, Richard M.; Busch, John Christian – 1986
This study explores the use of the modified caution index (MCI) for identifying judges whose patterns of recommendations suggest that their judgments might be based on incomplete information, flawed reasoning, or inattention to their standard-setting tasks. It also examines the effect on test standards and passing rates when the test standards of…
Descriptors: Criterion Referenced Tests, Error of Measurement, Evaluation Methods, High Schools