Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 2 |
Descriptor
Scores | 5 |
Test Items | 5 |
Comparative Analysis | 2 |
Criterion Referenced Tests | 2 |
Higher Education | 2 |
Item Response Theory | 2 |
Testing Problems | 2 |
Ability Grouping | 1 |
Classification | 1 |
College Admission | 1 |
College Students | 1 |
More ▼ |
Author
Hambleton, Ronald K. | 5 |
Baldwin, Peter | 1 |
Clauser, Brian E. | 1 |
Clauser, Jerome C. | 1 |
Lyren, Per-Erik | 1 |
Mazor, Kathleen M. | 1 |
Mills, Craig N. | 1 |
Rogers, H. Jane | 1 |
Publication Type
Journal Articles | 3 |
Reports - Research | 3 |
Reports - Evaluative | 2 |
Guides - Non-Classroom | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 1 |
Audience
Location
Estonia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
United States Medical… | 1 |
What Works Clearinghouse Rating
Clauser, Jerome C.; Hambleton, Ronald K.; Baldwin, Peter – Educational and Psychological Measurement, 2017
The Angoff standard setting method relies on content experts to review exam items and make judgments about the performance of the minimally proficient examinee. Unfortunately, at times content experts may have gaps in their understanding of specific exam content. These gaps are particularly likely to occur when the content domain is broad and/or…
Descriptors: Scores, Item Analysis, Classification, Decision Making
Lyren, Per-Erik; Hambleton, Ronald K. – International Journal of Testing, 2011
The equal ability distribution assumption associated with the equivalent groups equating design was investigated in the context of a selection test for admission to higher education. The purpose was to assess the consequences for the test-takers in terms of receiving improperly high or low scores compared to their peers, and to find strong…
Descriptors: Evidence, Test Items, Ability Grouping, Item Response Theory

Mazor, Kathleen M.; Hambleton, Ronald K.; Clauser, Brian E. – Applied Psychological Measurement, 1998
Studied whether matching on multiple test scores would reduce false-positive error rates compared to matching on a single number-correct score using simulation. False-positive error rates were reduced for most datasets. Findings suggest that assessing the dimensional structure of a test can be important in analysis of differential item functioning…
Descriptors: Error of Measurement, Item Bias, Scores, Test Items

Mills, Craig N.; Hambleton, Ronald K. – 1980
General guidelines exist for reporting and interpreting test scores, but there are short comings in the available technology, especially when applied to criterion-referenced tests. Concerns that have been expressed in the educational measurement literature address the uses of test scores, the manner of reporting scores, limited testing knowledge…
Descriptors: Criterion Referenced Tests, Educational Objectives, Elementary Secondary Education, Guidelines
Hambleton, Ronald K.; Rogers, H. Jane – 1986
This report was designed to respond to two major methodological shortcomings in the item bias literature: (1) misfitting test models; and (2) the use of significance tests. Specifically, the goals of the research were to describe a newly developed method known as the "plot method" for identifying potentially biased test items and to…
Descriptors: Criterion Referenced Tests, Culture Fair Tests, Difficulty Level, Estimation (Mathematics)