Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 3 |
Descriptor
Test Bias | 14 |
Test Items | 10 |
Latent Trait Theory | 8 |
Test Construction | 6 |
Item Analysis | 4 |
Statistical Analysis | 4 |
Testing Problems | 4 |
Comparative Analysis | 3 |
Computer Simulation | 3 |
Equated Scores | 3 |
Evaluation Methods | 3 |
More ▼ |
Author
Hambleton, Ronald K. | 14 |
Rogers, H. Jane | 7 |
Wells, Craig S. | 2 |
Baldwin, Su | 1 |
Dogan, Nuri | 1 |
Jirka, Stephen | 1 |
Karatonis, Ana | 1 |
Kirkpatrick, Robert | 1 |
Meng, Yu | 1 |
Murphy, Edward | 1 |
Sireci, Stephen G. | 1 |
More ▼ |
Publication Type
Reports - Research | 7 |
Journal Articles | 6 |
Reports - Evaluative | 5 |
Speeches/Meeting Papers | 5 |
Information Analyses | 2 |
Opinion Papers | 1 |
Reference Materials -… | 1 |
Reports - General | 1 |
Tests/Questionnaires | 1 |
Education Level
Grade 8 | 2 |
Elementary Education | 1 |
Elementary Secondary Education | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Secondary Education | 1 |
Audience
Researchers | 2 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
The Comparison of Differential Item Functioning Predicted through Experts and Statistical Techniques
Dogan, Nuri; Hambleton, Ronald K.; Yurtcu, Meltem; Yavuz, Sinan – Cypriot Journal of Educational Sciences, 2018
Validity is one of the psychometric properties of the achievement tests. To determine the validity, one of the examination is item bias studies, which are based on differential item functioning (DIF) analyses and field experts' opinion. In this study, field experts were asked to estimate the DIF levels of the items to compare the estimations…
Descriptors: Test Bias, Comparative Analysis, Predictor Variables, Statistical Analysis
Wells, Craig S.; Hambleton, Ronald K.; Kirkpatrick, Robert; Meng, Yu – Applied Measurement in Education, 2014
The purpose of the present study was to develop and evaluate two procedures flagging consequential item parameter drift (IPD) in an operational testing program. The first procedure was based on flagging items that exhibit a meaningful magnitude of IPD using a critical value that was defined to represent barely tolerable IPD. The second procedure…
Descriptors: Test Items, Test Bias, Equated Scores, Item Response Theory
Wells, Craig S.; Baldwin, Su; Hambleton, Ronald K.; Sireci, Stephen G.; Karatonis, Ana; Jirka, Stephen – Applied Measurement in Education, 2009
Score equity assessment is an important analysis to ensure inferences drawn from test scores are comparable across subgroups of examinees. The purpose of the present evaluation was to assess the extent to which the Grade 8 NAEP Math and Reading assessments for 2005 were equivalent across selected states. More specifically, the present study…
Descriptors: National Competency Tests, Test Bias, Equated Scores, Grade 8

Hambleton, Ronald K.; Rogers, H. Jane – Applied Measurement in Education, 1989
Item Response Theory and Mantel-Haenszel approaches for investigating differential item performance were compared to assess the level of agreement of the approaches in identifying potentially biased items. Subjects were 2,000 White and 2,000 Native American high school students. The Mantel-Haenszel method provides an acceptable approximation of…
Descriptors: American Indians, Comparative Testing, High School Students, High Schools
Hambleton, Ronald K.; Rogers, H. Jane – 1988
The agreement between item response theory-based and Mantel Haenszel (MH) methods in identifying biased items on tests was studied. Data came from item responses of four spaced samples of 1,000 examinees each--two samples of 1,000 Anglo-American and two samples of 1,000 Native American students taking the New Mexico High School Proficiency…
Descriptors: Comparative Analysis, High School Students, High Schools, Item Analysis
Rogers, H. Jane; Hambleton, Ronald K. – 1987
Although item bias statistics are widely recommended for use in test development and test analysis work, problems arise in their interpretation. The purpose of the present research was to evaluate the validity of logistic test models and computer simulation methods for providing a frame of reference for item bias statistic interpretations.…
Descriptors: Computer Simulation, Evaluation Methods, Item Analysis, Latent Trait Theory
Hambleton, Ronald K.; Swaminathan, H. – 1985
Comments are made on the review papers presented by six Dutch psychometricians: Ivo Molenaar, Wim van der Linden, Ed Roskam, Arnold Van den Wollenberg, Gideon Mellenbergh, and Dato de Gruijter. Molenaar has embraced a pragmatic viewpoint on Bayesian methods, using both empirical and pure approaches to solve educational research problems. Molenaar…
Descriptors: Bayesian Statistics, Decision Making, Elementary Secondary Education, Foreign Countries
Hambleton, Ronald K. – 1989
A brief overview of item response theory is provided, and a 186-item bibliography of books and articles on the subject dating from 1953 to June 1989 is presented. The overview includes a definition of the theory, a discussion of its development and application, and comparisons with classical test theory. All publications in the bibliography were…
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Software, Equated Scores
Rogers, H. Jane; Hambleton, Ronald K. – 1987
Though item bias statistics are widely recommended for use in test development and analysis, problems arise in their interpretation. This research evaluates logistic test models and computer simulation methods for providing a frame of reference for interpreting item bias statistics. Specifically, the intent was to produce simulated sampling…
Descriptors: Computer Simulation, Cutting Scores, Grade 9, Latent Trait Theory

Hambleton, Ronald K.; Murphy, Edward – Applied Measurement in Education, 1992
The validity of several criticisms of objective tests is addressed, and the viability of some alternatives to objective testing is discussed. Evidence against multiple-choice tests is not as strong as has been claimed. Authentic assessments may not always be better, and research about new forms of assessment is necessary. (SLD)
Descriptors: Achievement Tests, Educational Assessment, Literature Reviews, Measurement Techniques
Hambleton, Ronald K.; Rogers, H. Jane – 1988
Issues in preparing a review form to detect item bias in tests are discussed and the first draft of an item bias review form is presented. While stereotyping is the consistent representation of a given group in a particular light, bias is the presence of some characteristic of an item that results in differential performance of two individuals of…
Descriptors: Content Analysis, Culture Fair Tests, Ethnic Stereotypes, Evaluation Methods

Rogers, H. Jane; Hambleton, Ronald K. – Educational and Psychological Measurement, 1989
The validity of logistic test models and computer simulation methods for generating sampling distributions of item bias statistics was evaluated under the hypothesis of no item bias. Test data from 937 ninth-grade students were used to develop 7 steps for applying computer-simulated baseline statistics in test development. (SLD)
Descriptors: Computer Simulation, Educational Research, Evaluation Methods, Grade 9
Hambleton, Ronald K.; And Others – 1977
Latent trait theory supposes that, in testing situations, examinee performance on a test can be predicted (or explained) by defining examinee characteristics, referred to as traits, estimating scores for examinees on these traits and using the scores to predict or explain test performance (Lord and Novick, 1968). In view of the breakthroughs in…
Descriptors: Adaptive Testing, Bayesian Statistics, Cognitive Measurement, Computer Programs
Hambleton, Ronald K.; Rogers, H. Jane – 1986
This report was designed to respond to two major methodological shortcomings in the item bias literature: (1) misfitting test models; and (2) the use of significance tests. Specifically, the goals of the research were to describe a newly developed method known as the "plot method" for identifying potentially biased test items and to…
Descriptors: Criterion Referenced Tests, Culture Fair Tests, Difficulty Level, Estimation (Mathematics)