Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 1 |
Descriptor
Difficulty Level | 9 |
Test Bias | 9 |
Testing Problems | 9 |
Test Items | 6 |
Item Analysis | 4 |
Latent Trait Theory | 4 |
Higher Education | 3 |
Scores | 3 |
Black Students | 2 |
Culture Fair Tests | 2 |
Educational Assessment | 2 |
More ▼ |
Author
Baird, Jo-Anne | 1 |
Burton, Nancy W. | 1 |
Craig, Robert | 1 |
El Masri, Yasmine H. | 1 |
Graesser, Art | 1 |
Hambleton, Ronald K. | 1 |
Hill, Richard K. | 1 |
Ironson, Gail H. | 1 |
Israel, Glenn D. | 1 |
Legg, Sue M. | 1 |
Marco, Gary L. | 1 |
More ▼ |
Publication Type
Reports - Research | 7 |
Speeches/Meeting Papers | 5 |
Journal Articles | 3 |
Reports - Descriptive | 1 |
Reports - Evaluative | 1 |
Education Level
Elementary Secondary Education | 1 |
Secondary Education | 1 |
Audience
Location
California | 1 |
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
Program for International… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
El Masri, Yasmine H.; Baird, Jo-Anne; Graesser, Art – Assessment in Education: Principles, Policy & Practice, 2016
We investigate the extent to which language versions (English, French and Arabic) of the same science test are comparable in terms of item difficulty and demands. We argue that language is an inextricable part of the scientific literacy construct, be it intended or not by the examiner. This argument has considerable implications on methodologies…
Descriptors: International Assessment, Difficulty Level, Test Items, Language Variation
Plake, Barbara S.; And Others – 1983
Differential test performance by undergraduate males and females enrolled in a developmental educational psychology course (n=167) was reported on a quantitative examination as a function of item arrangement. Males were expected to perform better than females on tests whose items arranged easy to hard. Plake and Ansorge (1982) speculated this may…
Descriptors: Difficulty Level, Feedback, Higher Education, Scoring
Ironson, Gail H.; Craig, Robert – 1982
This study was designed to increase knowledge of the functioning of item bias techniques in detecting biased items. Previous studies have used computer-generated data or real data with unknown amounts of bias. The present project extends previous studies by using items that are logically generated and subjectively evaluated a priori to be biased…
Descriptors: Ability Grouping, Difficulty Level, Higher Education, Item Analysis
Burton, Nancy W.; And Others – 1976
Assessment exercises (items) in three different formats--multiple-choice with an "I don't know" (IDK) option, multiple-choice without the IDK, and open-ended--were placed at the beginning, middle and end of 45-minute assessment packages (instruments). A balanced incomplete blocks analysis of variance was computed to determine the biasing…
Descriptors: Age Differences, Difficulty Level, Educational Assessment, Guessing (Tests)

Israel, Glenn D.; Taylor, C. L. – Evaluation and Program Planning, 1990
Mail questionnaire items that are susceptible to order effects were examined using data from 168 questionnaires in a Florida Cooperative Extension Service evaluation. Order effects were found for multiple-response and attributive questions but not for single-response items. Order also interacted with question complexity, social desirability, and…
Descriptors: Adult Farmer Education, Difficulty Level, Educational Assessment, Error of Measurement
Legg, Sue M. – 1982
A case study of the Florida Teacher Certification Examination (FTCE) program was described to assist others launching the development of large scale item banks. FTCE has four subtests: Mathematics, Reading, Writing, and Professional Education. Rasch calibrated item banks have been developed for all subtests except Writing. The methods used to…
Descriptors: Cutting Scores, Difficulty Level, Field Tests, Item Analysis
Hill, Richard K. – 1979
Four problems faced by the staff of the California Assessment Program (CAP) were solved by applying Rasch scaling techniques: (1) item cultural bias in the Entry Level Test (ELT) given to all first grade pupils; (2) nonlinear regression analysis of the third grade Reading Test scores; (3) comparison of school growth from grades two to three, using…
Descriptors: Black Students, Cultural Differences, Data Analysis, Difficulty Level

Marco, Gary L. – Applied Measurement in Education, 1988
Four simulated mathematical and verbal test forms were produced by test assembly procedures proposed in legislative bills in California and New York in 1986 to minimize differences between majority and minority scores. Item response theory analyses of data for about 22,000 black and 28,000 White high-school students were conducted. (SLD)
Descriptors: Black Students, College Entrance Examinations, Comparative Analysis, Culture Fair Tests
Hambleton, Ronald K.; Rogers, H. Jane – 1986
This report was designed to respond to two major methodological shortcomings in the item bias literature: (1) misfitting test models; and (2) the use of significance tests. Specifically, the goals of the research were to describe a newly developed method known as the "plot method" for identifying potentially biased test items and to…
Descriptors: Criterion Referenced Tests, Culture Fair Tests, Difficulty Level, Estimation (Mathematics)