NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 9 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
El Masri, Yasmine H.; Baird, Jo-Anne; Graesser, Art – Assessment in Education: Principles, Policy & Practice, 2016
We investigate the extent to which language versions (English, French and Arabic) of the same science test are comparable in terms of item difficulty and demands. We argue that language is an inextricable part of the scientific literacy construct, be it intended or not by the examiner. This argument has considerable implications on methodologies…
Descriptors: International Assessment, Difficulty Level, Test Items, Language Variation
Plake, Barbara S.; And Others – 1983
Differential test performance by undergraduate males and females enrolled in a developmental educational psychology course (n=167) was reported on a quantitative examination as a function of item arrangement. Males were expected to perform better than females on tests whose items arranged easy to hard. Plake and Ansorge (1982) speculated this may…
Descriptors: Difficulty Level, Feedback, Higher Education, Scoring
Ironson, Gail H.; Craig, Robert – 1982
This study was designed to increase knowledge of the functioning of item bias techniques in detecting biased items. Previous studies have used computer-generated data or real data with unknown amounts of bias. The present project extends previous studies by using items that are logically generated and subjectively evaluated a priori to be biased…
Descriptors: Ability Grouping, Difficulty Level, Higher Education, Item Analysis
Burton, Nancy W.; And Others – 1976
Assessment exercises (items) in three different formats--multiple-choice with an "I don't know" (IDK) option, multiple-choice without the IDK, and open-ended--were placed at the beginning, middle and end of 45-minute assessment packages (instruments). A balanced incomplete blocks analysis of variance was computed to determine the biasing…
Descriptors: Age Differences, Difficulty Level, Educational Assessment, Guessing (Tests)
Peer reviewed Peer reviewed
Israel, Glenn D.; Taylor, C. L. – Evaluation and Program Planning, 1990
Mail questionnaire items that are susceptible to order effects were examined using data from 168 questionnaires in a Florida Cooperative Extension Service evaluation. Order effects were found for multiple-response and attributive questions but not for single-response items. Order also interacted with question complexity, social desirability, and…
Descriptors: Adult Farmer Education, Difficulty Level, Educational Assessment, Error of Measurement
Legg, Sue M. – 1982
A case study of the Florida Teacher Certification Examination (FTCE) program was described to assist others launching the development of large scale item banks. FTCE has four subtests: Mathematics, Reading, Writing, and Professional Education. Rasch calibrated item banks have been developed for all subtests except Writing. The methods used to…
Descriptors: Cutting Scores, Difficulty Level, Field Tests, Item Analysis
Hill, Richard K. – 1979
Four problems faced by the staff of the California Assessment Program (CAP) were solved by applying Rasch scaling techniques: (1) item cultural bias in the Entry Level Test (ELT) given to all first grade pupils; (2) nonlinear regression analysis of the third grade Reading Test scores; (3) comparison of school growth from grades two to three, using…
Descriptors: Black Students, Cultural Differences, Data Analysis, Difficulty Level
Peer reviewed Peer reviewed
Marco, Gary L. – Applied Measurement in Education, 1988
Four simulated mathematical and verbal test forms were produced by test assembly procedures proposed in legislative bills in California and New York in 1986 to minimize differences between majority and minority scores. Item response theory analyses of data for about 22,000 black and 28,000 White high-school students were conducted. (SLD)
Descriptors: Black Students, College Entrance Examinations, Comparative Analysis, Culture Fair Tests
Hambleton, Ronald K.; Rogers, H. Jane – 1986
This report was designed to respond to two major methodological shortcomings in the item bias literature: (1) misfitting test models; and (2) the use of significance tests. Specifically, the goals of the research were to describe a newly developed method known as the "plot method" for identifying potentially biased test items and to…
Descriptors: Criterion Referenced Tests, Culture Fair Tests, Difficulty Level, Estimation (Mathematics)