NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 14 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Andrés Christiansen; Rianne Janssen – Educational Assessment, Evaluation and Accountability, 2024
In international large-scale assessments, students may not be compelled to answer every test item: a student can decide to skip a seemingly difficult item or may drop out before the end of the test is reached. The way these missing responses are treated will affect the estimation of the item difficulty and student ability, and ultimately affect…
Descriptors: Test Items, Item Response Theory, Grade 4, International Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ozyeter, Neslihan Tugce – International Journal of Assessment Tools in Education, 2022
In education, examining students' learning in detail, determining their strengths and weaknesses and giving effective feedback have gained importance over time. The aim of this study is to determine the distribution of students' answers to the reading comprehension achievement test items which were written at different cognitive levels and to…
Descriptors: Student Evaluation, Feedback (Response), Scoring Rubrics, Reading Comprehension
Peer reviewed Peer reviewed
Direct linkDirect link
Cohen, Dale J.; Zhang, Jin; Wothke, Werner – Applied Measurement in Education, 2019
Construct-irrelevant cognitive complexity of some items in the statewide grade-level assessments may impose performance barriers for students with disabilities who are ineligible for alternate assessments based on alternate achievement standards. This has spurred research into whether items can be modified to reduce complexity without affecting…
Descriptors: Test Items, Accessibility (for Disabled), Students with Disabilities, Low Achievement
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kevelson, Marisol J. C. – ETS Research Report Series, 2019
This study presents estimates of Black-White, Hispanic-White, and income achievement gaps using data from two different types of reading and mathematics assessments: constructed-response assessments that were likely more cognitively demanding and state achievement tests that were likely less cognitively demanding (i.e., composed solely or largely…
Descriptors: Racial Differences, Achievement Gap, White Students, African American Students
Peer reviewed Peer reviewed
Direct linkDirect link
Palane, Nelladee McLeod; Howie, Sarah – Perspectives in Education, 2019
In this article, preProgress in Reading Literacy Study (prePIRLS) 2011 data is used to compare the performance of different language of instruction groupings (English, Afrikaans and African languages) in primary schools on the more complex, higher-order reading comprehension items tested in a large-scale international test. PrePIRLS 2011…
Descriptors: Reading Comprehension, Language of Instruction, Models, Elementary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Okumura, Taichi – Educational and Psychological Measurement, 2014
This study examined the empirical differences between the tendency to omit items and reading ability by applying tree-based item response (IRTree) models to the Japanese data of the Programme for International Student Assessment (PISA) held in 2009. For this purpose, existing IRTree models were expanded to contain predictors and to handle…
Descriptors: Foreign Countries, Item Response Theory, Test Items, Reading Ability
Yuan, Kun; Le, Vi-Nhuan – RAND Corporation, 2014
In 2010, the William and Flora Hewlett Foundation's Education Program has established the Deeper Learning Initiative, which focuses on students' development of deeper learning skills (i.e., the mastery of core academic content, critical-thinking, problem-solving, collaboration, communication, and "learn-how-to-learn" skills). Two test…
Descriptors: Test Items, Cognitive Processes, Difficulty Level, Skill Development
Peer reviewed Peer reviewed
Direct linkDirect link
Hein, Serge F.; Skaggs, Gary E. – Applied Measurement in Education, 2009
Only a small number of qualitative studies have investigated panelists' experiences during standard-setting activities or the thought processes associated with panelists' actions. This qualitative study involved an examination of the experiences of 11 panelists who participated in a prior, one-day standard-setting meeting in which either the…
Descriptors: Focus Groups, Standard Setting, Cutting Scores, Cognitive Processes
Peer reviewed Peer reviewed
Homan, Susan; And Others – Journal of Educational Measurement, 1994
A study was conducted with 782 elementary school students to determine whether the Homan-Hewitt Readability Formula could identify the readability of a single-sentence test item. Results indicate that a relationship exists between students' reading grade levels and responses to test items written at higher readability levels. (SLD)
Descriptors: Difficulty Level, Elementary Education, Elementary School Students, Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Wheldall, Kevin; Madelaine, Alison – Australasian Journal of Special Education, 2006
The aim of this study was to develop a means of tracking the reading performance of low-progress readers on a weekly basis, so as to inform instructional decision-making. A representative sample of 261 primary school children from Years 1 to 5 were tested on 21 different text passages taken from a developing passage reading test, the Wheldall…
Descriptors: Reading Tests, Test Reliability, Test Validity, Difficulty Level
Brutten, Sheila R.; And Others – 1988
A study attempted to estimate the instructional sensitivity of items in three reading comprehension tests in English as a second language (ESL). Instructional sensitivity is a test-item construct defined as the tendency for a test item to vary in difficulty as a function of instruction. Similar tasks were given to readers at different proficiency…
Descriptors: College Students, Comparative Analysis, Difficulty Level, English (Second Language)
Davis, John N.; Moriel, James, Jr. – 1980
In order to compose equivalent certification tests in the basic skills in reading, mathematics calculations and applications, and writing for high school graduation requirements, the San Jose Unified School District, California, used the Rasch computer program to arrive at the scaled difficulty level of test items. The procedure for achieving a…
Descriptors: Basic Skills, Criterion Referenced Tests, Cutting Scores, Difficulty Level
Hill, Richard K. – 1979
Four problems faced by the staff of the California Assessment Program (CAP) were solved by applying Rasch scaling techniques: (1) item cultural bias in the Entry Level Test (ELT) given to all first grade pupils; (2) nonlinear regression analysis of the third grade Reading Test scores; (3) comparison of school growth from grades two to three, using…
Descriptors: Black Students, Cultural Differences, Data Analysis, Difficulty Level
PDF pending restoration PDF pending restoration
Bauer, Ernest A.; And Others – 1979
The reading portion of the Michigan Educational Assessment Program (MEAP) was equated to the reading comprehension subtest of the Comprehensive Tests of Basic Skills (CTBS) using the Rasch Model. Both tests were administered to 366 low achieving fourth grade students. MEAP was treated as both a 95-item test and a 19-item (number of objectives…
Descriptors: Academic Standards, Criterion Referenced Tests, Difficulty Level, Educational Objectives