NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational Assessment45
Audience
What Works Clearinghouse Rating
Showing 1 to 15 of 45 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Haladyna, Thomas M.; Rodriguez, Michael C. – Educational Assessment, 2021
Full-information item analysis provides item developers and reviewers comprehensive empirical evidence of item quality, including option response frequency, point-biserial index (PBI) for distractors, mean-scores of respondents selecting each option, and option trace lines. The multi-serial index (MSI) is introduced as a more informative…
Descriptors: Test Items, Item Analysis, Reading Tests, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Im, Sukkeun; Lee, Jay – Educational Assessment, 2021
This study investigated test-taking engagement on the Spring 2019 administration of a large-scale state summative assessment. Through the identification of rapid-guessing behavior -- which is a validated indicator of disengagement -- the percentage of Grade 8 test events with meaningful amounts of rapid guessing was 5.5% in mathematics, 6.7% in…
Descriptors: Accountability, Test Results, Guessing (Tests), Summative Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Katherine T. Rhodes; Julie A. Washington; Sibylla Leon Guerrero – Educational Assessment, 2024
Little is known about mismatches between the language of mathematics testing instruments and the rich linguistic repertoires that African American children develop at home and in the community. The current study aims to provide a proof of concept and novel explanatory item response design that uses error analysis to investigate the relationship…
Descriptors: Black Dialects, African American Students, Language Usage, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wolf, Mikyung Kim; Yoo, Hanwook; Guzman-Orth, Danielle; Abedi, Jamal – Educational Assessment, 2022
Implementing a randomized controlled trial design, the present study investigated the effects of two types of accommodations, linguistic modification and a glossary, for English learners (ELs) taking a computer-based mathematics assessment. Process data including response time and clicks on glossary words were also examined to better interpret…
Descriptors: Testing Accommodations, English Language Learners, Computer Assisted Testing, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Russell, Michael; Szendey, Olivia; Kaplan, Larry – Educational Assessment, 2021
Differential Item Function (DIF) analysis is commonly employed to examine potential bias produced by a test item. Since its introduction DIF analyses have focused on potential bias related to broad categories of oppression, including gender, racial stratification, economic class, and ableness. More recently, efforts to examine the effects of…
Descriptors: Test Bias, Achievement Tests, Individual Characteristics, Disadvantaged
Peer reviewed Peer reviewed
Direct linkDirect link
Aydin, Utkun; Birgili, Bengi – Educational Assessment, 2023
Internationally, mathematics education reform has been directed toward characterizing educational goals that go beyond topic/content/skill descriptions and develop students' problem solving. The Revised Bloom's Taxonomy and MATH (Mathematical Assessment Task Hierarchy) Taxonomy characterize such goals. University entrance examinations have been…
Descriptors: Critical Thinking, Thinking Skills, Skill Development, Mathematics Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Kuhfeld, Megan R.; Cronin, John – Educational Assessment, 2022
The arrival of the COVID-19 pandemic had a profound effect on K-12 education. Most schools transitioned to remote instruction, and some used remote testing to assess student learning. Remote testing, however, is less controlled than in-school testing, leading to concerns regarding test-taking engagement. This study compared the disengagement of…
Descriptors: Computer Assisted Testing, COVID-19, Pandemics, Learner Engagement
Peer reviewed Peer reviewed
Direct linkDirect link
Yiyun Fan; Kristin L. K. Koskey; Dara Bright; Gabriel Matney; Jonathan Bostic; Toni A. May; Gregory E. Stone – Educational Assessment, 2024
Advancement of testing of mathematical problem-solving skills calls for open-ended, realistic tasks particularly susceptible to bias, compromising the score validity and fairness of tests. Informed by universal design principles, this study framed 360 prototype items developed for the "Problem-solving Measures Grades 6-8 Computer Adaptive…
Descriptors: Access to Education, Mathematics Education, Problem Solving, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Moon, Jung Aa; Keehner, Madeleine; Katz, Irvin R. – Educational Assessment, 2020
We investigated how item formats influence test takers' response tendencies under uncertainty. Adult participants solved content-equivalent math items in three formats: multiple-selection multiple-choice, grid with forced-choice (true-false) options, and grid with non-forced-choice options. Participants showed a greater tendency to commit (rather…
Descriptors: College Students, Test Wiseness, Test Format, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Russell, Michael; Moncaleano, Sebastian – Educational Assessment, 2019
Over the past decade, large-scale testing programs have employed technology-enhanced items (TEI) to improve the fidelity with which an item measures a targeted construct. This paper presents findings from a review of released TEIs employed by large-scale testing programs worldwide. Analyses examine the prevalence with which different types of TEIs…
Descriptors: Computer Assisted Testing, Fidelity, Elementary Secondary Education, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Walker, A. Adrienne; Jennings, Jeremy Kyle; Engelhard, George, Jr. – Educational Assessment, 2018
Individual person fit analyses provide important information regarding the validity of test score inferences for an "individual" test taker. In this study, we use data from an undergraduate statistics test (N = 1135) to illustrate a two-step method that researchers and practitioners can use to examine individual person fit. First, person…
Descriptors: Test Items, Test Validity, Scores, Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Ercikan, Kadriye; Guo, Hongwen; He, Qiwei – Educational Assessment, 2020
Comparing group is one of the key uses of large-scale assessment results, which are used to gain insights to inform policy and practice and to examine the comparability of scores and score meaning. Such comparisons typically focus on examinees' final answers and responses to test questions, ignoring response process differences groups may engage…
Descriptors: Data Use, Responses, Comparative Analysis, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Buono, Stephanie; Jang, Eunice Eunhee – Educational Assessment, 2021
Increasing linguistic diversity in classrooms has led researchers to examine the validity and fairness of standardized achievement tests, specifically concerning whether test score interpretations are free of bias and score use is fair for all students. This study examined whether mathematics achievement test items that contain complex language…
Descriptors: English Language Learners, Standardized Tests, Achievement Tests, Culture Fair Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Ralston, Nicole C.; Li, Min; Taylor, Catherine – Educational Assessment, 2018
Elementary school students often exhibit a variety of conceptions associated with algebraic thinking that their teachers fail to recognize or understand. It is crucial that elementary school teachers possess knowledge of the variety of student conceptions and also have abilities to address varying states of conceptions. Otherwise, students who are…
Descriptors: Elementary School Students, Student Evaluation, Mathematics Tests, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Steedle, Jeffrey T.; Morrison, Kristin M. – Educational Assessment, 2019
Assessment items are commonly field tested prior to operational use to observe statistical item properties such as difficulty. Item parameter estimates from field testing may be used to assign scores via pre-equating or computer adaptive designs. This study examined differences between item difficulty estimates based on field test and operational…
Descriptors: Field Tests, Test Items, Statistics, Difficulty Level
Previous Page | Next Page »
Pages: 1  |  2  |  3