NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational Assessment18
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 18 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Haladyna, Thomas M.; Rodriguez, Michael C. – Educational Assessment, 2021
Full-information item analysis provides item developers and reviewers comprehensive empirical evidence of item quality, including option response frequency, point-biserial index (PBI) for distractors, mean-scores of respondents selecting each option, and option trace lines. The multi-serial index (MSI) is introduced as a more informative…
Descriptors: Test Items, Item Analysis, Reading Tests, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Sparks, Jesse R.; van Rijn, Peter W.; Deane, Paul – Educational Assessment, 2021
Effectively evaluating the credibility and accuracy of multiple sources is critical for college readiness. We developed 24 source evaluation tasks spanning four predicted difficulty levels of a hypothesized learning progression (LP) and piloted these tasks to evaluate the utility of an LP-based approach to designing formative literacy assessments.…
Descriptors: Middle School Students, Information Sources, Grade 6, Grade 7
Peer reviewed Peer reviewed
Direct linkDirect link
Becker, Anthony; Nekrasova-Beker, Tatiana – Educational Assessment, 2018
While previous research has identified numerous factors that contribute to item difficulty, studies involving large-scale reading tests have provided mixed results. This study examined five selected-response item types used to measure reading comprehension in the Pearson Test of English Academic: a) multiple-choice (choose one answer), b)…
Descriptors: Reading Comprehension, Test Items, Reading Tests, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Steedle, Jeffrey T.; Morrison, Kristin M. – Educational Assessment, 2019
Assessment items are commonly field tested prior to operational use to observe statistical item properties such as difficulty. Item parameter estimates from field testing may be used to assign scores via pre-equating or computer adaptive designs. This study examined differences between item difficulty estimates based on field test and operational…
Descriptors: Field Tests, Test Items, Statistics, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Seeratan, Kavita L.; McElhaney, Kevin W.; Mislevy, Jessica; McGhee, Raymond, Jr.; Conger, Dylan; Long, Mark C. – Educational Assessment, 2020
We describe the conceptualization, design, development, validation, and testing of a summative instrument that measures high school students' ability to analyze and evaluate data, construct scientific explanations, and formulate scientific arguments in biology and chemistry disciplinary contexts. Data from 1,405 students were analyzed to evaluate…
Descriptors: High School Students, Science Process Skills, Student Evaluation, Science Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Turkan, Sultan; Lopez, Alexis; Lawless, Rene; Tolentino, Florencia – Educational Assessment, 2019
In this article we explore the use of pictorial glossaries as an accommodation for English learners (ELs) with entry and emerging levels of English language proficiency. Drawing on survey responses from 98 middle school ELs and cognitive interviews with 10 of the survey participants, we examined the participants' preferences and experiences with…
Descriptors: Glossaries, Pictorial Stimuli, English Language Learners, Language Proficiency
Peer reviewed Peer reviewed
Direct linkDirect link
Traynor, Anne – Educational Assessment, 2017
Variation in test performance among examinees from different regions or national jurisdictions is often partially attributed to differences in the degree of content correspondence between local school or training program curricula, and the test of interest. This posited relationship between test-curriculum correspondence, or "alignment,"…
Descriptors: Test Items, Test Construction, Alignment (Education), Curriculum
Peer reviewed Peer reviewed
Direct linkDirect link
Schneider, M. Christina; Huff, Kristen L.; Egan, Karla L.; Gaines, Margie L.; Ferrara, Steve – Educational Assessment, 2013
A primary goal of standards-based statewide achievement tests is to classify students into achievement levels that enable valid inferences about student content area knowledge and skill. Explicating how knowledge and skills are expected to differ in complexity in achievement level descriptors, and how that complexity is related to empirical item…
Descriptors: Test Items, Difficulty Level, Achievement Tests, Test Interpretation
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Viger, Steven G. – Educational Assessment, 2011
An important part of test development is ensuring alignment between test forms and content standards. One common way of measuring alignment is the Webb (1997, 2007) alignment procedure. This article investigates (a) how well item writers understand components of the definition of Depth of Knowledge (DOK) from the Webb alignment procedure and (b)…
Descriptors: Test Items, Difficulty Level, Test Construction, Alignment (Education)
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Ou Lydia; Lee, Hee-Sun; Linn, Marcia C. – Educational Assessment, 2011
Both multiple-choice and constructed-response items have known advantages and disadvantages in measuring scientific inquiry. In this article we explore the function of explanation multiple-choice (EMC) items and examine how EMC items differ from traditional multiple-choice and constructed-response items in measuring scientific reasoning. A group…
Descriptors: Science Tests, Multiple Choice Tests, Responses, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Ou Lydia; Lee, Hee-Sun; Linn, Marcia C. – Educational Assessment, 2010
To improve student science achievement in the United States we need inquiry-based instruction that promotes coherent understanding and assessments that are aligned with the instruction. Instead, current textbooks often offer fragmented ideas and most assessments only tap recall of details. In this study we implemented 10 inquiry-based science…
Descriptors: Inquiry, Active Learning, Science Achievement, Science Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Abedi, Jamal – Educational Assessment, 2009
This study compared performance of both English language learners (ELLs) and non-ELL students in Grades 4 and 8 under accommodated and nonaccommodated testing conditions. The accommodations used in this study included a computerized administration of a math test with a pop-up glossary, a customized English dictionary, extra testing time, and…
Descriptors: Computer Assisted Testing, Testing Accommodations, Mathematics Tests, Grade 4
Peer reviewed Peer reviewed
Direct linkDirect link
Leighton, Jacqueline P.; Gokiert, Rebecca J. – Educational Assessment, 2008
The purpose of the present investigation was to identify the relationship among different indicators of uncertainty that lead to potential item misalignment. The item-based indicators included ratings of ambiguity and cognitive complexity. The student-based indicators included (a) frequency of cognitive monitoring per item, (b) levels of…
Descriptors: Test Items, Cognitive Processes, Item Analysis, Self Concept
Peer reviewed Peer reviewed
Direct linkDirect link
Wolf, Mikyung Kim; Leon, Seth – Educational Assessment, 2009
The purpose of the present study is to examine the language characteristics of a few states' large-scale assessments of mathematics and science and investigate whether the language demands of the items are associated with the degree of differential item functioning (DIF) for English language learner (ELL) students. A total of 542 items from 11…
Descriptors: Mathematics Tests, Science Tests, Measurement, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Martiniello, Maria – Educational Assessment, 2009
This article examines nonmathematical linguistic complexity as a source of differential item functioning (DIF) in math word problems for English language learners (ELLs). Specifically, this study investigates the relationship between item measures of linguistic complexity, nonlinguistic forms of representation and DIF measures based on item…
Descriptors: Mathematics Tests, Grade 4, Test Bias, Word Problems (Mathematics)
Previous Page | Next Page ยป
Pages: 1  |  2