Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 7 |
Since 2006 (last 20 years) | 17 |
Descriptor
Difficulty Level | 18 |
Test Items | 18 |
Mathematics Tests | 9 |
Science Tests | 6 |
Grade 4 | 5 |
Grade 8 | 5 |
English (Second Language) | 4 |
Grade 7 | 4 |
Language Usage | 4 |
Middle School Students | 4 |
Multiple Choice Tests | 4 |
More ▼ |
Source
Educational Assessment | 18 |
Author
Lee, Hee-Sun | 2 |
Linn, Marcia C. | 2 |
Liu, Ou Lydia | 2 |
Abedi, Jamal | 1 |
Becker, Anthony | 1 |
Belton-Kocher, Evelyn | 1 |
Conger, Dylan | 1 |
Custer, Michael | 1 |
Deane, Paul | 1 |
Egan, Karla L. | 1 |
Ferrara, Steve | 1 |
More ▼ |
Publication Type
Journal Articles | 18 |
Reports - Research | 14 |
Reports - Evaluative | 3 |
Reports - Descriptive | 1 |
Tests/Questionnaires | 1 |
Education Level
Middle Schools | 7 |
Elementary Secondary Education | 6 |
Secondary Education | 6 |
Elementary Education | 5 |
Grade 4 | 5 |
Grade 8 | 5 |
Junior High Schools | 5 |
Grade 7 | 4 |
Intermediate Grades | 4 |
Grade 6 | 3 |
High Schools | 2 |
More ▼ |
Audience
Location
California | 2 |
Alabama | 1 |
Georgia | 1 |
Indiana | 1 |
Kansas | 1 |
Massachusetts | 1 |
Michigan | 1 |
Minnesota | 1 |
New Jersey | 1 |
Ohio | 1 |
Oregon | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 2 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Haladyna, Thomas M.; Rodriguez, Michael C. – Educational Assessment, 2021
Full-information item analysis provides item developers and reviewers comprehensive empirical evidence of item quality, including option response frequency, point-biserial index (PBI) for distractors, mean-scores of respondents selecting each option, and option trace lines. The multi-serial index (MSI) is introduced as a more informative…
Descriptors: Test Items, Item Analysis, Reading Tests, Mathematics Tests
Sparks, Jesse R.; van Rijn, Peter W.; Deane, Paul – Educational Assessment, 2021
Effectively evaluating the credibility and accuracy of multiple sources is critical for college readiness. We developed 24 source evaluation tasks spanning four predicted difficulty levels of a hypothesized learning progression (LP) and piloted these tasks to evaluate the utility of an LP-based approach to designing formative literacy assessments.…
Descriptors: Middle School Students, Information Sources, Grade 6, Grade 7
Becker, Anthony; Nekrasova-Beker, Tatiana – Educational Assessment, 2018
While previous research has identified numerous factors that contribute to item difficulty, studies involving large-scale reading tests have provided mixed results. This study examined five selected-response item types used to measure reading comprehension in the Pearson Test of English Academic: a) multiple-choice (choose one answer), b)…
Descriptors: Reading Comprehension, Test Items, Reading Tests, Test Format
Steedle, Jeffrey T.; Morrison, Kristin M. – Educational Assessment, 2019
Assessment items are commonly field tested prior to operational use to observe statistical item properties such as difficulty. Item parameter estimates from field testing may be used to assign scores via pre-equating or computer adaptive designs. This study examined differences between item difficulty estimates based on field test and operational…
Descriptors: Field Tests, Test Items, Statistics, Difficulty Level
Seeratan, Kavita L.; McElhaney, Kevin W.; Mislevy, Jessica; McGhee, Raymond, Jr.; Conger, Dylan; Long, Mark C. – Educational Assessment, 2020
We describe the conceptualization, design, development, validation, and testing of a summative instrument that measures high school students' ability to analyze and evaluate data, construct scientific explanations, and formulate scientific arguments in biology and chemistry disciplinary contexts. Data from 1,405 students were analyzed to evaluate…
Descriptors: High School Students, Science Process Skills, Student Evaluation, Science Tests
Turkan, Sultan; Lopez, Alexis; Lawless, Rene; Tolentino, Florencia – Educational Assessment, 2019
In this article we explore the use of pictorial glossaries as an accommodation for English learners (ELs) with entry and emerging levels of English language proficiency. Drawing on survey responses from 98 middle school ELs and cognitive interviews with 10 of the survey participants, we examined the participants' preferences and experiences with…
Descriptors: Glossaries, Pictorial Stimuli, English Language Learners, Language Proficiency
Traynor, Anne – Educational Assessment, 2017
Variation in test performance among examinees from different regions or national jurisdictions is often partially attributed to differences in the degree of content correspondence between local school or training program curricula, and the test of interest. This posited relationship between test-curriculum correspondence, or "alignment,"…
Descriptors: Test Items, Test Construction, Alignment (Education), Curriculum
Schneider, M. Christina; Huff, Kristen L.; Egan, Karla L.; Gaines, Margie L.; Ferrara, Steve – Educational Assessment, 2013
A primary goal of standards-based statewide achievement tests is to classify students into achievement levels that enable valid inferences about student content area knowledge and skill. Explicating how knowledge and skills are expected to differ in complexity in achievement level descriptors, and how that complexity is related to empirical item…
Descriptors: Test Items, Difficulty Level, Achievement Tests, Test Interpretation
Wyse, Adam E.; Viger, Steven G. – Educational Assessment, 2011
An important part of test development is ensuring alignment between test forms and content standards. One common way of measuring alignment is the Webb (1997, 2007) alignment procedure. This article investigates (a) how well item writers understand components of the definition of Depth of Knowledge (DOK) from the Webb alignment procedure and (b)…
Descriptors: Test Items, Difficulty Level, Test Construction, Alignment (Education)
Liu, Ou Lydia; Lee, Hee-Sun; Linn, Marcia C. – Educational Assessment, 2011
Both multiple-choice and constructed-response items have known advantages and disadvantages in measuring scientific inquiry. In this article we explore the function of explanation multiple-choice (EMC) items and examine how EMC items differ from traditional multiple-choice and constructed-response items in measuring scientific reasoning. A group…
Descriptors: Science Tests, Multiple Choice Tests, Responses, Test Items
Liu, Ou Lydia; Lee, Hee-Sun; Linn, Marcia C. – Educational Assessment, 2010
To improve student science achievement in the United States we need inquiry-based instruction that promotes coherent understanding and assessments that are aligned with the instruction. Instead, current textbooks often offer fragmented ideas and most assessments only tap recall of details. In this study we implemented 10 inquiry-based science…
Descriptors: Inquiry, Active Learning, Science Achievement, Science Instruction
Abedi, Jamal – Educational Assessment, 2009
This study compared performance of both English language learners (ELLs) and non-ELL students in Grades 4 and 8 under accommodated and nonaccommodated testing conditions. The accommodations used in this study included a computerized administration of a math test with a pop-up glossary, a customized English dictionary, extra testing time, and…
Descriptors: Computer Assisted Testing, Testing Accommodations, Mathematics Tests, Grade 4
Leighton, Jacqueline P.; Gokiert, Rebecca J. – Educational Assessment, 2008
The purpose of the present investigation was to identify the relationship among different indicators of uncertainty that lead to potential item misalignment. The item-based indicators included ratings of ambiguity and cognitive complexity. The student-based indicators included (a) frequency of cognitive monitoring per item, (b) levels of…
Descriptors: Test Items, Cognitive Processes, Item Analysis, Self Concept
Wolf, Mikyung Kim; Leon, Seth – Educational Assessment, 2009
The purpose of the present study is to examine the language characteristics of a few states' large-scale assessments of mathematics and science and investigate whether the language demands of the items are associated with the degree of differential item functioning (DIF) for English language learner (ELL) students. A total of 542 items from 11…
Descriptors: Mathematics Tests, Science Tests, Measurement, Test Bias
Martiniello, Maria – Educational Assessment, 2009
This article examines nonmathematical linguistic complexity as a source of differential item functioning (DIF) in math word problems for English language learners (ELLs). Specifically, this study investigates the relationship between item measures of linguistic complexity, nonlinguistic forms of representation and DIF measures based on item…
Descriptors: Mathematics Tests, Grade 4, Test Bias, Word Problems (Mathematics)
Previous Page | Next Page ยป
Pages: 1 | 2