Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 12 |
Since 2006 (last 20 years) | 18 |
Descriptor
Standardized Tests | 49 |
Test Format | 49 |
Test Items | 49 |
Test Construction | 18 |
Foreign Countries | 13 |
Achievement Tests | 12 |
Difficulty Level | 12 |
Higher Education | 12 |
Multiple Choice Tests | 12 |
Mathematics Tests | 9 |
Reading Tests | 9 |
More ▼ |
Source
Author
Huntley, Renee M. | 2 |
Kalogrides, Demetra | 2 |
Lawrence, Ida M. | 2 |
Podolsky, Anne | 2 |
Turhan, Ahmet | 2 |
Abedi, Jamal | 1 |
Ackerman, Terry | 1 |
Babiar, Tasha Calvert | 1 |
Bass, Lori A. | 1 |
Bayley, Robert | 1 |
Binici, Salih | 1 |
More ▼ |
Publication Type
Education Level
Location
Netherlands | 2 |
Sweden | 2 |
United Kingdom | 2 |
Australia | 1 |
Florida | 1 |
Germany | 1 |
Israel | 1 |
Japan | 1 |
Kansas | 1 |
Massachusetts | 1 |
Oregon | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Jiajing Huang – ProQuest LLC, 2022
The nonequivalent-groups anchor-test (NEAT) data-collection design is commonly used in large-scale assessments. Under this design, different test groups take different test forms. Each test form has its own unique items and all test forms share a set of common items. If item response theory (IRT) models are applied to analyze the test data, the…
Descriptors: Item Response Theory, Test Format, Test Items, Test Construction
Goran Trajkovski; Heather Hayes – Digital Education and Learning, 2025
This book explores the transformative role of artificial intelligence in educational assessment, catering to researchers, educators, administrators, policymakers, and technologists involved in shaping the future of education. It delves into the foundations of AI-assisted assessment, innovative question types and formats, data analysis techniques,…
Descriptors: Artificial Intelligence, Educational Assessment, Computer Uses in Education, Test Format
Magdalen Beiting-Parrish – ProQuest LLC, 2022
The following is a five-chapter dissertation surrounding the use of text mining techniques for better understanding the language of mathematics items from standardized tests to improve linguistic equity of these items to support assessment of English Language Learners. Introduction: The dissertation begins with an overview of the problem that…
Descriptors: Mathematics Tests, Test Items, Item Analysis, Standardized Tests
Bryant, William – Practical Assessment, Research & Evaluation, 2017
As large-scale standardized tests move from paper-based to computer-based delivery, opportunities arise for test developers to make use of items beyond traditional selected and constructed response types. Technology-enhanced items (TEIs) have the potential to provide advantages over conventional items, including broadening construct measurement,…
Descriptors: Standardized Tests, Test Items, Computer Assisted Testing, Test Format
Mason, Rihana S.; Bass, Lori A. – Early Education and Development, 2020
Research Findings Research suggests children from low-income environments have vocabularies that differ from those of their higher-income peers. They may have basic knowledge of many words of which children from higher income environments have acquired sub- or supra-ordinate knowledge. This study sought to determine if children from low-income…
Descriptors: Receptive Language, Disadvantaged Environment, Vocabulary Development, Standardized Tests
Lina Anaya; Nagore Iriberri; Pedro Rey-Biel; Gema Zamarro – Annenberg Institute for School Reform at Brown University, 2021
Standardized assessments are widely used to determine access to educational resources with important consequences for later economic outcomes in life. However, many design features of the tests themselves may lead to psychological reactions influencing performance. In particular, the level of difficulty of the earlier questions in a test may…
Descriptors: Test Construction, Test Wiseness, Test Items, Difficulty Level
Woodcock, Stuart; Howard, Steven J.; Ehrich, John – School Psychology, 2020
Standardized testing is ubiquitous in educational assessment, but questions have been raised about the extent to which these test scores accurately reflect students' genuine knowledge and skills. To more rigorously investigate this issue, the current study employed a within-subject experimental design to examine item format effects on primary…
Descriptors: Elementary School Students, Grade 3, Test Items, Test Format
Shin, Sun-Young; Lee, Senyung; Lidster, Ryan – Language Testing, 2021
In this study we investigated the potential for a shared-first-language (shared-L1) effect on second language (L2) listening test scores using differential item functioning (DIF) analyses. We did this in order to understand how accented speech may influence performance at the item level, while controlling for key variables including listening…
Descriptors: Listening Comprehension Tests, Language Tests, Native Language, Scores
Reardon, Sean F.; Kalogrides, Demetra; Fahle, Erin M.; Podolsky, Anne; Zárate, Rosalía C. – Educational Researcher, 2018
Prior research suggests that males outperform females, on average, on multiple-choice items compared to their relative performance on constructed-response items. This paper characterizes the extent to which gender achievement gaps on state accountability tests across the United States are associated with those tests' item formats. Using roughly 8…
Descriptors: Test Items, Test Format, Gender Differences, Achievement Gap
Reardon, Sean; Fahle, Erin; Kalogrides, Demetra; Podolsky, Anne; Zarate, Rosalia – Society for Research on Educational Effectiveness, 2016
Prior research demonstrates the existence of gender achievement gaps and the variation in the magnitude of these gaps across states. This paper characterizes the extent to which the variation of gender achievement gaps on standardized tests across the United States can be explained by differing state accountability test formats. A comprehensive…
Descriptors: Test Format, Gender Differences, Achievement Gap, Standardized Tests
Saß, Steffani; Schütte, Kerstin – Journal of Psychoeducational Assessment, 2016
Solving test items might require abilities in test-takers other than the construct the test was designed to assess. Item and student characteristics such as item format or reading comprehension can impact the test result. This experiment is based on cognitive theories of text and picture comprehension. It examines whether integration aids, which…
Descriptors: Reading Difficulties, Science Tests, Test Items, Visual Aids
Kevelson, Marisol J. C. – ETS Research Report Series, 2019
This study presents estimates of Black-White, Hispanic-White, and income achievement gaps using data from two different types of reading and mathematics assessments: constructed-response assessments that were likely more cognitively demanding and state achievement tests that were likely less cognitively demanding (i.e., composed solely or largely…
Descriptors: Racial Differences, Achievement Gap, White Students, African American Students
Partnership for Assessment of Readiness for College and Careers, 2015
The Partnership for Assessment of Readiness for College and Careers (PARCC) is a group of states working together to develop a modern assessment that replaces previous state standardized tests. It provides better information for teachers and parents to identify where a student needs help, or is excelling, so they are able to enhance instruction to…
Descriptors: Literacy, Language Arts, Scoring Formulas, Scoring
Keller, Lisa A.; Keller, Robert R. – Applied Measurement in Education, 2015
Equating test forms is an essential activity in standardized testing, with increased importance with the accountability systems in existence through the mandate of Adequate Yearly Progress. It is through equating that scores from different test forms become comparable, which allows for the tracking of changes in the performance of students from…
Descriptors: Item Response Theory, Rating Scales, Standardized Tests, Scoring Rubrics
Babiar, Tasha Calvert – Journal of Applied Measurement, 2011
Traditionally, women and minorities have not been fully represented in science and engineering. Numerous studies have attributed these differences to gaps in science achievement as measured by various standardized tests. Rather than describe mean group differences in science achievement across multiple cultures, this study focused on an in-depth…
Descriptors: Test Bias, Science Achievement, Standardized Tests, Grade 8