Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 14 |
Since 2006 (last 20 years) | 32 |
Descriptor
Mathematics Tests | 38 |
Test Items | 18 |
Statistical Analysis | 15 |
Comparative Analysis | 12 |
Item Response Theory | 10 |
Difficulty Level | 9 |
Scores | 9 |
Item Analysis | 8 |
Models | 8 |
College Entrance Examinations | 7 |
Reading Tests | 7 |
More ▼ |
Source
ETS Research Report Series | 38 |
Author
Graf, Edith Aurora | 3 |
Lawless, René | 3 |
Attali, Yigal | 2 |
Bridgeman, Brent | 2 |
Brownstein, Beth | 2 |
Deane, Paul | 2 |
Frankel, Lois | 2 |
Futagi, Yoko | 2 |
Higgins, Derrick | 2 |
Kannan, Priya | 2 |
Liu, Jinghua | 2 |
More ▼ |
Publication Type
Journal Articles | 38 |
Reports - Research | 37 |
Tests/Questionnaires | 2 |
Collected Works - General | 1 |
Numerical/Quantitative Data | 1 |
Reports - Evaluative | 1 |
Education Level
Audience
Location
New Jersey | 4 |
Kentucky | 2 |
Massachusetts | 2 |
Australia | 1 |
Delaware | 1 |
Louisiana (New Orleans) | 1 |
Michigan | 1 |
Pennsylvania (Philadelphia) | 1 |
Texas | 1 |
Washington | 1 |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 3 |
Assessments and Surveys
What Works Clearinghouse Rating
Blair Lehman; Jesse R. Sparks; Jonathan Steinberg – ETS Research Report Series, 2024
Over the last 20 years, many methods have been proposed to use process data (e.g., response time) to detect changes in engagement during the test-taking process. However, many of these methods were developed and evaluated in highly similar testing contexts: 30 or more single-select multiple-choice items presented in a linear, fixed sequence in…
Descriptors: National Competency Tests, Secondary School Mathematics, Secondary School Students, Mathematics Tests
Guo, Hongwen; Ercikan, Kadriye – ETS Research Report Series, 2021
In this report, we demonstrate use of differential response time (DRT) methodology, an extension of differential item functioning methodology, for examining differences in how students from different backgrounds engage with assessment tasks. We analyze response time data from a digitally delivered mathematics assessment to examine timing…
Descriptors: Test Wiseness, English Language Learners, Reaction Time, Mathematics Tests
Identifying Teachers' Needs for Results from Interim Unit Assessments. Research Report. ETS RR-21-08
Kannan, Priya; Bryant, Andrew D.; Shao, Shiyi; Wylie, E. Caroline – ETS Research Report Series, 2021
Interim assessments have been defined variously in different contexts and can be used for predictive purposes or instructional purposes. In this paper, we present results from a study where we evaluated reporting needs for interim assessments designed for instructional purposes and intended to be used at the end of defined curriculum units.…
Descriptors: Educational Assessment, Student Evaluation, Units of Study, Mathematics Tests
Stone, Elizabeth; Wylie, E. Caroline – ETS Research Report Series, 2019
We describe the summative assessment component within a K-12 assessment program and our development of a validity argument to support its claims with respect to intended uses and interpretations. First, we describe the "Winsight"® assessment program theory of action, a logic model elucidating mechanisms for how use of the assessment…
Descriptors: Summative Evaluation, Educational Assessment, Test Validity, Test Use
Lopez, Alexis A.; Guzman-Orth, Danielle; Zapata-Rivera, Diego; Forsyth, Carolyn M.; Luce, Christine – ETS Research Report Series, 2021
Substantial progress has been made toward applying technology enhanced conversation-based assessments (CBAs) to measure the English-language proficiency of English learners (ELs). CBAs are conversation-based systems that use conversations among computer-animated agents and a test taker. We expanded the design and capability of prior…
Descriptors: Accuracy, English Language Learners, Language Proficiency, Language Tests
Lopez, Alexis A.; Tolentino, Florencia – ETS Research Report Series, 2020
In this study we investigated how English learners (ELs) interacted with "®" summative English language arts (ELA) and mathematics items, the embedded online tools, and accessibility features. We focused on how EL students navigated the assessment items; how they selected or constructed their responses; how they interacted with the…
Descriptors: English Language Learners, Student Evaluation, Language Arts, Summative Evaluation
Robin, Frédéric; Bejar, Isaac; Liang, Longjuan; Rijmen, Frank – ETS Research Report Series, 2016
Exploratory and confirmatory factor analyses of domestic data from the" GRE"® revised General Test, introduced in 2011, were conducted separately for the verbal (VBL) and quantitative (QNT) reasoning measures to evaluate the unidimensionality and local independence assumptions required by item response theory (IRT). Results based on data…
Descriptors: College Entrance Examinations, Graduate Study, Verbal Tests, Mathematics Tests
Fu, Jianbin – ETS Research Report Series, 2016
The multidimensional item response theory (MIRT) models with covariates proposed by Haberman and implemented in the "mirt" program provide a flexible way to analyze data based on item response theory. In this report, we discuss applications of the MIRT models with covariates to longitudinal test data to measure skill differences at the…
Descriptors: Item Response Theory, Longitudinal Studies, Test Bias, Goodness of Fit
Frankel, Lois; Brownstein, Beth; Soiffer, Neil; Hansen, Eric – ETS Research Report Series, 2016
The work described in this report is the first phase of a project to provide easy-to-use tools for authoring and rendering secondary-school algebra-level math expressions in synthesized speech that is useful for students with blindness or low vision. This report describes the initial development, software implementation, and evaluation of the…
Descriptors: Algebra, Automation, Secondary School Mathematics, Artificial Speech
Kannan, Priya; Sgammato, Adrienne – ETS Research Report Series, 2017
Logistic regression (LR)-based methods have become increasingly popular for predicting and articulating cut scores. However, the precision of predictive relationships is largely dependent on the underlying correlations between the predictor and the criterion. In two simulation studies, we evaluated the impact of varying the underlying grade-level…
Descriptors: Regression (Statistics), Cutting Scores, Prediction, Accuracy
Fife, James H.; James, Kofi; Peters, Stephanie – ETS Research Report Series, 2020
The concept of variability is central to statistics. In this research report, we review mathematics education research on variability and, based on that review and on feedback from an expert panel, propose a learning progression (LP) for variability. The structure of the proposed LP consists of 5 levels of sophistication in understanding…
Descriptors: Mathematics Education, Statistics Education, Feedback (Response), Research Reports
Lu, Ying; Yen, Wendy M. – ETS Research Report Series, 2014
This article explores the use of longitudinal regression as a tool for identifying scoring inaccuracies. Student progression patterns, as evaluated through longitudinal regressions, typically are more stable from year to year than are scale score distributions and statistics, which require representative samples to conduct credibility checks.…
Descriptors: Quality Control, Regression (Statistics), Scoring, Accuracy
Frankel, Lois; Brownstein, Beth – ETS Research Report Series, 2016
The work described in this report is the second phase of a project to provide easy-to-use tools for authoring and rendering secondary-school algebra-level math expressions in synthesized speech that is useful for students with blindness or low vision. This report describes the development and results of the second feedback study performed for our…
Descriptors: Artificial Speech, Suprasegmentals, Cues, Usability
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Kevelson, Marisol J. C. – ETS Research Report Series, 2019
This study presents estimates of Black-White, Hispanic-White, and income achievement gaps using data from two different types of reading and mathematics assessments: constructed-response assessments that were likely more cognitively demanding and state achievement tests that were likely less cognitively demanding (i.e., composed solely or largely…
Descriptors: Racial Differences, Achievement Gap, White Students, African American Students