Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 8 |
Since 2006 (last 20 years) | 13 |
Descriptor
Error of Measurement | 13 |
Test Items | 13 |
Grade 3 | 9 |
Elementary School Students | 8 |
Item Response Theory | 7 |
Grade 2 | 5 |
Mathematics Tests | 5 |
Difficulty Level | 4 |
Test Construction | 4 |
Goodness of Fit | 3 |
Grade 1 | 3 |
More ▼ |
Source
Author
Alonzo, Julie | 2 |
James S. Kim | 2 |
Joshua B. Gilbert | 2 |
Luke W. Miratrix | 2 |
Tindal, Gerald | 2 |
Arce, Alvaro J. | 1 |
Battista, Carmela | 1 |
Bulut, Okan | 1 |
Campana, Kayla V. | 1 |
Coverdale, Bradley J. | 1 |
Dawes, Jillian M. | 1 |
More ▼ |
Publication Type
Reports - Research | 9 |
Journal Articles | 8 |
Reports - Descriptive | 3 |
Numerical/Quantitative Data | 2 |
Reports - Evaluative | 1 |
Tests/Questionnaires | 1 |
Education Level
Grade 3 | 13 |
Elementary Education | 10 |
Early Childhood Education | 8 |
Primary Education | 8 |
Grade 2 | 6 |
Grade 1 | 4 |
Grade 4 | 4 |
Grade 5 | 4 |
Elementary Secondary Education | 3 |
Grade 6 | 2 |
Grade 7 | 2 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
Measures of Academic Progress | 1 |
What Works Clearinghouse Rating
Gilbert, Joshua B.; Kim, James S.; Miratrix, Luke W. – Journal of Educational and Behavioral Statistics, 2023
Analyses that reveal how treatment effects vary allow researchers, practitioners, and policymakers to better understand the efficacy of educational interventions. In practice, however, standard statistical methods for addressing heterogeneous treatment effects (HTE) fail to address the HTE that may exist "within" outcome measures. In…
Descriptors: Test Items, Item Response Theory, Computer Assisted Testing, Program Effectiveness
Joshua B. Gilbert; James S. Kim; Luke W. Miratrix – Annenberg Institute for School Reform at Brown University, 2024
Longitudinal models of individual growth typically emphasize between-person predictors of change but ignore how growth may vary "within" persons because each person contributes only one point at each time to the model. In contrast, modeling growth with multi-item assessments allows evaluation of how relative item performance may shift…
Descriptors: Vocabulary Development, Item Response Theory, Test Items, Student Development
Joshua B. Gilbert; James S. Kim; Luke W. Miratrix – Applied Measurement in Education, 2024
Longitudinal models typically emphasize between-person predictors of change but ignore how growth varies "within" persons because each person contributes only one data point at each time. In contrast, modeling growth with multi-item assessments allows evaluation of how relative item performance may shift over time. While traditionally…
Descriptors: Vocabulary Development, Item Response Theory, Test Items, Student Development
Precision of Single-Skill Math CBM Time-Series Data: The Effect of Probe Stratification and Set Size
Solomon, Benjamin G.; Payne, Lexy L.; Campana, Kayla V.; Marr, Erin A.; Battista, Carmela; Silva, Alex; Dawes, Jillian M. – Journal of Psychoeducational Assessment, 2020
Comparatively little research exists on single-skill math (SSM) curriculum-based measurements (CBMs) for the purpose of monitoring growth, as may be done in practice or when monitoring intervention effectiveness within group or single-case research. Therefore, we examined a common variant of SSM-CBM: 1 digit × 1 digit multiplication. Reflecting…
Descriptors: Curriculum Based Assessment, Mathematics Tests, Mathematics Skills, Multiplication
Gorgun, Guher; Bulut, Okan – Educational and Psychological Measurement, 2021
In low-stakes assessments, some students may not reach the end of the test and leave some items unanswered due to various reasons (e.g., lack of test-taking motivation, poor time management, and test speededness). Not-reached items are often treated as incorrect or not-administered in the scoring process. However, when the proportion of…
Descriptors: Scoring, Test Items, Response Style (Tests), Mathematics Tests
Li, Feifei – ETS Research Report Series, 2017
An information-correction method for testlet-based tests is introduced. This method takes advantage of both generalizability theory (GT) and item response theory (IRT). The measurement error for the examinee proficiency parameter is often underestimated when a unidimensional conditional-independence IRT model is specified for a testlet dataset. By…
Descriptors: Item Response Theory, Generalizability Theory, Tests, Error of Measurement
Li, Sylvia; Meyer, Patrick – NWEA, 2019
This simulation study examines the measurement precision, item exposure rates, and the depth of the MAP® Growth™ item pools under various grade-level restrictions. Unlike most summative assessments, MAP Growth allows examinees to see items from any grade level, regardless of the examinee's actual grade level. It does not limit the test to items…
Descriptors: Achievement Tests, Item Banks, Test Items, Instructional Program Divisions
Schoen, Robert C.; Yang, Xiaotong; Liu, Sicong; Paek, Insu – Grantee Submission, 2017
The Early Fractions Test v2.2 is a paper-pencil test designed to measure mathematics achievement of third- and fourth-grade students in the domain of fractions. The purpose, or intended use, of the Early Fractions Test v2.2 is to serve as a measure of student outcomes in a randomized trial designed to estimate the effect of an educational…
Descriptors: Psychometrics, Mathematics Tests, Mathematics Achievement, Fractions
Arce, Alvaro J.; Wang, Ze – International Journal of Testing, 2012
The traditional approach to scale modified-Angoff cut scores transfers the raw cuts to an existing raw-to-scale score conversion table. Under the traditional approach, cut scores and conversion table raw scores are not only seen as interchangeable but also as originating from a common scaling process. In this article, we propose an alternative…
Descriptors: Generalizability Theory, Item Response Theory, Cutting Scores, Scaling
Schafer, William D.; Coverdale, Bradley J.; Luxenberg, Harlan; Jin, Ying – Practical Assessment, Research & Evaluation, 2011
There are relatively few examples of quantitative approaches to quality control in educational assessment and accountability contexts. Among the several techniques that are used in other fields, Shewart charts have been found in a few instances to be applicable in educational settings. This paper describes Shewart charts and gives examples of how…
Descriptors: Charts, Quality Control, Educational Assessment, Statistical Analysis
Rogers, W. Todd; Lin, Jie; Rinaldi, Christia M. – Applied Measurement in Education, 2011
The evidence gathered in the present study supports the use of the simultaneous development of test items for different languages. The simultaneous approach used in the present study involved writing an item in one language (e.g., French) and, before moving to the development of a second item, translating the item into the second language (e.g.,…
Descriptors: Test Items, Item Analysis, Achievement Tests, French
Alonzo, Julie; Liu, Kimy; Tindal, Gerald – Behavioral Research and Teaching, 2007
In this technical report, the authors describe the development and piloting of reading comprehension measures as part of a comprehensive progress monitoring literacy assessment system developed in 2006 for use with students in Kindergarten through fifth grade. They begin with a brief overview of the two conceptual frameworks underlying the…
Descriptors: Reading Comprehension, Emergent Literacy, Test Construction, Literacy Education
Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2007
In this technical report, the authors describe the development alternate forms of three types of early literacy measures as part of a comprehensive progress monitoring literacy assessment system developed in 2006 for use with students in Kindergarten through fourth grade. They begin with a brief overview of the two conceptual frameworks underlying…
Descriptors: Emergent Literacy, Measures (Individuals), Naming, Alphabets