Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 13 |
Descriptor
Error of Measurement | 13 |
Equated Scores | 9 |
Comparative Analysis | 6 |
Computation | 6 |
Statistical Analysis | 6 |
Statistical Bias | 5 |
Accuracy | 4 |
Scores | 4 |
Correlation | 3 |
Regression (Statistics) | 3 |
Reliability | 3 |
More ▼ |
Source
ETS Research Report Series | 7 |
Journal of Educational… | 3 |
Educational Measurement:… | 1 |
Educational Testing Service | 1 |
Journal of Educational and… | 1 |
Author
Moses, Tim | 13 |
Holland, Paul | 2 |
Kim, Sooyeon | 2 |
Grant, Mary | 1 |
Kim, YoungKoung | 1 |
Liu, Jinghua | 1 |
McHale, Fred | 1 |
Oh, Hyeonjoo | 1 |
Puhan, Gautam | 1 |
Yoo, Hanwook Henry | 1 |
Publication Type
Journal Articles | 12 |
Reports - Research | 10 |
Reports - Evaluative | 3 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Moses, Tim; Kim, YoungKoung – Journal of Educational Measurement, 2017
The focus of this article is on scale score transformations that can be used to stabilize conditional standard errors of measurement (CSEMs). Three transformations for stabilizing the estimated CSEMs are reviewed, including the traditional arcsine transformation, a recently developed general variance stabilization transformation, and a new method…
Descriptors: Error of Measurement, Scores, Comparative Analysis, Item Response Theory
Kim, Sooyeon; Moses, Tim; Yoo, Hanwook Henry – ETS Research Report Series, 2015
The purpose of this inquiry was to investigate the effectiveness of item response theory (IRT) proficiency estimators in terms of estimation bias and error under multistage testing (MST). We chose a 2-stage MST design in which 1 adaptation to the examinees' ability levels takes place. It includes 4 modules (1 at Stage 1, 3 at Stage 2) and 3 paths…
Descriptors: Item Response Theory, Computation, Statistical Bias, Error of Measurement
Moses, Tim – Educational Measurement: Issues and Practice, 2014
This module describes and extends X-to-Y regression measures that have been proposed for use in the assessment of X-to-Y scaling and equating results. Measures are developed that are similar to those based on prediction error in regression analyses but that are directly suited to interests in scaling and equating evaluations. The regression and…
Descriptors: Scaling, Regression (Statistics), Equated Scores, Comparative Analysis
Moses, Tim – ETS Research Report Series, 2013
The purpose of this report is to review ETS psychometric contributions that focus on test scores. Two major sections review contributions based on assessing test scores' measurement characteristics and other contributions about using test scores as predictors in correlational and regression relationships. An additional section reviews additional…
Descriptors: Psychometrics, Scores, Correlation, Regression (Statistics)
Moses, Tim – Journal of Educational Measurement, 2012
The focus of this paper is assessing the impact of measurement errors on the prediction error of an observed-score regression. Measures are presented and described for decomposing the linear regression's prediction error variance into parts attributable to the true score variance and the error variances of the dependent variable and the predictor…
Descriptors: Error of Measurement, Prediction, Regression (Statistics), True Scores
Moses, Tim; Liu, Jinghua – Educational Testing Service, 2011
In equating research and practice, equating functions that are smooth are typically assumed to be more accurate than equating functions with irregularities. This assumption presumes that population test score distributions are relatively smooth. In this study, two examples were used to reconsider common beliefs about smoothing and equating. The…
Descriptors: Equated Scores, Data Analysis, Scores, Methods
Oh, Hyeonjoo; Moses, Tim – Journal of Educational Measurement, 2012
This study investigated differences between two approaches to chained equipercentile (CE) equating (one- and bi-direction CE equating) in nearly equal groups and relatively unequal groups. In one-direction CE equating, the new form is linked to the anchor in one sample of examinees and the anchor is linked to the reference form in the other…
Descriptors: Equated Scores, Statistical Analysis, Comparative Analysis, Differences
Moses, Tim; Holland, Paul – ETS Research Report Series, 2008
The purpose of this paper is to extend von Davier, Holland, and Thayer's (2004b) framework of kernel equating so that it can incorporate raw data and traditional equipercentile equating methods. One result of this more general framework is that previous equating methodology research can be viewed more comprehensively. Another result is that the…
Descriptors: Equated Scores, Error of Measurement, Statistical Analysis, Computation
Moses, Tim – Journal of Educational and Behavioral Statistics, 2008
Equating functions are supposed to be population invariant, meaning that the choice of subpopulation used to compute the equating function should not matter. The extent to which equating functions are population invariant is typically assessed in terms of practical difference criteria that do not account for equating functions' sampling…
Descriptors: Equated Scores, Error of Measurement, Sampling, Evaluation Methods
Puhan, Gautam; Moses, Tim; Grant, Mary; McHale, Fred – ETS Research Report Series, 2008
A single group (SG) equating design with nearly equivalent test forms (SiGNET) design was developed by Grant (2006) to equate small volume tests. The basis of this design is that examinees take two largely overlapping test forms within a single administration. The scored items for the operational form are divided into mini-tests called testlets.…
Descriptors: Data Collection, Equated Scores, Item Sampling, Sample Size
Moses, Tim; Holland, Paul – ETS Research Report Series, 2007
The purpose of this study was to empirically evaluate the impact of loglinear presmoothing accuracy on equating bias and variability across chained and post-stratification equating methods, kernel and percentile-rank continuization methods, and sample sizes. The results of evaluating presmoothing on equating accuracy generally agreed with those of…
Descriptors: Equated Scores, Statistical Analysis, Accuracy, Sample Size
Moses, Tim; Kim, Sooyeon – ETS Research Report Series, 2007
This study evaluated the impact of unequal reliability on test equating methods in the nonequivalent groups with anchor test (NEAT) design. Classical true score-based models were compared in terms of their assumptions about how reliability impacts test scores. These models were related to treatment of population ability differences by different…
Descriptors: Reliability, Equated Scores, Test Items, Statistical Analysis
Moses, Tim – ETS Research Report Series, 2006
Population invariance is an important requirement of test equating. An equating function is said to be population invariant when the choice of (sub)population used to compute the equating function does not matter. In recent studies, the extent to which equating functions are population invariant is typically addressed in terms of practical…
Descriptors: Equated Scores, Computation, Error of Measurement, Statistical Analysis