Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 19 |
Descriptor
Source
Journal of Educational… | 32 |
Author
Kolen, Michael J. | 5 |
Moses, Tim | 3 |
von Davier, Alina A. | 3 |
Chen, Haiwen | 2 |
Harris, Deborah J. | 2 |
Holland, Paul W. | 2 |
Kim, Sooyeon | 2 |
Liu, Jinghua | 2 |
Puhan, Gautam | 2 |
Sinharay, Sandip | 2 |
Wiberg, Marie | 2 |
More ▼ |
Publication Type
Journal Articles | 30 |
Reports - Research | 20 |
Reports - Evaluative | 10 |
Information Analyses | 1 |
Education Level
Audience
Practitioners | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
General Educational… | 1 |
Iowa Tests of Educational… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Moses, Tim – Journal of Educational Measurement, 2022
One result of recent changes in testing is that previously established linking frameworks may not adequately address challenges in current linking situations. Test linking through equating, concordance, vertical scaling or battery scaling may not represent linkings for the scores of tests developed to measure constructs differently for different…
Descriptors: Measures (Individuals), Educational Assessment, Test Construction, Comparative Analysis
Kim, Hyung Jin; Brennan, Robert L.; Lee, Won-Chan – Journal of Educational Measurement, 2017
In equating, when common items are internal and scoring is conducted in terms of the number of correct items, some pairs of total scores ("X") and common-item scores ("V") can never be observed in a bivariate distribution of "X" and "V"; these pairs are called "structural zeros." This simulation…
Descriptors: Test Items, Equated Scores, Comparative Analysis, Methods
Fitzpatrick, Joseph; Skorupski, William P. – Journal of Educational Measurement, 2016
The equating performance of two internal anchor test structures--miditests and minitests--is studied for four IRT equating methods using simulated data. Originally proposed by Sinharay and Holland, miditests are anchors that have the same mean difficulty as the overall test but less variance in item difficulties. Four popular IRT equating methods…
Descriptors: Difficulty Level, Test Items, Comparative Analysis, Test Construction
Albano, Anthony D. – Journal of Educational Measurement, 2015
Research on equating with small samples has shown that methods with stronger assumptions and fewer statistical estimates can lead to decreased error in the estimated equating function. This article introduces a new approach to linear observed-score equating, one which provides flexible control over how form difficulty is assumed versus estimated…
Descriptors: Equated Scores, Sample Size, Sampling, Statistical Inference
Häggström, Jenny; Wiberg, Marie – Journal of Educational Measurement, 2014
The selection of bandwidth in kernel equating is important because it has a direct impact on the equated test scores. The aim of this article is to examine the use of double smoothing when selecting bandwidths in kernel equating and to compare double smoothing with the commonly used penalty method. This comparison was made using both an equivalent…
Descriptors: Equated Scores, Data Analysis, Comparative Analysis, Simulation
Moses, Tim – Journal of Educational Measurement, 2013
The purpose of this study was to evaluate the use of adjoined and piecewise linear approximations (APLAs) of raw equipercentile equating functions as a postsmoothing equating method. APLAs are less familiar than other postsmoothing equating methods (i.e., cubic splines), but their use has been described in historical equating practices of…
Descriptors: Equated Scores, Accuracy, Simulation, Comparative Analysis
Powers, Sonya; Kolen, Michael J. – Journal of Educational Measurement, 2014
Accurate equating results are essential when comparing examinee scores across exam forms. Previous research indicates that equating results may not be accurate when group differences are large. This study compared the equating results of frequency estimation, chained equipercentile, item response theory (IRT) true-score, and IRT observed-score…
Descriptors: Accuracy, Equated Scores, Differences, Groups
Jiang, Yanlin; von Davier, Alina A.; Chen, Haiwen – Journal of Educational Measurement, 2012
This article presents a method for evaluating equating results. Within the kernel equating framework, the percent relative error (PRE) for chained equipercentile equating was computed under the nonequivalent groups with anchor test (NEAT) design. The method was applied to two data sets to obtain the PRE, which can be used to measure equating…
Descriptors: Equated Scores, Evaluation, Error of Measurement, Computation
Chen, Haiwen – Journal of Educational Measurement, 2012
In this article, linear item response theory (IRT) observed-score equating is compared under a generalized kernel equating framework with Levine observed-score equating for nonequivalent groups with anchor test design. Interestingly, these two equating methods are closely related despite being based on different methodologies. Specifically, when…
Descriptors: Tests, Item Response Theory, Equated Scores, Statistical Analysis
Puhan, Gautam – Journal of Educational Measurement, 2012
Tucker and chained linear equatings were evaluated in two testing scenarios. In Scenario 1, referred to as rater comparability scoring and equating, the anchor-to-total correlation is often very high for the new form but moderate for the reference form. This may adversely affect the results of Tucker equating, especially if the new and reference…
Descriptors: Testing, Scoring, Equated Scores, Statistical Analysis
Wiberg, Marie; van der Linden, Wim J. – Journal of Educational Measurement, 2011
Two methods of local linear observed-score equating for use with anchor-test and single-group designs are introduced. In an empirical study, the two methods were compared with the current traditional linear methods for observed-score equating. As a criterion, the bias in the equated scores relative to true equating based on Lord's (1980)…
Descriptors: Equated Scores, Statistical Analysis, Comparative Analysis, Statistical Bias
Zu, Jiyun; Liu, Jinghua – Journal of Educational Measurement, 2010
Equating of tests composed of both discrete and passage-based multiple choice items using the nonequivalent groups with anchor test design is popular in practice. In this study, we compared the effect of discrete and passage-based anchor items on observed score equating via simulation. Results suggested that an anchor with a larger proportion of…
Descriptors: Equated Scores, Test Items, Multiple Choice Tests, Comparative Analysis
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
Using data from a large-scale exam, in this study we compared various designs for equating constructed-response (CR) tests to determine which design was most effective in producing equivalent scores across the two tests to be equated. In the context of classical equating methods, four linking designs were examined: (a) an anchor set containing…
Descriptors: Equated Scores, Responses, Tests, Measurement
Oh, Hyeonjoo; Moses, Tim – Journal of Educational Measurement, 2012
This study investigated differences between two approaches to chained equipercentile (CE) equating (one- and bi-direction CE equating) in nearly equal groups and relatively unequal groups. In one-direction CE equating, the new form is linked to the anchor in one sample of examinees and the anchor is linked to the reference form in the other…
Descriptors: Equated Scores, Statistical Analysis, Comparative Analysis, Differences
Sinharay, Sandip; Holland, Paul W. – Journal of Educational Measurement, 2010
The nonequivalent groups with anchor test (NEAT) design involves missing data that are missing by design. Three equating methods that can be used with a NEAT design are the frequency estimation equipercentile equating method, the chain equipercentile equating method, and the item-response-theory observed-score-equating method. We suggest an…
Descriptors: Equated Scores, Item Response Theory, Comparative Analysis, Evaluation