Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 4 |
Descriptor
Author
Baghi, Heibatollah | 1 |
Cook, Linda L. | 1 |
Cope, Ronald T. | 1 |
Goodman, Joshua | 1 |
Haberman, Shelby | 1 |
Hills, John R. | 1 |
Kim, Sooyeon | 1 |
Lee, Yi-Hsuan | 1 |
Meyers, Jason L. | 1 |
Murphy, Stephen | 1 |
Petersen, Nancy S. | 1 |
More ▼ |
Publication Type
Speeches/Meeting Papers | 5 |
Journal Articles | 4 |
Reports - Evaluative | 4 |
Reports - Research | 4 |
Numerical/Quantitative Data | 1 |
Reports - Descriptive | 1 |
Education Level
Elementary Secondary Education | 1 |
Audience
Researchers | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Wang, Lin; Qian, Jiahe; Lee, Yi-Hsuan – ETS Research Report Series, 2013
The purpose of this study was to evaluate the combined effects of reduced equating sample size and shortened anchor test length on item response theory (IRT)-based linking and equating results. Data from two independent operational forms of a large-scale testing program were used to establish the baseline results for evaluating the results from…
Descriptors: Test Construction, Item Response Theory, Testing Programs, Simulation
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement
Meyers, Jason L.; Murphy, Stephen; Goodman, Joshua; Turhan, Ahmet – Pearson, 2012
Operational testing programs employing item response theory (IRT) applications benefit from of the property of item parameter invariance whereby item parameter estimates obtained from one sample can be applied to other samples (when the underlying assumptions are satisfied). In theory, this feature allows for applications such as computer-adaptive…
Descriptors: Equated Scores, Test Items, Test Format, Item Response Theory
Kim, Sooyeon; von Davier, Alina A.; Haberman, Shelby – ETS Research Report Series, 2006
This study addresses the sample error and linking bias that occur with small and unrepresentative samples in a non-equivalent groups anchor test (NEAT) design. We propose a linking method called the "synthetic function," which is a weighted average of the identity function (the trivial equating function for forms that are known to be…
Descriptors: Equated Scores, Sample Size, Test Items, Statistical Bias

Segall, Daniel O. – Psychometrika, 1994
An asymptotic expression for the reliability of a linearly equated test is developed using normal theory. Reliability is expressed as the product of test reliability before equating and an adjustment term that is a function of the sample sizes used to estimate the linear equating transformation. The approach is illustrated. (SLD)
Descriptors: Equated Scores, Error of Measurement, Estimation (Mathematics), Sample Size
Cope, Ronald T. – 1995
This paper deals with the problems that arise in performance assessment from the granularity that results from having a small number of tasks or prompts and raters of responses to these tasks or prompts. Two problems are discussed in detail: (1) achieving a satisfactory degree of reliability; and (2) equating or adjusting for differences of…
Descriptors: Difficulty Level, Educational Assessment, Equated Scores, High Stakes Tests
Baghi, Heibatollah; And Others – 1995
Issues related to linking tests with constructed response items were explored, specifically by comparing single-group and anchor-test designs to link raw scores from alternate forms of performance-based student assessments in the context of Delaware's assessment program using performance-based assessment. This study explored use of the two test…
Descriptors: Comparative Analysis, Constructed Response, Correlation, Educational Assessment
Hills, John R.; And Others – 1985
This study investigated the feasibility of using the three-parameter model in Florida's minimum competency testing program. LOGIST 4 was used to analyze 1984 Statewide Student Assessment Tests (SSAT)-II data, exploring possibilities that easiness of the test would cause problems in the estimation of the a and c parameters. These problems and…
Descriptors: Elementary Secondary Education, Equated Scores, Feasibility Studies, Goodness of Fit
Cook, Linda L.; Petersen, Nancy S. – 1986
This paper examines how various equating methods are affected by: (1) sampling error; (2) sample characteristics; and (3) characteristics of anchor test items. It reviews empirical studies that investigated the invariance of equating transformations, and it discusses empirical and simulation studies that focus on how the properties of anchor tests…
Descriptors: Educational Research, Equated Scores, Error of Measurement, Evaluation Methods