Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 7 |
Descriptor
Source
ETS Research Report Series | 2 |
Cambridge Assessment | 1 |
College Board | 1 |
Ministerial Council on… | 1 |
Online Submission | 1 |
Pearson | 1 |
Author
Allalouf, Avi | 1 |
Bramley, Tom | 1 |
Brennan, Robert L. | 1 |
Donovan, Jenny | 1 |
Eignor, Daniel R. | 1 |
Goodman, Joshua | 1 |
Hedges, Larry V. | 1 |
Hendrickson, Amy B. | 1 |
Holland, Paul | 1 |
Hutton, Penny | 1 |
Kim, Dong-In | 1 |
More ▼ |
Publication Type
Numerical/Quantitative Data | 13 |
Reports - Research | 11 |
Speeches/Meeting Papers | 5 |
Journal Articles | 2 |
Reports - Evaluative | 2 |
Education Level
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Secondary Education | 2 |
Grade 6 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Location
Australia | 1 |
United Kingdom (England) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Comprehensive Tests of Basic… | 1 |
Law School Admission Test | 1 |
Medical College Admission Test | 1 |
National Assessment of… | 1 |
Program for International… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Tomkowicz, Joanna; Kim, Dong-In; Wan, Ping – Online Submission, 2022
In this study we evaluated the stability of item parameters and student scores, using the pre-equated (pre-pandemic) parameters from Spring 2019 and post-equated (post-pandemic) parameters from Spring 2021 in two calibration and equating designs related to item parameter treatment: re-estimating all anchor parameters (Design 1) and holding the…
Descriptors: Equated Scores, Test Items, Evaluation Methods, Pandemics
Bramley, Tom – Cambridge Assessment, 2018
The aim of the research reported here was to get some idea of the accuracy of grade boundaries (cut-scores) obtained by applying the 'similar items method' described in Bramley & Wilson (2016). In this method experts identify items on the current version of a test that are sufficiently similar to items on previous versions for them to be…
Descriptors: Accuracy, Cutting Scores, Test Items, Item Analysis
Lee, Eunjung; Lee, Won-Chan; Brennan, Robert L. – College Board, 2012
In almost all high-stakes testing programs, test equating is necessary to ensure that test scores across multiple test administrations are equivalent and can be used interchangeably. Test equating becomes even more challenging in mixed-format tests, such as Advanced Placement Program® (AP®) Exams, that contain both multiple-choice and constructed…
Descriptors: Test Construction, Test Interpretation, Test Norms, Test Reliability
Livingston, Samuel A.; Kim, Sooyeon – ETS Research Report Series, 2010
A series of resampling studies investigated the accuracy of equating by four different methods in a random groups equating design with samples of 400, 200, 100, and 50 test takers taking each form. Six pairs of forms were constructed. Each pair was constructed by assigning items from an existing test taken by 9,000 or more test takers. The…
Descriptors: Equated Scores, Accuracy, Sample Size, Sampling
Meyers, Jason L.; Murphy, Stephen; Goodman, Joshua; Turhan, Ahmet – Pearson, 2012
Operational testing programs employing item response theory (IRT) applications benefit from of the property of item parameter invariance whereby item parameter estimates obtained from one sample can be applied to other samples (when the underlying assumptions are satisfied). In theory, this feature allows for applications such as computer-adaptive…
Descriptors: Equated Scores, Test Items, Test Format, Item Response Theory
Hendrickson, Amy B.; Kolen, Michael J. – 2001
This study compared various equating models and procedures for a sample of data from the Medical College Admission Test(MCAT), considering how item response theory (IRT) equating results compare with classical equipercentile results and how the results based on use of various IRT models, observed score versus true score, direct versus linked…
Descriptors: Equated Scores, Higher Education, Item Response Theory, Models
Sinharay, Sandip; Holland, Paul – ETS Research Report Series, 2006
It is a widely held belief that anchor tests should be miniature versions (i.e., minitests), with respect to content and statistical characteristics of the tests being equated. This paper examines the foundations for this belief. It examines the requirement of statistical representativeness of anchor tests that are content representative. The…
Descriptors: Test Items, Equated Scores, Evaluation Methods, Difficulty Level
Reese, Lynda M.; Pashley, Peter J. – 1999
This study investigated the practical effects of local item dependence (LID) on item response theory (IRT) true-score equating. A scenario was defined that emulated the Law School Admission Test (LSAT) preequating model, and data were generated to assess the impact of different degrees of LID on final equating outcomes. An extreme amount of LID…
Descriptors: College Entrance Examinations, Equated Scores, Item Response Theory, Law Schools
Rapp, Joel; Allalouf, Avi – 2002
This study examined the cross-lingual equating process adopted by a large scale testing system in which target language (TL) forms are equated to the source language (SL) forms using a set of translated items. The focus was on evaluating the degree of error inherent in the routine cross-lingual equating of the Verbal Reasoning subtest of the…
Descriptors: College Applicants, College Entrance Examinations, Equated Scores, High Stakes Tests
Hedges, Larry V.; Vevea, Jack L. – 1997
This study investigates the amount of uncertainty added to National Assessment of Educational Progress (NAEP) estimates by equating error under both ideal and less than ideal circumstances. Data from past administrations are used to guide simulations of various equating designs and error due to equating is estimated empirically. The design…
Descriptors: Ability, Elementary Secondary Education, Equated Scores, Error of Measurement
de la Torre, Jimmy; Patz, Richard J. – 2001
This paper seeks to extend the application of Markov chain Monte Carlo (MCMC) methods in item response theory (IRT) to include the estimation of equating relationships along with the estimation of test item parameters. A method is proposed that incorporates estimation of the equating relationship in the item calibration phase. Item parameters from…
Descriptors: Achievement Tests, Bayesian Statistics, Equated Scores, Estimation (Mathematics)
Eignor, Daniel R. – 1985
The feasibility of pre-equating, or establishing conversions from raw to scaled scores through the use of pretest data before operationally administering a test, was investigated for the Scholastic Aptitude Test (SAT). Item-response theory based equating methods were used to estimate item parameters on SAT pretest data, instead of using final form…
Descriptors: College Entrance Examinations, Equated Scores, Estimation (Mathematics), Feasibility Studies
Wu, Margaret; Donovan, Jenny; Hutton, Penny; Lennon, Melissa – Ministerial Council on Education, Employment, Training and Youth Affairs (NJ1), 2008
In July 2001, the Ministerial Council on Education, Employment, Training and Youth Affairs (MCEETYA) agreed to the development of assessment instruments and key performance measures for reporting on student skills, knowledge and understandings in primary science. It directed the newly established Performance Measurement and Reporting Taskforce…
Descriptors: Foreign Countries, Scientific Literacy, Science Achievement, Comparative Analysis