Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 2 |
Descriptor
Models | 4 |
Test Items | 4 |
Computer Assisted Testing | 2 |
Item Response Theory | 2 |
Psychometrics | 2 |
Scoring | 2 |
Testing | 2 |
Adaptive Testing | 1 |
Bias | 1 |
Computation | 1 |
Computer Software Evaluation | 1 |
More ▼ |
Source
Educational Testing Service | 4 |
Author
Davey, Tim | 1 |
DeCarlo, Lawrence T. | 1 |
Herbert, Erin | 1 |
Rizavi, Saba | 1 |
Way, Walter D. | 1 |
von Davier, Alina A. | 1 |
von Davier, Matthias | 1 |
Publication Type
Reports - Research | 2 |
Guides - General | 1 |
Reports - Descriptive | 1 |
Reports - Evaluative | 1 |
Education Level
Adult Education | 1 |
Audience
Practitioners | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Educational Testing Service, 2011
Choosing whether to test via computer is the most difficult and consequential decision the designers of a testing program can make. The decision is difficult because of the wide range of choices available. Designers can choose where and how often the test is made available, how the test items look and function, how those items are combined into…
Descriptors: Test Items, Testing Programs, Testing, Computer Assisted Testing
DeCarlo, Lawrence T. – Educational Testing Service, 2010
A basic consideration in large-scale assessments that use constructed response (CR) items, such as essays, is how to allocate the essays to the raters that score them. Designs that are used in practice are incomplete, in that each essay is scored by only a subset of the raters, and also unbalanced, in that the number of essays scored by each rater…
Descriptors: Test Items, Responses, Essay Tests, Scoring
von Davier, Matthias; von Davier, Alina A. – Educational Testing Service, 2004
This paper examines item response theory (IRT) scale transformations and IRT scale linking methods used in the Non-Equivalent Groups with Anchor Test (NEAT) design to equate two tests, X and Y. It proposes a unifying approach to the commonly used IRT linking methods: mean-mean, mean-var linking, concurrent calibration, Stocking and Lord and…
Descriptors: Measures (Individuals), Item Response Theory, Item Analysis, Models
Rizavi, Saba; Way, Walter D.; Davey, Tim; Herbert, Erin – Educational Testing Service, 2004
Item parameter estimates vary for a variety of reasons, including estimation error, characteristics of the examinee samples, and context effects (e.g., item location effects, section location effects, etc.). Although we expect variation based on theory, there is reason to believe that observed variation in item parameter estimates exceeds what…
Descriptors: Adaptive Testing, Test Items, Computation, Context Effect