Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 3 |
Descriptor
Language Tests | 4 |
Test Items | 4 |
English (Second Language) | 2 |
Mathematics Tests | 2 |
Bias | 1 |
Coding | 1 |
College Students | 1 |
Computer Assisted Testing | 1 |
Correlation | 1 |
Cutting Scores | 1 |
Difficulty Level | 1 |
More ▼ |
Source
Educational Testing Service | 4 |
Author
DeCarlo, Lawrence T. | 1 |
Kim, Sooyeon | 1 |
Kostin, Irene | 1 |
Nissan, Susan | 1 |
Sawaki, Yasuyo | 1 |
Walker, Michael E. | 1 |
Publication Type
Reports - Research | 3 |
Information Analyses | 1 |
Reports - Evaluative | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
California | 1 |
Canada | 1 |
Connecticut | 1 |
Georgia | 1 |
Indiana | 1 |
Iowa | 1 |
Michigan | 1 |
Wisconsin | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Test of English as a Foreign… | 2 |
What Works Clearinghouse Rating
DeCarlo, Lawrence T. – Educational Testing Service, 2010
A basic consideration in large-scale assessments that use constructed response (CR) items, such as essays, is how to allocate the essays to the raters that score them. Designs that are used in practice are incomplete, in that each essay is scored by only a subset of the raters, and also unbalanced, in that the number of essays scored by each rater…
Descriptors: Test Items, Responses, Essay Tests, Scoring
Kim, Sooyeon; Walker, Michael E. – Educational Testing Service, 2011
This study examines the use of subpopulation invariance indices to evaluate the appropriateness of using a multiple-choice (MC) item anchor in mixed-format tests, which include both MC and constructed-response (CR) items. Linking functions were derived in the nonequivalent groups with anchor test (NEAT) design using an MC-only anchor set for 4…
Descriptors: Test Format, Multiple Choice Tests, Test Items, Gender Differences
Sawaki, Yasuyo; Nissan, Susan – Educational Testing Service, 2009
The study investigated the criterion-related validity of the "Test of English as a Foreign Language"[TM] Internet-based test (TOEFL[R] iBT) Listening section by examining its relationship to a criterion measure designed to reflect language-use tasks that university students encounter in everyday academic life: listening to academic…
Descriptors: Test Validity, Language Tests, English (Second Language), Computer Assisted Testing
Kostin, Irene – Educational Testing Service, 2004
The purpose of this study is to explore the relationship between a set of item characteristics and the difficulty of TOEFL[R] dialogue items. Identifying characteristics that are related to item difficulty has the potential to improve the efficiency of the item-writing process The study employed 365 TOEFL dialogue items, which were coded on 49…
Descriptors: Statistical Analysis, Difficulty Level, Language Tests, English (Second Language)