Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 10 |
Descriptor
Source
| Educational Testing Service | 12 |
Author
| Haberman, Shelby J. | 3 |
| Deane, Paul | 2 |
| Quinlan, Thomas | 2 |
| Sinharay, Sandip | 2 |
| Bennett, Randy Elliot | 1 |
| Breland, Hunter | 1 |
| Cahalan-Laitusis, Cara | 1 |
| Dorans, Neil J. | 1 |
| Higgins, Derrick | 1 |
| Jia, Helena | 1 |
| Kim, Hae-Jin | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 8 |
| Numerical/Quantitative Data | 3 |
| Reports - Evaluative | 3 |
| Information Analyses | 1 |
| Reports - Descriptive | 1 |
Education Level
| Elementary Secondary Education | 4 |
| Grade 8 | 2 |
| Higher Education | 2 |
| Postsecondary Education | 2 |
| Elementary Education | 1 |
| Grade 7 | 1 |
| High Schools | 1 |
| Junior High Schools | 1 |
| Middle Schools | 1 |
| Secondary Education | 1 |
Audience
Location
| Japan | 1 |
| South Korea | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Test of English as a Foreign… | 3 |
| Graduate Record Examinations | 2 |
| National Merit Scholarship… | 1 |
| Preliminary Scholastic… | 1 |
| Test of English for… | 1 |
What Works Clearinghouse Rating
Ricker-Pedley, Kathryn L. – Educational Testing Service, 2011
A pseudo-experimental study was conducted to examine the link between rater accuracy calibration performances and subsequent accuracy during operational scoring. The study asked 45 raters to score a 75-response calibration set and then a 100-response (operational) set of responses from a retired Graduate Record Examinations[R] (GRE[R]) writing…
Descriptors: Scoring, Accuracy, College Entrance Examinations, Writing Tests
Deane, Paul; Quinlan, Thomas; Kostin, Irene – Educational Testing Service, 2011
ETS has recently instituted the Cognitively Based Assessments of, for, and as Learning (CBAL) research initiative to create a new generation of assessment designed from the ground up to enhance learning. It is intended as a general approach, covering multiple subject areas including reading, writing, and math. This paper is concerned with the…
Descriptors: Automation, Scoring, Educational Assessment, Writing Tests
Use of Continuous Exponential Families to Link Forms via Anchor Tests. Research Report. ETS RR-11-11
Haberman, Shelby J.; Yan, Duanli – Educational Testing Service, 2011
Continuous exponential families are applied to linking test forms via an internal anchor. This application combines work on continuous exponential families for single-group designs and work on continuous exponential families for equivalent-group designs. Results are compared to those for kernel and equipercentile equating in the case of chained…
Descriptors: Equated Scores, Statistical Analysis, Language Tests, Mathematics Tests
Sinharay, Sandip; Haberman, Shelby J.; Jia, Helena – Educational Testing Service, 2011
Standard 3.9 of the "Standards for Educational and Psychological Testing" (American Educational Research Association, American Psychological Association, & National Council for Measurement in Education, 1999) demands evidence of model fit when an item response theory (IRT) model is used to make inferences from a data set. We applied two recently…
Descriptors: Item Response Theory, Goodness of Fit, Statistical Analysis, Language Tests
Bennett, Randy Elliot – Educational Testing Service, 2011
CBAL, an acronym for Cognitively Based Assessment of, for, and as Learning, is a research initiative intended to create a model for an innovative K-12 assessment system that provides summative information for policy makers, as well as formative information for classroom instructional purposes. This paper summarizes empirical results from 16 CBAL…
Descriptors: Educational Assessment, Elementary Secondary Education, Summative Evaluation, Formative Evaluation
Haberman, Shelby J. – Educational Testing Service, 2011
Alternative approaches are discussed for use of e-rater[R] to score the TOEFL iBT[R] Writing test. These approaches involve alternate criteria. In the 1st approach, the predicted variable is the expected rater score of the examinee's 2 essays. In the 2nd approach, the predicted variable is the expected rater score of 2 essay responses by the…
Descriptors: Writing Tests, Scoring, Essays, Language Tests
Deane, Paul – Educational Testing Service, 2011
This paper presents a socio-cognitive framework for connecting writing pedagogy and writing assessment with modern social and cognitive theories of writing. It focuses on providing a general framework that highlights the connections between writing competency and other literacy skills; identifies key connections between literacy instruction,…
Descriptors: Writing (Composition), Writing Evaluation, Writing Tests, Cognitive Ability
Powers, Donald E.; Kim, Hae-Jin; Yu, Feng; Weng, Vincent Z.; VanWinkle, Waverely – Educational Testing Service, 2009
To facilitate the interpretation of test scores from the new TOEIC[R] (Test of English for International Communications[TM]) speaking and writing tests as measures of English-language proficiency, we administered a self-assessment inventory to TOEIC examinees in Japan and Korea, to gather their perceptions of their ability to perform a variety of…
Descriptors: English for Special Purposes, Language Tests, Writing Tests, Speech Tests
Sinharay, Sandip; Dorans, Neil J.; Liang, Longjuan – Educational Testing Service, 2009
To ensure fairness, it is important to better understand the relationship of language proficiency to standard psychometric analysis procedures. This paper examines how results of differential item functioning (DIF) analysis are affected by an increase in the proportion of examinees who report that English is not their first language in the…
Descriptors: Test Bias, Language Proficiency, English (Second Language), Measurement
Quinlan, Thomas; Higgins, Derrick; Wolff, Susanne – Educational Testing Service, 2009
This report evaluates the construct coverage of the e-rater[R[ scoring engine. The matter of construct coverage depends on whether one defines writing skill, in terms of process or product. Originally, the e-rater engine consisted of a large set of components with a proven ability to predict human holistic scores. By organizing these capabilities…
Descriptors: Guides, Writing Skills, Factor Analysis, Writing Tests
Cahalan-Laitusis, Cara – Educational Testing Service, 2004
With the addition of a writing component to many high-stakes tests (e.g., college admission tests and high school exit exams), a review of writing accommodations for students with disabilities becomes critical. This paper reviews the accommodation policies of 20 states with high school exit exams of writing and three testing organizations that…
Descriptors: College Entrance Examinations, Assistive Technology, Writing Tests, Testing
Breland, Hunter; Lee, Yong-Won; Najarian, Michelle; Muraki, Eiji – Educational Testing Service, 2004
This investigation of the comparability of writing assessment prompts was conducted in two phases. In an exploratory Phase I, 47 writing prompts administered in the computer-based Test of English as a Foreign Language[TM] (TOEFL[R] CBT) from July through December 1998 were examined. Logistic regression procedures were used to estimate prompt…
Descriptors: Writing Evaluation, Quality Control, Gender Differences, Writing Tests


