Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Essay Tests | 20 |
Scoring | 9 |
Higher Education | 6 |
Writing Skills | 6 |
Correlation | 5 |
Multiple Choice Tests | 5 |
Writing Evaluation | 5 |
Test Validity | 4 |
Interrater Reliability | 3 |
Scores | 3 |
Test Construction | 3 |
More ▼ |
Source
Journal of Educational… | 20 |
Author
Chase, Clinton I. | 3 |
Bridgeman, Brent | 2 |
Hughes, David C. | 2 |
Akeju, S. A. | 1 |
Benton, Stephen L. | 1 |
Biesbrock, Edieann F. | 1 |
Blok, H. | 1 |
Breland, Hunter M. | 1 |
Cross, Lawrence H. | 1 |
DeCarlo, Lawrence T. | 1 |
Eiting, Mindert H. | 1 |
More ▼ |
Publication Type
Journal Articles | 17 |
Reports - Research | 15 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 2 |
Education Level
Audience
Researchers | 2 |
Laws, Policies, & Programs
Assessments and Surveys
Test of Standard Written… | 2 |
Advanced Placement… | 1 |
Flesch Reading Ease Formula | 1 |
Graduate Record Examinations | 1 |
Metropolitan Achievement Tests | 1 |
National Teacher Examinations | 1 |
What Works Clearinghouse Rating
DeCarlo, Lawrence T.; Kim, YoungKoung; Johnson, Matthew S. – Journal of Educational Measurement, 2011
The hierarchical rater model (HRM) recognizes the hierarchical structure of data that arises when raters score constructed response items. In this approach, raters' scores are not viewed as being direct indicators of examinee proficiency but rather as indicators of essay quality; the (latent categorical) quality of an examinee's essay in turn…
Descriptors: Responses, Essay Tests, Models, Scores

Englehard, George, Jr. – Journal of Educational Measurement, 1996
A new method for evaluating rater accuracy within the context of performance assessments is described. It uses an extended Rasch measurement model, FACETS, which is illustrated with 373 benchmark papers from the Georgia High School Graduation Writing Test rated by 20 operational raters and an expert panel. (SLD)
Descriptors: Essay Tests, Evaluation Methods, Evaluators, Performance Based Assessment

Veal, L. Ramon; Biesbrock, Edieann F. – Journal of Educational Measurement, 1971
Reports the development of an experimental essay test for use with young children based on the now-obsolete STEP Essay Tests. The child's essay on an assigned topic is rated in comparison with previously rated and scaled essays which serve as performance models. (Author)
Descriptors: Essay Tests, Grade 2, Grade 3, Test Construction

Hogan, Thomas P.; Mishler, Carol – Journal of Educational Measurement, 1980
The relationship between scores on objective tests of language skills and on free writing tasks was analyzed for third and eighth graders. Correlations between scores were of the same magnitude as reported for college students. Differences in relationships between free-writing performance and objective test scores are discussed. (Author/RD)
Descriptors: Correlation, Elementary Education, Essay Tests, Language Skills

Bridgeman, Brent; Morgan, Rick; Wang, Ming-mei – Journal of Educational Measurement, 1997
Test results of 915 high school students taking a history examination with a choice of topics show that students were generally able to pick the topic on which they could get the highest score. Implications for fair scoring when topic choice is allowed are discussed. (SLD)
Descriptors: Essay Tests, High School Students, History, Performance Factors

Hales, Loyde W.; Tokar, Edward – Journal of Educational Measurement, 1975
Investigates the effect of initial blocks of either very good or poor essay question responses on the grades assigned to subsequent essay responses. (Author/DEP)
Descriptors: Adaptation Level Theory, Essay Tests, Grading, Performance

Blok, H. – Journal of Educational Measurement, 1985
Raters judged essays on two occasions making it possible to address the question of whether multiple ratings, however obtained, represent the same true scores. Multiple ratings of a given rater did represent the same true scores, but ratings of different raters did not. Reliability, validity, and invalidity coefficients were computed. (Author/DWH)
Descriptors: Analysis of Variance, Elementary Education, Essay Tests, Evaluators

Hughes, David C.; Keeling, Brian – Journal of Educational Measurement, 1984
Several studies have shown that essays receive higher marks when preceded by poor quality scripts than when preceded by good quality scripts. This study investigated the effectiveness of providing scorers with model essays to reduce the influence of context. Context effects persisted despite the scoring procedures used. (Author/EGS)
Descriptors: Context Effect, Essay Tests, Essays, High Schools

And Others; Hughes, David C. – Journal of Educational Measurement, 1980
The effect of context on the scoring of essays was examined by arranging that the scoring of the criterion essay would be preceded either by five superior essays or by five inferior essays. The contrast in essay quality had the hypothesized effect. Other effects were not significant. (CTM)
Descriptors: Essay Tests, High Schools, Holistic Evaluation, Scoring

Chase, Clinton I. – Journal of Educational Measurement, 1983
Proposition analysis was used to equate the text base of two essays with different readability levels. Easier reading essays were given higher scores than difficult reading essays. The results appear to identify another noncontent influence on essay test scores, leaving increasingly less variance for differences in content. (Author/PN)
Descriptors: Content Analysis, Difficulty Level, Essay Tests, Higher Education

Akeju, S. A. – Journal of Educational Measurement, 1972
Study was an attempt to evaluate the West African Examinations Council efforts in terms of the extent to which its marking procedures have ensured high reader reliability for the English Language Essay examination, a test which was designed to measure writing ability. (Author)
Descriptors: Essay Tests, Examiners, Foreign Countries, Multiple Choice Tests

Powers, Donald E.; Fowles, Mary E. – Journal of Educational Measurement, 1996
Approximately 300 prospective graduate students each wrote two essays for the Graduate Record Examinations in 40-minute and 60-minute time periods. Analysis revealed that performance was, on average, significantly better with the 60-minute limit. There was no interaction between self-described test-taking style (fast versus slow) and time limits.…
Descriptors: College Entrance Examinations, College Students, Essay Tests, Higher Education

van den Bergh, Huub; Eiting, Mindert H. – Journal of Educational Measurement, 1989
A method of assessing rater reliability via a design of overlapping rater teams is presented. Covariances or correlations of ratings can be analyzed with LISREL models. Models in which the rater reliabilities are congeneric, tau-equivalent, or parallel can be tested. Two examples based on essay ratings are presented. (TJH)
Descriptors: Analysis of Covariance, Computer Simulation, Correlation, Elementary Secondary Education

Benton, Stephen L.; Kiewra, Kenneth A. – Journal of Educational Measurement, 1986
This paper assessed the relationships among holistic writing ability, the Test of Standard Written English, and four tests of organizational ability. Findings showed a significant correlation between writing ability and the tests. It was concluded that tests assessing organizational strategies ought to be included in assessments of writing…
Descriptors: Correlation, Essay Tests, Higher Education, Holistic Evaluation

Chase, Clinton I. – Journal of Educational Measurement, 1986
It is hypothesized that the readers of an essay respond to a variable in terms of its context with other variables. Sex, race, reader expectation, and quality of handwriting were crossed to study their interaction effects. Results showed complex interactions of expectations, writing, and sex within race. (Author/LMO)
Descriptors: Analysis of Variance, Elementary Secondary Education, Essay Tests, Handwriting
Previous Page | Next Page ยป
Pages: 1 | 2