NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers6
Location
Arizona1
Turkey1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 23 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Atilgan, Hakan – Eurasian Journal of Educational Research, 2019
Purpose: This study intended to examine the generalizability and reliability of essay ratings within the scope of the generalizability (G) theory. Specifically, the effect of raters on the generalizability and reliability of students' essay ratings was examined. Furthermore, variations of the generalizability and reliability coefficients with…
Descriptors: Foreign Countries, Essay Tests, Test Reliability, Interrater Reliability
Michelle Herridge – ProQuest LLC, 2021
Evaluation of student written work during summative assessments is an important and critical task for instructors at all educational levels. Nevertheless, few research studies exist that provide insights into how different instructors approach this task. Chemistry faculty (FIs) and graduate student instructors (GSIs) regularly engage in the…
Descriptors: Science Instruction, Chemistry, College Faculty, Teaching Assistants
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rios, Joseph A.; Sparks, Jesse R.; Zhang, Mo; Liu, Ou Lydia – ETS Research Report Series, 2017
Proficiency with written communication (WC) is critical for success in college and careers. As a result, institutions face a growing challenge to accurately evaluate their students' writing skills to obtain data that can support demands of accreditation, accountability, or curricular improvement. Many current standardized measures, however, lack…
Descriptors: Test Construction, Test Validity, Writing Tests, College Outcomes Assessment
Scharf, Davida – ProQuest LLC, 2013
Purpose: The goal of the study was to test an intervention using a brief essay as an instrument for evaluating higher-order information literacy skills in college students, while accounting for prior conditions such as socioeconomic status and prior academic achievement, and identify other predictors of information literacy through an evaluation…
Descriptors: Information Literacy, Intervention, Student Evaluation, College Students
Peer reviewed Peer reviewed
Jackson, E. A. – European Journal of Engineering Education, 1988
Investigates the marker-marker reliability of an examination for a third-year degree course in circuit theory. Reports that the coefficient of correlation between markers fell within the range 0.94 to 0.98. (YP)
Descriptors: College Science, Engineering Education, Essay Tests, Interrater Reliability
Peer reviewed Peer reviewed
Spaulding, Cheryl L. – Journal of Reading, 1989
Reviews "Written Language Assessment" (WLA), a new standardized test to evaluate children's and adolescents' written language competence by having students write essays instead of answer multiple choice questions. Finds problems with the WLA in terms of interrater reliability. (RS)
Descriptors: Elementary Secondary Education, Essay Tests, Interrater Reliability, Standardized Tests
Weare, Jane; And Others – 1987
This annotated bibliography was developed upon noting a deficiency of information in the literature regarding the training of raters for establishing agreement. The ERIC descriptor, "Interrater Reliability", was used to locate journal articles. Some of the 33 resulting articles focus on mathematical concepts and present formulas for computing…
Descriptors: Annotated Bibliographies, Cloze Procedure, Correlation, Essay Tests
Paden, Patricia A. – 1986
Two factors which may affect the ratings assigned to an essay test are investigated: (1) context effects; and (2) score level effects. Context effects exist in essay scoring if an essay is rated higher when preceded by poor quality essays than when preceded by high quality essays. A score level effect is defined as a change in the score (value)…
Descriptors: Context Effect, Essay Tests, Holistic Evaluation, Interrater Reliability
Bunch, Michael B.; Littlefair, Wendy – 1988
A total of 2,000 essays written by 1,000 students was submitted to generalizability analyses for domain-referenced tests. Each student had written one essay on each of two prompts representing two models of discourse. Each essay was read by six readers and judged on a scale of from 1 to 4. No reader read essays from both prompts. Reader agreement…
Descriptors: Cutting Scores, Essay Tests, Generalizability Theory, Interrater Reliability
Ackerman, Terry A. – 1986
The purpose of this paper is to compare the precision of direct and indirect measures of writing assessment using the test information functions from a graded response Item Response Theory (IRT) model. Subjects were 192 sophomore English students from a parochial high school in Wisconsin. Both direct and indirect measures of writing ability were…
Descriptors: Correlation, Essay Tests, High Schools, Interrater Reliability
Mitchell, Karen J.; Anderson, Judith A. – 1987
The Association of American Medical Colleges is conducting research to develop, implement, and evaluate a Medical College Admission Test (MCAT) essay testing program. Essay administration in the spring and fall of 1985 and 1986 suggested that additional research was needed on the development of topics which elicit similar skills and meet standard…
Descriptors: College Entrance Examinations, Essay Tests, Estimation (Mathematics), Generalizability Theory
Shale, Doug – 1986
This study is an attempt at a cohesive characterization of the concept of essay reliability. As such, it takes as a basic premise that previous and current practices in reporting reliability estimates for essay tests have certain shortcomings. The study provides an analysis of these shortcomings--partly to encourage a fuller understanding of the…
Descriptors: Analysis of Variance, Correlation, Error of Measurement, Essay Tests
Peterson, Gary W. – 1983
Even though several national testing firms have developed measures to evaluate the effectiveness of baccalaureate education, there continues to be a general reluctance on the part of faculty in colleges and universities to accept these measures as criteria on which to evaluate educational programs. Some of the resistance appears to lie in the lack…
Descriptors: Bachelors Degrees, Cognitive Processes, Difficulty Level, Essay Tests
Breland, Hunter M.; And Others – 1987
Six university English departments collaborated in this examination of the differences between multiple-choice and essay tests in evaluating writing skills. The study also investigated ways the two tools can complement one another, ways to improve cost effectiveness of essay testing, and ways to integrate assessment and the educational process.…
Descriptors: Comparative Testing, Efficiency, Essay Tests, Higher Education
Breland, Hunter M. – 1983
Direct assessment of writing skill, usually considered to be synonymous with assessment by means of writing samples, is reviewed in terms of its history and with respect to evidence of its reliability and validity. Reliability is examined as it is influenced by reader inconsistency, domain sampling, and other sources of error. Validity evidence is…
Descriptors: Essay Tests, Evaluation Needs, Higher Education, Interrater Reliability
Previous Page | Next Page »
Pages: 1  |  2