Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 10 |
Descriptor
Evaluators | 10 |
Grade 7 | 6 |
Interrater Reliability | 5 |
Foreign Countries | 4 |
Correlation | 3 |
Measures (Individuals) | 3 |
Middle School Students | 3 |
Scores | 3 |
Scoring Rubrics | 3 |
Test Validity | 3 |
Writing Evaluation | 3 |
More ▼ |
Source
Author
Publication Type
Journal Articles | 9 |
Reports - Research | 8 |
Reports - Evaluative | 2 |
Education Level
Grade 7 | 10 |
Middle Schools | 5 |
Elementary Education | 4 |
Grade 6 | 4 |
Grade 8 | 4 |
Junior High Schools | 3 |
Secondary Education | 3 |
Grade 5 | 2 |
Elementary Secondary Education | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
More ▼ |
Audience
Policymakers | 1 |
Teachers | 1 |
Location
Turkey | 3 |
California | 1 |
Ohio | 1 |
Texas | 1 |
Thailand | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Nazira Tursynbayeva; Umur Öç; Ismail Karakaya – International Journal of Assessment Tools in Education, 2024
This study aimed to measure the effect of rater training given to improve the peer assessment skills of secondary school students on rater behaviors using the many-facet Rasch Measurement model. The research employed a single-group pretest-posttest design. Since all raters scored all students, the analyses were carried out in a fully crossed (s x…
Descriptors: Evaluators, Training, Behavior, Peer Evaluation
Erman Aslanoglu, Aslihan; Sata, Mehmet – Participatory Educational Research, 2021
When students present writing tasks that require higher order thinking skills to work, one of the most important problems is scoring these writing tasks objectively. The fact that raters give scores below or above their performance based on several environmental factors affects the consistency of the measurements. Inconsistencies in scoring…
Descriptors: Interrater Reliability, Evaluators, Error of Measurement, Writing Evaluation
Raczynski, Kevin; Cohen, Allan – Applied Measurement in Education, 2018
The literature on Automated Essay Scoring (AES) systems has provided useful validation frameworks for any assessment that includes AES scoring. Furthermore, evidence for the scoring fidelity of AES systems is accumulating. Yet questions remain when appraising the scoring performance of AES systems. These questions include: (a) which essays are…
Descriptors: Essay Tests, Test Scoring Machines, Test Validity, Evaluators
Gargani, John; Strong, Michael – Journal of Teacher Education, 2014
Teacher observations have become a national education phenomenon, mandated by federal policies and promoted by philanthropists. They are crucial components of teacher evaluation systems that often have high stakes for teachers and school systems, but have sparked little innovation. Recent calls to make teacher evaluations better, faster, and…
Descriptors: Teacher Evaluation, Observation, Evidence, Measures (Individuals)
Reed, Deborah K.; Sturges, Keith M. – Remedial and Special Education, 2013
Researchers have expressed concern about "implementation" fidelity in intervention research but have not extended that concern to "assessment" fidelity, or the extent to which pre-/posttests are administered and interpreted as intended. When studying reading interventions, data gathering heavily influences the identification of…
Descriptors: Reading Tests, Fidelity, Pretests Posttests, Intervention
Barth, Amy E.; Stuebing, Karla K.; Fletcher, Jack M.; Cirino, Paul T.; Romain, Melissa; Francis, David; Vaughn, Sharon – Reading Psychology, 2012
We evaluated the reliability and validity of two oral reading fluency scores for 1-minute equated passages: median score and mean score. These scores were calculated from measures of reading fluency administered up to five times over the school year to students in grades six to eight (n = 1,317). Both scores were highly reliable with strong…
Descriptors: Reading Fluency, Test Validity, Test Reliability, Scores
Della-Piana, Gabriel Mario; Gardner, Michael – Journal of MultiDisciplinary Evaluation, 2011
Background: Professional standards for validity of achievement tests have long reflected a consensus that validity is the degree to which evidence and theory support interpretations of test scores entailed by the intended uses of tests. Yet there are convincing lines of evidence that the standards are not adequately followed in practice, that…
Descriptors: Achievement Tests, Test Validity, Scores, Standards
Sondergeld, Toni A.; Beltyukova, Svetlana A.; Fox, Christine M.; Stone, Gregory E. – Mid-Western Educational Researcher, 2012
Scientifically based research used to inform evidence based school reform efforts has been required by the federal government in order to receive grant funding since the reenactment of No Child Left Behind (2002). Educational evaluators are thus faced with the challenge to use rigorous research designs to establish causal relationships. However,…
Descriptors: Research Design, Research Tools, Simulation, Educational Research
Beyreli, Latif; Ari, Gokhan – Educational Sciences: Theory and Practice, 2009
In this study, the purpose was to determine whether there was concordance among raters in the assessment of the writing performance using analytic rubric; furthermore, factors affecting the assessment process were examined. The analytic rubric used in the study consists of three sections and ten properties: External structure (format, spelling and…
Descriptors: Sentences, Scoring Rubrics, Program Effectiveness, Writing Evaluation
Corcoran, Thomas B. – Consortium for Policy Research in Education, 2008
This is the first report on the evaluation of the Inquiry Based Science and Technology Education Program (IN-STEP), an innovative and ambitious science education initiative for lower secondary schools being undertaken by a public-private partnership in Thailand funded by MSD-Thailand, an affiliate of Merck & Co. IN-STEP is a public-private…
Descriptors: Feedback (Response), Schools of Education, Evaluators, Science Interests