NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
Elementary and Secondary…2
What Works Clearinghouse Rating
Showing 1 to 15 of 193 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Lazenby, Katherine; Tenney, Kristin; Marcroft, Tina A.; Komperda, Regis – Chemistry Education Research and Practice, 2023
Assessment instruments that generate quantitative data on attributes (cognitive, affective, behavioral, "etc.") of participants are commonly used in the chemistry education community to draw conclusions in research studies or inform practice. Recently, articles and editorials have stressed the importance of providing evidence for the…
Descriptors: Chemistry, Periodicals, Journal Articles, Science Education
Peer reviewed Peer reviewed
Direct linkDirect link
Elayne P. Colón; Lori M. Dassa; Thomas M. Dana; Nathan P. Hanson – Action in Teacher Education, 2024
To meet accreditation expectations, teacher preparation programs must demonstrate their candidates are evaluated using summative assessment tools that yield sound, reliable, and valid data. These tools are primarily used by the clinical experience team -- university supervisors and mentor teachers. Institutional beliefs regarding best practices…
Descriptors: Student Teachers, Teacher Interns, Evaluation Methods, Interrater Reliability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Holcomb, T. Scott; Lambert, Richard; Bottoms, Bryndle L. – Journal of Educational Supervision, 2022
In this study, various statistical indexes of agreement were calculated using empirical data from a group of evaluators (n = 45) of early childhood teachers. The group of evaluators rated ten fictitious teacher profiles using the North Carolina Teacher Evaluation Process (NCTEP) rubric. The exact and adjacent agreement percentages were calculated…
Descriptors: Interrater Reliability, Teacher Evaluation, Statistical Analysis, Early Childhood Teachers
Peer reviewed Peer reviewed
Direct linkDirect link
Braun, Robert; Ravn, Tine; Frankus, Elisabeth – Research Ethics, 2020
In this paper we reflect on the looming question of what constitutes expertise in ethics. Based on an empirical program that involved qualitative and quantitative as well as participatory research elements we show that expertise in research ethics and integrity is based on experience in the assessment processes. We then connect traditional…
Descriptors: Integrity, Ethics, Participatory Research, Qualitative Research
Peer reviewed Peer reviewed
Direct linkDirect link
Lambie, Glenn W.; Mullen, Patrick R.; Swank, Jacqueline M.; Blount, Ashley – Measurement and Evaluation in Counseling and Development, 2018
Supervisors evaluated counselors-in-training at multiple points during their practicum experience using the Counseling Competencies Scale (CCS; N = 1,070). The CCS evaluations were randomly split to conduct exploratory factor analysis and confirmatory factor analysis, resulting in a 2-factor model (61.5% of the variance explained).
Descriptors: Counselor Training, Counseling, Measures (Individuals), Competence
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Garcia-Garzon, Eduardo; Abad, Francisco J.; Garrido, Luis E. – Journal of Intelligence, 2019
There has been increased interest in assessing the quality and usefulness of short versions of the Raven's Progressive Matrices. A recent proposal, composed of the last twelve matrices of the Standard Progressive Matrices (SPM-LS), has been depicted as a valid measure of "g." Nonetheless, the results provided in the initial validation…
Descriptors: Intelligence Tests, Test Validity, Evaluation Methods, Undergraduate Students
Peer reviewed Peer reviewed
Direct linkDirect link
Looney, Marilyn A. – Measurement in Physical Education and Exercise Science, 2018
The purpose of this article was two-fold (1) provide an overview of the commonly reported and under-reported absolute agreement indices in the kinesiology literature for continuous data; and (2) present examples of these indices for hypothetical data along with recommendations for future use. It is recommended that three types of information be…
Descriptors: Interrater Reliability, Evaluation Methods, Kinetics, Indexes
Peer reviewed Peer reviewed
Direct linkDirect link
Margulieux, Lauren; Ketenci, Tuba Ayer; Decker, Adrienne – Computer Science Education, 2019
Background and context: The variables that researchers measure and how they measure them are central in any area of research, including computing education. Which research questions can be asked and how they are answered depends on measurement. Objective: To summarize the commonly used variables and measurements in computing education and to…
Descriptors: Measurement Techniques, Standards, Evaluation Methods, Computer Science Education
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A.; Peterson, Meghan E. – Language Testing, 2018
The use of assessments that require rater judgment (i.e., rater-mediated assessments) has become increasingly popular in high-stakes language assessments worldwide. Using a systematic literature review, the purpose of this study is to identify and explore the dominant methods for evaluating rating quality within the context of research on…
Descriptors: Language Tests, Evaluators, Evaluation Methods, Interrater Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Radley, Keith C.; Dart, Evan H.; Wright, Sarah J. – School Psychology Quarterly, 2018
Research based on single-case designs (SCD) are frequently utilized in educational settings to evaluate the effect of an intervention on student behavior. Visual analysis is the primary method of evaluation of SCD, despite research noting concerns regarding reliability of the procedure. Recent research suggests that characteristics of the graphic…
Descriptors: Graphs, Evaluation Methods, Data, Intervention
Peer reviewed Peer reviewed
Direct linkDirect link
Cousineau, Denis; Laurencelle, Louis – Educational and Psychological Measurement, 2017
Assessing global interrater agreement is difficult as most published indices are affected by the presence of mixtures of agreements and disagreements. A previously proposed method was shown to be specifically sensitive to global agreement, excluding mixtures, but also negatively biased. Here, we propose two alternatives in an attempt to find what…
Descriptors: Interrater Reliability, Evaluation Methods, Statistical Bias, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Stefanic, Nicholas; Randles, Clint – Music Education Research, 2015
The purpose of this study was to explore the reliability of measures of both individual and group creative work using the consensual assessment technique (CAT). CAT was used to measure individual and group creativity among a population of pre-service music teachers enrolled in a secondary general music class (n = 23) and was evaluated from…
Descriptors: Music Education, Creativity, Preservice Teachers, Music Teachers
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Regional Educational Laboratory West, 2020
These are the appendixes for the report, "The Association between Teachers' Use of Formative Assessment Practices and Students' Use of Self-Regulated Learning Strategies." Two appendixes are included in this document. Appendix A are the methods of the study. This includes the reliability of the teacher and student surveys and the…
Descriptors: Formative Evaluation, Learning Strategies, Elementary School Students, Elementary School Teachers
Peer reviewed Peer reviewed
Direct linkDirect link
Raczynski, Kevin; Cohen, Allan – Applied Measurement in Education, 2018
The literature on Automated Essay Scoring (AES) systems has provided useful validation frameworks for any assessment that includes AES scoring. Furthermore, evidence for the scoring fidelity of AES systems is accumulating. Yet questions remain when appraising the scoring performance of AES systems. These questions include: (a) which essays are…
Descriptors: Essay Tests, Test Scoring Machines, Test Validity, Evaluators
Christ, Theodore J.; Desjardins, Christopher David – Journal of Psychoeducational Assessment, 2018
Curriculum-Based Measurement of Oral Reading (CBM-R) is often used to monitor student progress and guide educational decisions. Ordinary least squares regression (OLSR) is the most widely used method to estimate the slope, or rate of improvement (ROI), even though published research demonstrates OLSR's lack of validity and reliability, and…
Descriptors: Bayesian Statistics, Curriculum Based Assessment, Oral Reading, Least Squares Statistics
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  12  |  13