Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 8 |
Since 2016 (last 10 years) | 16 |
Since 2006 (last 20 years) | 53 |
Descriptor
Evaluation Methods | 83 |
Generalizability Theory | 83 |
Reliability | 26 |
Test Reliability | 17 |
Interrater Reliability | 15 |
Student Evaluation | 14 |
Performance Based Assessment | 13 |
Test Construction | 12 |
Test Validity | 12 |
Scores | 11 |
Higher Education | 10 |
More ▼ |
Source
Author
Publication Type
Education Level
Audience
Practitioners | 1 |
Researchers | 1 |
Location
California | 2 |
United Kingdom | 2 |
Asia | 1 |
Canada | 1 |
Finland (Helsinki) | 1 |
Michigan | 1 |
Oklahoma | 1 |
Tennessee | 1 |
Turkey | 1 |
Laws, Policies, & Programs
Individuals with Disabilities… | 1 |
Assessments and Surveys
Behavior Assessment System… | 1 |
Eysenck Personality Inventory | 1 |
Myers Briggs Type Indicator | 1 |
Teacher Rating Scale | 1 |
What Works Clearinghouse Rating
Mandviwalla, Munir; Schuff, David; Miller, Laurel; Chacko, Manoj – IEEE Transactions on Learning Technologies, 2023
In this article, we develop and evaluate a novel system and computing platform to structure, measure, and improve student development using points. We define student development broadly as the achievement of learning to do, know, live together, and be. The system leverages individual agency, social influences, content generation and sharing,…
Descriptors: Student Development, Academic Achievement, Systems Approach, Design
Daniel McNeish – Grantee Submission, 2023
Factor analysis is often used to model scales created to measure latent constructs, and internal structure validity evidence is commonly assessed with indices like SRMR, RMSEA, and CFI. These indices are essentially effect size measures and definitive benchmarks regarding which values connote reasonable fit have been elusive. Simulations from the…
Descriptors: Models, Testing, Indexes, Factor Analysis
Bonifay, Wes – Grantee Submission, 2022
Traditional statistical model evaluation typically relies on goodness-of-fit testing and quantifying model complexity by counting parameters. Both of these practices may result in overfitting and have thereby contributed to the generalizability crisis. The information-theoretic principle of minimum description length addresses both of these…
Descriptors: Statistical Analysis, Models, Goodness of Fit, Evaluation Methods
Comparison of the Results of the Generalizability Theory with the Inter-Rater Agreement Coefficients
Eser, Mehmet Taha; Aksu, Gökhan – International Journal of Curriculum and Instruction, 2022
The agreement between raters is examined within the scope of the concept of "inter-rater reliability". Although there are clear definitions of the concepts of agreement between raters and reliability between raters, there is no clear information about the conditions under which agreement and reliability level methods are appropriate to…
Descriptors: Generalizability Theory, Interrater Reliability, Evaluation Methods, Test Theory
Sturgis, Paul W.; Marchand, Leslie; Miller, M. David; Xu, Wei; Castiglioni, Analia – Association for Institutional Research, 2022
This article introduces generalizability theory (G-theory) to institutional research and assessment practitioners, and explains how it can be utilized to evaluate the reliability of assessment procedures in order to improve student learning outcomes. The fundamental concepts associated with G-theory are briefly discussed, followed by a discussion…
Descriptors: Generalizability Theory, Institutional Research, Reliability, Computer Software
Zhiwen Tang – ProQuest LLC, 2021
Artificial intelligence (AI) aims to build intelligent systems that can interact with and assist humans. During the interaction, a system learns the requirements from the human user and adapts to the needs to complete tasks. A popular type of interactive system is retrieval-based, where the system uses a retrieval function to retrieve relevant…
Descriptors: Artificial Intelligence, Intelligent Tutoring Systems, Objectives, Reinforcement
Akaeze, Hope O.; Wu, Jamie Heng-Chieh; Lawrence, Frank R.; Weber, Everett P. – Journal of Psychoeducational Assessment, 2023
This paper reports an investigation into the psychometric properties of the COR-Advantage1.5 (COR-Adv1.5) assessment tool, a criterion-referenced observation-based instrument designed to assess the developmental abilities of children from birth through kindergarten. Using data from 8534 children participating in a state-funded preschool program…
Descriptors: Criterion Referenced Tests, Evaluation Methods, Measures (Individuals), Measurement Techniques
May, Henry; Blackman, Horatio; Van Horne, Sam; Tilley, Katherine; Farley-Ripple, Elizabeth N.; Shewchuk, Samantha; Agboh, Darren; Micklos, Deborah Amsden – Center for Research Use in Education, 2022
In this technical report, the Center for Research Use in Education (CRUE) presents the methodological design of a large-scale quantitative investigation of research use by school-based practitioners through the "Survey of Evidence in Education for Schools (SEE-S)." It documents the major technical aspects of the development of SEE-S,…
Descriptors: Surveys, Schools, Educational Research, Research Utilization
Tipton, Elizabeth; Fellers, Lauren; Caverly, Sarah; Vaden-Kiernan, Michael; Borman, Geoffrey; Sullivan, Kate; Ruiz de Castilla, Veronica – Journal of Research on Educational Effectiveness, 2016
Recently, statisticians have begun developing methods to improve the generalizability of results from large-scale experiments in education. This work has included the development of methods for improved site selection when random sampling is infeasible, including the use of stratification and targeted recruitment strategies. This article provides…
Descriptors: Generalizability Theory, Site Selection, Experiments, Comparative Analysis
Gresham, Frank M.; Dart, Evan H.; Collins, Tai A. – School Psychology Review, 2017
The concept of treatment integrity is an essential component to databased decision making within a response-to-intervention model. Although treatment integrity is a topic receiving increased attention in the school-based intervention literature, relatively few studies have been conducted regarding the technical adequacy of treatment integrity…
Descriptors: Fidelity, Generalizability Theory, Observation, Measurement Techniques
McLaughlin, Tara W.; Snyder, Patricia A.; Algina, James – Grantee Submission, 2017
The Learning Target Rating Scale (LTRS) is a measure designed to evaluate the quality of teacher-developed learning targets for embedded instruction for early learning. In the present study, we examined the measurement dependability of LTRS scores by conducting a generalizability study (G-study). We used a partially nested, three-facet model to…
Descriptors: Generalizability Theory, Scores, Rating Scales, Evaluation Methods
Lin, Chih-Kai – Language Testing, 2017
Sparse-rated data are common in operational performance-based language tests, as an inevitable result of assigning examinee responses to a fraction of available raters. The current study investigates the precision of two generalizability-theory methods (i.e., the rating method and the subdividing method) specifically designed to accommodate the…
Descriptors: Data Analysis, Language Tests, Generalizability Theory, Accuracy
Kim, Young-Suk Grace; Schatschneider, Christopher; Wanzek, Jeanne; Gatlin, Brandy; Al Otaiba, Stephanie – Reading and Writing: An Interdisciplinary Journal, 2017
We examined how raters and tasks influence measurement error in writing evaluation and how many raters and tasks are needed to reach a desirable level of 0.90 and 0.80 reliabilities for children in Grades 3 and 4. A total of 211 children (102 boys) were administered three tasks in narrative and expository genres, respectively, and their written…
Descriptors: Writing Evaluation, Elementary School Students, Grade 3, Grade 4
Kim, Young-Suk Grace; Schatschneider, Christopher; Wanzek, Jeanne; Gatlin, Brandy; Al Otaiba, Stephanie – Grantee Submission, 2017
We examined how raters and tasks influence measurement error in writing evaluation and how many raters and tasks are needed to reach a desirable level of 0.90 and 0.80 reliabilities for children in Grades 3 and 4. A total of 211 children (102 boys) were administered three tasks in narrative and expository genres, respectively, and their written…
Descriptors: Writing Evaluation, Elementary School Students, Grade 3, Grade 4
Zhang, Bo; Xiao, Yunnan; Luo, Juan – Language Testing in Asia, 2015
Previous studies comparing holistic scoring to analytic scoring of second language writing have given mixed results. Some of them suffer from methodological drawbacks, such as limited writing sample size, limited number of raters, and lack of direct comparison of the two methods. Based on 300 writing samples graded by 14 raters, this research…
Descriptors: Evaluators, Reliability, Scores, Holistic Approach