Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 5 |
Descriptor
Generalizability Theory | 7 |
Scores | 7 |
Interrater Reliability | 3 |
Educational Testing | 2 |
Evaluation | 2 |
Measurement | 2 |
Test Reliability | 2 |
Bilingualism | 1 |
Classroom Observation… | 1 |
Comparative Analysis | 1 |
Competence | 1 |
More ▼ |
Source
Educational Psychologist | 1 |
Educational Research and… | 1 |
Educational Researcher | 1 |
Journal of Consulting and… | 1 |
Journal of Educational and… | 1 |
Multivariate Behavioral… | 1 |
National Center on Improving… | 1 |
Author
Abedi, Jamal | 1 |
Atkins, David C. | 1 |
Baldwin, Scott A. | 1 |
Charalambous, Charalambos Y. | 1 |
Hill, Heather C. | 1 |
Imel, Zac E. | 1 |
Kraft, Matthew A. | 1 |
Li, Min | 1 |
Longford, N. T. | 1 |
Pentimonti, J. | 1 |
Petscher, Y. | 1 |
More ▼ |
Publication Type
Reports - Descriptive | 7 |
Journal Articles | 6 |
Opinion Papers | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Petscher, Y.; Pentimonti, J.; Stanley, C. – National Center on Improving Literacy, 2019
Validity is broadly defined as how well something measures what it's supposed to measure. The reliability and validity of scores from assessments are two concepts that are closely knit together and feed into each other.
Descriptors: Screening Tests, Scores, Test Validity, Test Reliability
Shavelson, Richard J. – Educational Psychologist, 2013
E. L. Thorndike contributed significantly to the field of educational and psychological testing as well as more broadly to psychological studies in education. This article follows in his testing legacy. I address the escalating demand, across societal sectors, to measure individual and group competencies. In formulating an approach to measuring…
Descriptors: Competence, Psychology, Psychological Testing, Psychological Studies
Baldwin, Scott A.; Imel, Zac E.; Atkins, David C. – Journal of Consulting and Clinical Psychology, 2012
Objective: Crits-Christoph, Connolly Gibbons, Hamilton, Ring-Kurtz, and Gallop (2011) used generalizability theory to critique the measurement of the therapeutic alliance in psychotherapy research, showing that the dependability of alliance scores may be quite low, which in turn can lead to attenuated alliance-outcome correlation estimates. Method…
Descriptors: Psychotherapy, Generalizability Theory, Counselor Client Relationship, Scores
Solano-Flores, Guillermo; Li, Min – Educational Research and Evaluation, 2013
We discuss generalizability (G) theory and the fair and valid assessment of linguistic minorities, especially emergent bilinguals. G theory allows examination of the relationship between score variation and language variation (e.g., variation of proficiency across languages, language modes, and social contexts). Studies examining score variation…
Descriptors: Measurement, Testing, Language Proficiency, Test Construction
Hill, Heather C.; Charalambous, Charalambos Y.; Kraft, Matthew A. – Educational Researcher, 2012
In recent years, interest has grown in using classroom observation as a means to several ends, including teacher development, teacher evaluation, and impact evaluation of classroom-based interventions. Although education practitioners and researchers have developed numerous observational instruments for these purposes, many developers fail to…
Descriptors: Generalizability Theory, Observation, Classroom Observation Techniques, Evaluation

Longford, N. T. – Journal of Educational and Behavioral Statistics, 1994
Presents a model-based approach to rater reliability for essays read by multiple raters. The approach is motivated by generalizability theory, and variation of rater severity and rater inconsistency is considered in the presence of between-examinee variations. Illustrates methods with data from standardized educational tests. (Author/SLD)
Descriptors: Educational Testing, Essay Tests, Generalizability Theory, Interrater Reliability

Abedi, Jamal – Multivariate Behavioral Research, 1996
The Interrater/Test Reliability System (ITRS) is described. The ITRS is a comprehensive computer tool used to address questions of interrater reliability that computes several different indices of interrater reliability and the generalizability coefficient over raters and topics. The system is available in IBM compatible or Macintosh format. (SLD)
Descriptors: Computer Software, Computer Software Evaluation, Evaluation Methods, Evaluators