Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 3 |
Descriptor
Comparative Analysis | 7 |
Generalizability Theory | 7 |
Test Reliability | 7 |
Item Response Theory | 3 |
Scores | 3 |
Test Theory | 3 |
Accuracy | 1 |
Bayesian Statistics | 1 |
Children | 1 |
Clinical Diagnosis | 1 |
College Admission | 1 |
More ▼ |
Source
Advances in Health Sciences… | 1 |
Applied Measurement in… | 1 |
Educational and Psychological… | 1 |
IEEE Transactions on Learning… | 1 |
Author
Publication Type
Journal Articles | 4 |
Reports - Research | 3 |
Reports - Descriptive | 2 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 2 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Childrens Depression Inventory | 1 |
What Works Clearinghouse Rating
Zaidi, Nikki L.; Swoboda, Christopher M.; Kelcey, Benjamin M.; Manuel, R. Stephen – Advances in Health Sciences Education, 2017
The extant literature has largely ignored a potentially significant source of variance in multiple mini-interview (MMI) scores by "hiding" the variance attributable to the sample of attributes used on an evaluation form. This potential source of hidden variance can be defined as rating items, which typically comprise an MMI evaluation…
Descriptors: Interviews, Scores, Generalizability Theory, Monte Carlo Methods
Uto, Masaki; Ueno, Maomi – IEEE Transactions on Learning Technologies, 2016
As an assessment method based on a constructivist approach, peer assessment has become popular in recent years. However, in peer assessment, a problem remains that reliability depends on the rater characteristics. For this reason, some item response models that incorporate rater parameters have been proposed. Those models are expected to improve…
Descriptors: Item Response Theory, Peer Evaluation, Bayesian Statistics, Simulation
Brennan, Robert L. – Applied Measurement in Education, 2011
Broadly conceived, reliability involves quantifying the consistencies and inconsistencies in observed scores. Generalizability theory, or G theory, is particularly well suited to addressing such matters in that it enables an investigator to quantify and distinguish the sources of inconsistencies in observed scores that arise, or could arise, over…
Descriptors: Generalizability Theory, Test Theory, Test Reliability, Item Response Theory
Eason, Sandra H. – 1989
Generalizability theory provides a technique for accurately estimating the reliability of measurements. The power of this theory is based on the simultaneous analysis of multiple sources of error variances. Equally important, generalizability theory considers relationships among the sources of measurement error. Just as multivariate inferential…
Descriptors: Comparative Analysis, Generalizability Theory, Test Reliability, Test Theory
Linacre, John M. – 1993
Generalizability theory (G-theory) and many-facet Rasch measurement (Rasch) manage the variability inherent when raters rate examinees on test items. The purpose of G-theory is to estimate test reliability in a raw score metric. Unadjusted examinee raw scores are reported as measures. A variance component is estimated for the examinee…
Descriptors: Comparative Analysis, Equations (Mathematics), Estimation (Mathematics), Evaluators

Crowley, Susan L.; And Others – Educational and Psychological Measurement, 1994
Dependability of the Children's Depression Inventory (CDI) was studied using both generalizability and classical test score analyses with a sample of 164 elementary school students. Results suggest that sources of error variance interact to decrease dependability of CDI scores. Depression in children might be better assessed through multiple…
Descriptors: Children, Clinical Diagnosis, Comparative Analysis, Depression (Psychology)
Phillips, Gary W., Ed. – 1996
Recently, there has been a significant expansion in the use of performance assessment in large scale testing programs. Although there has been significant support from curriculum and policy stakeholders, the technical feasibility of large scale performance assessments has remained a question. This report is intended to contribute to the debate by…
Descriptors: Comparative Analysis, Generalizability Theory, Performance Based Assessment, Psychometrics