Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Error of Measurement | 8 |
Interrater Reliability | 8 |
Mathematical Models | 8 |
Higher Education | 3 |
Rating Scales | 3 |
Correlation | 2 |
Evaluation Methods | 2 |
Generalizability Theory | 2 |
Measurement Techniques | 2 |
Observation | 2 |
Regression (Statistics) | 2 |
More ▼ |
Author
Cason, Carolyn L. | 1 |
Cason, Gerald J. | 1 |
Chen, Hsueh-Chih | 1 |
Chen, Po-Hsi | 1 |
Hung, Su-Pin | 1 |
Raymond, Mark R. | 1 |
Rowley, Glenn L. | 1 |
Viswesvaran, Chockalingam | 1 |
Webber, Larry | 1 |
Zegers, Frits E. | 1 |
Zwick, Rebecca | 1 |
More ▼ |
Publication Type
Reports - Research | 5 |
Journal Articles | 3 |
Reports - Evaluative | 3 |
Speeches/Meeting Papers | 3 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Researchers | 2 |
Location
Taiwan (Taipei) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Hung, Su-Pin; Chen, Po-Hsi; Chen, Hsueh-Chih – Creativity Research Journal, 2012
Product assessment is widely applied in creative studies, typically as an important dependent measure. Within this context, this study had 2 purposes. First, the focus of this research was on methods for investigating possible rater effects, an issue that has not received a great deal of attention in past creativity studies. Second, the…
Descriptors: Item Response Theory, Creativity, Interrater Reliability, Undergraduate Students
Zwick, Rebecca – 1986
Most currently used measures of inter-rater agreement for the nominal case incorporate a correction for "chance agreement." The definition of chance agreement is not the same for all coefficients, however. Three chance-corrected coefficients are Cohen's Kappa; Scott's Pi; and the S index of Bennett, Goldstein, and Alpert, which has…
Descriptors: Error of Measurement, Interrater Reliability, Mathematical Models, Measurement Techniques

Zegers, Frits E. – Applied Psychological Measurement, 1991
The degree of agreement between two raters rating several objects for a single characteristic can be expressed through an association coefficient, such as the Pearson product-moment correlation. How to select an appropriate association coefficient, and the desirable properties and uses of a class of such coefficients--the Euclidean…
Descriptors: Classification, Correlation, Data Interpretation, Equations (Mathematics)
Cason, Gerald J.; Cason, Carolyn L. – 1989
The use of three remedies for errors in the measurement of ability that arise from differences in rater stringency is discussed. Models contrasted are: (1) Conventional; (2) Handicap; and (3) deterministic Rater Response Theory (RRT). General model requirements, power, bias of measures, computing cost, and complexity are contrasted. Contrasts are…
Descriptors: Ability, Achievement Rating, Error of Measurement, Evaluation Methods
Rowley, Glenn L. – 1986
Classroom researchers are frequently urged to provide evidence of the reliability of their data. In the case of observational data, three approaches to this have emerged: observer agreement, generalizability theory, and measurement error. Generalizability theory provides the most powerful approach given an adequate data collection design, but…
Descriptors: Classroom Observation Techniques, Classroom Research, Correlation, Elementary Education
Webber, Larry; And Others – 1986
Generalizability theory, which subsumes classical measurement theory as a special case, provides a general model for estimating the reliability of observational rating data by estimating the variance components of the measurement design. Research data from the "Heart Smart" health intervention program were analyzed as a heuristic tool.…
Descriptors: Behavior Rating Scales, Cardiovascular System, Error of Measurement, Generalizability Theory
van der Linden, Wim J. – 1982
A latent trait method is presented to investigate the possibility that Angoff or Nedelsky judges specify inconsistent probabilities in standard setting techniques for objectives-based instructional programs. It is suggested that judges frequently specify a low probability of success for an easy item but a large probability for a hard item. The…
Descriptors: Criterion Referenced Tests, Cutting Scores, Error of Measurement, Interrater Reliability

Raymond, Mark R.; Viswesvaran, Chockalingam – Journal of Educational Measurement, 1993
Three variations of a least squares regression model are presented that are suitable for determining and correcting for rating error in designs in which examinees are evaluated by a subset of possible raters. Models are applied to ratings from 4 administrations of a medical certification examination in which 40 raters and approximately 115…
Descriptors: Error of Measurement, Evaluation Methods, Higher Education, Interrater Reliability