Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Author
Allison, Howard K., II | 1 |
Gillmore, Gerald M. | 1 |
Gradman, Harry L. | 1 |
Hanania, Edith | 1 |
Hathcoat, John D. | 1 |
Kleinke, David J. | 1 |
Penn, Jeremy D. | 1 |
Stewart, Krista J. | 1 |
Thompson, Bruce | 1 |
Whitely, Susan E. | 1 |
Publication Type
Reports - Research | 6 |
Journal Articles | 4 |
Information Analyses | 2 |
Speeches/Meeting Papers | 2 |
Reports - Evaluative | 1 |
Reports - General | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Oklahoma | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Hathcoat, John D.; Penn, Jeremy D. – Research & Practice in Assessment, 2012
Critics of standardized testing have recommended replacing standardized tests with more authentic assessment measures, such as classroom assignments, projects, or portfolios rated by a panel of raters using common rubrics. Little research has examined the consistency of scores across multiple authentic assignments or the implications of this…
Descriptors: Generalizability Theory, Performance Based Assessment, Writing Across the Curriculum, Standardized Tests

Whitely, Susan E. – Applied Psychological Measurement, 1979
A model which gives maximum likelihood estimates of measurement error within the context of a simplex model for practice effects is presented. The appropriateness of the model is tested for five traits, and error estimates are compared to the classical formula estimates. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Models

Stewart, Krista J. – Psychology in the Schools, 1987
Evaluated the technical aspects of three Wechsler Intelligence Scale for Children-Revised (WISC-R) administrations of five psychology graduate students using the WISC-R Administration Observational Checklist (WAOC) to evaluate interrater agreement. Students performed significantly better on the second than on the first observation, with…
Descriptors: Educational Diagnosis, Error Patterns, Examiners, Graduate Students

Kleinke, David J. – Applied Psychological Measurement, 1979
Lord's, Millman's and Saupe's methods of approximating the standard error of measurement are reviewed. Through an empirical demonstration involving 200 university classroom tests, all three approximations are shown to be biased. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Formulas
Gillmore, Gerald M.; And Others – 1976
Generalizability theory is described and applied to data collection methods for student ratings of instruction. Three generalizability coefficients (generalizing over both students and items, generalizing over students only, and generalizing over items only) are defined, and the appropriateness of each for various purposes is discussed.…
Descriptors: Analysis of Variance, College Faculty, College Students, Course Evaluation
Gradman, Harry L.; Hanania, Edith – 1988
A study investigated the variability of language performance on different types of testing task, global versus discrete-focus. Three tests (cloze, multiple-choice, and fill-in-the-blank) were developed to measure learners' knowledge of five verb forms. The tests, containing corresponding items designed to elicit equivalent structures, were…
Descriptors: Cloze Procedure, College Students, Comparative Analysis, English (Second Language)
Thompson, Bruce – 1994
Dissertations are an important component of the effort to generate knowledge. Thus, dissertation quality may be seen by accreditation and coordinating-board reviewers as a noteworthy reflection on the quality of doctoral programs themselves. The present study reviews methodological errors within Ph.D. dissertations. The illustrative errors are…
Descriptors: Behavioral Science Research, Case Studies, Doctoral Dissertations, Error Patterns
Allison, Howard K., II – 1978
To assess the impact of various rating scale formats in rating student teachers, a Likert-type scale (Highly agree/slightly agree/neither/slightly disagree/highly disagree) was derived from the graphic scale currently used by the cooperating teachers and university supervisors. They rated the music student teachers on an eight-point scale from…
Descriptors: Analysis of Variance, Behavior Rating Scales, Comparative Testing, Cooperating Teachers