Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 6 |
Descriptor
Generalizability Theory | 9 |
Test Reliability | 9 |
Item Response Theory | 3 |
Comparative Analysis | 2 |
Computation | 2 |
Computer Software | 2 |
Interrater Reliability | 2 |
Item Analysis | 2 |
Language Proficiency | 2 |
Language Tests | 2 |
Scores | 2 |
More ▼ |
Source
Author
Salmani-Nodoushan, Mohammad… | 2 |
Abedi, Jamal | 1 |
Alonzo, Julie | 1 |
Anderson, Daniel | 1 |
Cronbach, Lee J. | 1 |
Huebner, Alan | 1 |
Lai, Cheng-Fei | 1 |
Lucht, Marissa | 1 |
Park, Jasmine, Bitnara | 1 |
Pentimonti, J. | 1 |
Petscher, Y. | 1 |
More ▼ |
Publication Type
Reports - Descriptive | 9 |
Journal Articles | 5 |
Opinion Papers | 1 |
Education Level
Early Childhood Education | 1 |
Elementary Education | 1 |
Grade 1 | 1 |
Primary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Petscher, Y.; Pentimonti, J.; Stanley, C. – National Center on Improving Literacy, 2019
Validity is broadly defined as how well something measures what it's supposed to measure. The reliability and validity of scores from assessments are two concepts that are closely knit together and feed into each other.
Descriptors: Screening Tests, Scores, Test Validity, Test Reliability
Huebner, Alan; Lucht, Marissa – Practical Assessment, Research & Evaluation, 2019
Generalizability theory is a modern, powerful, and broad framework used to assess the reliability, or dependability, of measurements. While there exist classic works that explain the basic concepts and mathematical foundations of the method, there is currently a lack of resources addressing computational resources for those researchers wishing to…
Descriptors: Generalizability Theory, Test Reliability, Computer Software, Statistical Analysis
Uto, Masaki; Ueno, Maomi – IEEE Transactions on Learning Technologies, 2016
As an assessment method based on a constructivist approach, peer assessment has become popular in recent years. However, in peer assessment, a problem remains that reliability depends on the rater characteristics. For this reason, some item response models that incorporate rater parameters have been proposed. Those models are expected to improve…
Descriptors: Item Response Theory, Peer Evaluation, Bayesian Statistics, Simulation
Anderson, Daniel; Park, Jasmine, Bitnara; Lai, Cheng-Fei; Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2012
This technical report is one in a series of five describing the reliability (test/retest/and alternate form) and G-Theory/D-Study research on the easy CBM reading measures, grades 1-5. Data were gathered in the spring 2011 from a convenience sample of students nested within classrooms at a medium-sized school district in the Pacific Northwest. Due…
Descriptors: Test Reliability, Generalizability Theory, Curriculum Based Assessment, Elementary School Students
Salmani-Nodoushan, Mohammad Ali – Journal on Educational Psychology, 2009
A good test is one that has at least three qualities: reliability, or the precision with which a test measures what it is supposed to measure; validity, i.e., if the test really measures what it is supposed to measure, and practicality, or if the test, no matter how sound theoretically, is practicable in reality. These are the sine qua non for any…
Descriptors: Generalizability Theory, Testing, Language Tests, Item Response Theory
Salmani-Nodoushan, Mohammad Ali – Online Submission, 2009
A good test is one that has at least three qualities: reliability, or the precision with which a test measures what it is supposed to measure; validity, i.e., if the test really measures what it is supposed to measure; and practicality, or if the test, no matter how sound theoretically, is practicable in reality. These are the sine qua non for…
Descriptors: Generalizability Theory, Testing, Language Tests, Item Response Theory
Cronbach, Lee J. – Center for Research on Evaluation Standards and Student Testing CRESST, 2004
Where the accuracy of a measurement is important, whether for scientific or practical purposes, the investigator should evaluate how much random error affects the measurement. New research may not be necessary when a procedure has been studied enough to establish how much error it involves. But, with new measures, or measures being transferred…
Descriptors: Error of Measurement, Test Reliability, Generalizability Theory, Educational Research

Abedi, Jamal – Multivariate Behavioral Research, 1996
The Interrater/Test Reliability System (ITRS) is described. The ITRS is a comprehensive computer tool used to address questions of interrater reliability that computes several different indices of interrater reliability and the generalizability coefficient over raters and topics. The system is available in IBM compatible or Macintosh format. (SLD)
Descriptors: Computer Software, Computer Software Evaluation, Evaluation Methods, Evaluators
Phillips, Gary W., Ed. – 1996
Recently, there has been a significant expansion in the use of performance assessment in large scale testing programs. Although there has been significant support from curriculum and policy stakeholders, the technical feasibility of large scale performance assessments has remained a question. This report is intended to contribute to the debate by…
Descriptors: Comparative Analysis, Generalizability Theory, Performance Based Assessment, Psychometrics