Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 3 |
Descriptor
Error of Measurement | 9 |
Reliability | 9 |
Test Reliability | 9 |
Statistical Analysis | 3 |
True Scores | 3 |
Analysis of Variance | 2 |
Classification | 2 |
Correlation | 2 |
Criterion Referenced Tests | 2 |
Cutting Scores | 2 |
Mastery Tests | 2 |
More ▼ |
Source
ACT, Inc. | 1 |
Applied Psychological… | 1 |
Journal of Educational… | 1 |
Journal of Experimental… | 1 |
Language Testing | 1 |
Author
Livingston, Samuel A. | 2 |
Harris, Chester W. | 1 |
Harris, Deborah J. | 1 |
Hsin-Yun Lee | 1 |
Hunyh, Hunyh | 1 |
Li, Dongmei | 1 |
Li, Minzi | 1 |
Li-Jen Weng | 1 |
Olejnik, Stephen F. | 1 |
Porter, Andrew C. | 1 |
Powers, Sonya | 1 |
More ▼ |
Publication Type
Reports - Research | 6 |
Journal Articles | 4 |
Numerical/Quantitative Data | 2 |
Guides - General | 1 |
Guides - Non-Classroom | 1 |
Reports - Descriptive | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 1 |
What Works Clearinghouse Rating
Hsin-Yun Lee; You-Lin Chen; Li-Jen Weng – Journal of Experimental Education, 2024
The second version of Kaiser's Measure of Sampling Adequacy (MSA[subscript 2]) has been widely applied to assess the factorability of data in psychological research. The MSA[subscript 2] is developed in the population and little is known about its behavior in finite samples. If estimated MSA[subscript 2]s are biased due to sampling errors,…
Descriptors: Error of Measurement, Reliability, Sampling, Statistical Bias
Li, Minzi; Zhang, Xian – Language Testing, 2021
This meta-analysis explores the correlation between self-assessment (SA) and language performance. Sixty-seven studies with 97 independent samples involving more than 68,500 participants were included in our analysis. It was found that the overall correlation between SA and language performance was 0.466 (p < 0.01). Moderator analysis was…
Descriptors: Meta Analysis, Self Evaluation (Individuals), Likert Scales, Research Reports
Powers, Sonya; Li, Dongmei; Suh, Hongwook; Harris, Deborah J. – ACT, Inc., 2016
ACT reporting categories and ACT Readiness Ranges are new features added to the ACT score reports starting in fall 2016. For each reporting category, the number correct score, the maximum points possible, the percent correct, and the ACT Readiness Range, along with an indicator of whether the reporting category score falls within the Readiness…
Descriptors: Scores, Classification, College Entrance Examinations, Error of Measurement

Hunyh, Hunyh; Saunders, Joseph C. – 1979
Comparisons were made among various methods of estimating the reliability of pass-fail decisions based on mastery tests. The reliability indices that are considered are p, the proportion of agreements between two estimates, and kappa, the proportion of agreements corrected for chance. Estimates of these two indices were made on the basis of…
Descriptors: Cutting Scores, Error of Measurement, Mastery Tests, Reliability
Livingston, Samuel A. – 1976
A distinction is made between reliability of measurement and reliability of classification; the "criterion-referenced reliability coefficient" describes the former. Application of this coefficient to the probability distribution of possible scores for a single student yields a meaningful way to describe the reliability of a single score. (Author)
Descriptors: Classification, Criterion Referenced Tests, Error of Measurement, Measurement

Harris, Chester W. – 1971
Livingston's work is a careful analysis of what occurs when one pools two populations with different means, but similar variances and reliability coefficients. However, his work fails to advance reliability theory for the special case of criterion-referenced testing. See ED 042 802 for Livingston's paper. (MS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Error of Measurement, Reliability

Livingston, Samuel A.; Wingersky, Marilyn A. – Journal of Educational Measurement, 1979
Procedures are described for studying the reliability of decisions based on specific passing scores with tests made up of discrete items and designed to measure continuous rather than categorical traits. These procedures are based on the estimation of the joint distribution of true scores and observed scores. (CTM)
Descriptors: Cutting Scores, Decision Making, Efficiency, Error of Measurement

Whitely, Susan E. – Applied Psychological Measurement, 1979
Two sources of inconsistency were separated by reanalyzing data from a major study on short-term consistency. Little evidence was found for generalizability or behavioral predictability. Results supported the assumption that measurement error from short-term fluctuations is not due to systematic individual differences in response consistency.…
Descriptors: Behavior Change, Cognitive Processes, College Freshmen, Error of Measurement
Olejnik, Stephen F.; Porter, Andrew C. – 1978
The statistical properties of two methods of estimating gain scores for groups in quasi-experiments are compared: (1) gains in scores standardized separately for each group; and (2) analysis of covariance with estimated true pretest scores. The fan spread hypothesis is assumed for groups but not necessarily assumed for members of the groups.…
Descriptors: Academic Achievement, Achievement Gains, Analysis of Covariance, Analysis of Variance