Publication Date
| In 2026 | 1 |
| Since 2025 | 168 |
| Since 2022 (last 5 years) | 1021 |
| Since 2017 (last 10 years) | 2336 |
| Since 2007 (last 20 years) | 6522 |
Descriptor
| Reliability | 9761 |
| Validity | 3866 |
| Foreign Countries | 2823 |
| Measures (Individuals) | 1892 |
| Correlation | 1522 |
| Factor Analysis | 1460 |
| Statistical Analysis | 1278 |
| Questionnaires | 1084 |
| Scores | 1064 |
| Student Attitudes | 1034 |
| Psychometrics | 979 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Researchers | 181 |
| Practitioners | 101 |
| Teachers | 61 |
| Administrators | 42 |
| Policymakers | 33 |
| Students | 21 |
| Counselors | 10 |
| Media Staff | 5 |
| Community | 1 |
| Parents | 1 |
| Support Staff | 1 |
| More ▼ | |
Location
| Turkey | 454 |
| Australia | 155 |
| Canada | 144 |
| China | 127 |
| United States | 127 |
| Taiwan | 107 |
| United Kingdom | 100 |
| Nigeria | 98 |
| California | 95 |
| Netherlands | 91 |
| Indonesia | 86 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 3 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 2 |
Peer reviewedThornton, Billy W.; Croskey, Frank L. – Educational and Psychological Measurement, 1975
Describes method of indexing interobserver reliability from data relating to mutually exclusive nominal categories. Program computes reliability estimates from nominal data collected in a time interval manner. A specified critical time interval, order of responses, and number of responses are important factors in computing amount of agreement…
Descriptors: Computer Programs, Observation, Reliability, Time
Krause, Merton – J Gen Psychol, 1969
Descriptors: Measurement Instruments, Reliability, Research, Validity
Orange Coast Junior Coll. District, Costa Mesa, CA. – 1969
A 9-item questionnaire was given to registering students for fall and spring 1969. For easier handling, its form was changed for the next semester. By accident, the two versions were handed out and 687 students answered both in quick succession. The staff decided to compare the two forms to see if the answers differed. Two questions were…
Descriptors: Questionnaires, Reliability, Two Year Colleges
Peer reviewedKoslowsky, Meni; Bailit, Howard – Educational and Psychological Measurement, 1975
An equation, introduced by Goodman and Kruskal for obtaining a reliability measure of one item was expanded. This formula determines inter-rater reliability for a series of items across many subjects. The statistic that results is easily interpreted and in many ways is analogous to the conventional reliability for quantitative data. (Author/BJG)
Descriptors: Error Patterns, Reliability, Research Problems
Peer reviewedBehrens, Elke; Brambring, Michael – International Journal of Rehabilitation Research, 1987
A German version of the "International Classification of Impairments, Disabilities, and Handicaps" was tested in respect to interrater reliability both on the method of relative agreement between raters and on the measure "kappa." Unsatisfactory reliability was found. (Author/DB)
Descriptors: Classification, Disabilities, Foreign Countries, Reliability
Peer reviewedMendoza, Jorge L.; Mumford, Michael – Journal of Educational Statistics, 1987
This paper delineates the factors that influence the joint impact of attenuation and range restriction on the magnitude of the correlation coefficient. A theoretical framework for the generation of appropriate correction formulas is established. (TJH)
Descriptors: Ability, Correlation, Predictive Validity, Reliability
Roberts, J. Kyle; Onwuegbuzie, Anthony J.; Eby, J. Robert – 2001
This paper suggests that although data from a homogenous sample might yield less reliable scores than did an inducted sample, these data should not be discarded until further examination of the data is conducted. The paper presents two statistics for monitoring data homogeneity and one statistic for correcting alpha when homogeneity is large. The…
Descriptors: Error of Measurement, Reliability, Scores
Weber, Deborah A. – 2002
Greater understanding and use of confidence intervals is central to changes in statistical practice (G. Cumming and S. Finch, 2001). Reliability coefficients and confidence intervals for reliability coefficients can be computed using a variety of methods. Estimating confidence intervals includes both central and noncentral distribution approaches.…
Descriptors: Reliability, Research Reports, Statistical Distributions
Kane, Michael; Case, Susan – 2003
The scores on two distinct tests (e.g., essay and objective) are often combined into a composite score, which is used to make decisions. The validity of the observed composite can sometimes be evaluated relative to a separate criterion. In cases where no criterion is available, the observed composite has generally been evaluated in terms of its…
Descriptors: Reliability, Simulation, Validity, Weighted Scores
Peer reviewedMugford, Len – Reading, 1970
Presents ...a method for assessing the level of difficulty of reading material in the 7 to 15 year range, together with a supplementary method for dealing with material below the 7 year level." (Author)
Descriptors: Measurement Instruments, Prediction, Readability, Reliability
Peer reviewedKraemer, Helena Chmura – Psychometrika, 1979
Coefficient Kappa is generally defined in terms of procedures of computation rather than in terms of a population. Here a population definition is proposed. Factors influencing the magnitude of Kappa are identified. Strategies to improve reliability are proposed, including that of combining multiple unreliable diagnoses. (Author/CTM)
Descriptors: Correlation, Models, Reliability, Research Design
Peer reviewedMilligan, Glenn W. – Psychometrika, 1981
A Monte Carlo evaluation of 30 internal criteria for cluster analysis was conducted using four hierarchical clustering techniques. The results indicated that a subset of internal criteria was identified which appear to be valid indices of correct cluster recovery. (Author/JKS)
Descriptors: Cluster Analysis, Criteria, Reliability, Validity
Peer reviewedSmith, Philip L. – Journal of Educational Measurement, 1981
This study explores a strategy for improving the stability of variance component estimates when only small samples are available, using a series of small, less complex generalizability (G) study designs as a surrogate for a single large design. (Author/BW)
Descriptors: Models, Reliability, Research Design, Sampling
Peer reviewedHartmann, Donald P. – Educational and Psychological Measurement, 1976
In at least two situations, use of the Spearman-Brown prophesy formula yields overestimates of the interobserver reliability of composite scores. The appropriate formulas for estimating the interobserver reliability of composite scores, as well as efficient means of estimating the elements in these formulas are presented. (Author)
Descriptors: Correlation, Measurement Techniques, Observation, Reliability
Peer reviewedVacha-Haase, Tammi; Henson, Robin K.; Caruso, John C. – Educational and Psychological Measurement, 2002
Reliability generalization (RG) is a measurement meta-analytic method used to explore the variability in score reliability estimates and to characterize possible sources of the variance. Summarizes some RG considerations, and suggests how confidence intervals might be portrayed graphically. (SLD)
Descriptors: Generalization, Meta Analysis, Reliability, Scores


