NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 436 to 450 of 1,166 results Save | Export
Peer reviewed Peer reviewed
D'Amato, Rik Carl; And Others – Journal of School Psychology, 1988
Investigated the overlap between the Wechsler Intelligence Scale for Children - Revised (WISC-R) and the Halstead-Reitan Neuropsychological Battery (HRNB) in light of their use in diagnosing children's learning problems using scores for children (N=1,181) on the WISC-R and the HRNB. Results showed primary overlap between measures was attributed to…
Descriptors: Adolescents, Children, Intelligence Tests, Test Items
Peer reviewed Peer reviewed
Engelhard, George, Jr. – Educational and Psychological Measurement, 1992
A historical perspective is provided of the concept of invariance in measurement theory, describing sample-invariant item calibration and item-invariant measurement of individuals. Invariance as a key measurement concept is illustrated through the measurement theories of E. L. Thorndike, L. L. Thurstone, and G. Rasch. (SLD)
Descriptors: Behavioral Sciences, Educational History, Measurement Techniques, Psychometrics
Peer reviewed Peer reviewed
Raykov, Tenko – Applied Psychological Measurement, 1998
Examines the relationship between Cronbach's coefficient alpha and the reliability of a composite of a prespecified set of interrelated nonhomogeneous components through simulation. Shows that alpha can over- or underestimate scale reliability at the population level. Illustrates the bias in terms of structural parameters. (SLD)
Descriptors: Reliability, Simulation, Statistical Bias, Structural Equation Models
Peer reviewed Peer reviewed
MacMillan, Peter D. – Journal of Experimental Education, 2000
Compared classical test theory (CTT), generalizability theory (GT), and multifaceted Rasch model (MFRM) approaches to detecting and correcting for rater variability using responses of 4,930 high school students graded by 3 raters on 9 scales. The MFRM approach identified far more raters as different than did the CTT analysis. GT and Rasch…
Descriptors: Generalizability Theory, High School Students, High Schools, Interrater Reliability
Peer reviewed Peer reviewed
Brennan, Robert L. – Educational Measurement: Issues and Practice, 1997
The history of generalizability theory (G theory) is told from the perspective of one researcher's experiences, describing psychometric and scientific perspectives that influenced the development of G theory and its adoption. Work that remains to be done in the field is outlined. (SLD)
Descriptors: Educational Testing, Generalizability Theory, Measurement, Psychometrics
Peer reviewed Peer reviewed
Wollack, James A. – Applied Psychological Measurement, 1997
Introduces a new Item Response Theory (IRT) based statistic for detecting answer copying. Compares this omega statistic with the best classical test theory-based statistic under various conditions, and finds omega superior based on Type I error rate and power. (SLD)
Descriptors: Cheating, Identification, Item Response Theory, Power (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Kupermintz, Haggai – Journal of Educational Measurement, 2004
A decision-theoretic approach to the question of reliability in categorically scored examinations is explored. The concepts of true scores and errors are discussed as they deviate from conventional psychometric definitions and measurement error in categorical scores is cast in terms of misclassifications. A reliability measure based on…
Descriptors: Test Reliability, Error of Measurement, Psychometrics, Test Theory
Mislevy, Robert J.; Behrens, John T.; Bennett, Randy E.; Demark, Sarah F.; Frezzo, Dennis C.; Levy, Roy; Robinson, Daniel H.; Rutstein, Daisy Wise; Shute, Valerie J.; Stanley, Ken; Winters, Fielding I. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2007
People use external knowledge representations (EKRs) to identify, depict, transform, store, share, and archive information. Learning how to work with EKRs is central to becoming proficient in virtually every discipline. As such, EKRs play central roles in curriculum, instruction, and assessment. Five key roles of EKRs in educational assessment are…
Descriptors: Educational Assessment, Computer Networks, Test Construction, Computer Assisted Testing
Peer reviewed Peer reviewed
Huynh, Huynh – Psychometrika, 1977
A model for the setting of mastery cut scores is presented. The model, based on the beta-binomial test distribution, allows for hand calculation of cut scores. The model provides a simple way to explore the consequences of selecting a particular cut score. (Author/JKS)
Descriptors: Career Development, Cutting Scores, Mastery Tests, Mathematical Models
Peer reviewed Peer reviewed
Brennan, Robert L.; And Others – Applied Psychological Measurement, 1988
Seven papers on technical and practical issues in equating are presented. Problems related to the use of conventional and item response theory equating methods, using pre- and post-smoothing to increase equipercentile equating's precision, and linear equating models for common-item nonequivalent-population design are discussed. (SLD)
Descriptors: Equated Scores, Latent Trait Theory, Research Problems, Scaling
Peer reviewed Peer reviewed
Pumfrey, Peter D. – Journal of Research in Reading, 1987
Discusses, for the benefit of research workers and other test users, the ongoing controversy concerning the relative merits of conventional test theory and Rasch scaling in the construction of reading tests. Concludes that a great deal of further research is required to see whether these approaches are educationally valid. (JD)
Descriptors: Reading Research, Reading Tests, Test Construction, Test Format
Peer reviewed Peer reviewed
Gardner, Robert C.; Erdle, Stephen – Educational and Psychological Measurement, 1986
This article evaluated criticisms by Stevens and Aleamoni (1986) of an article by Gardner and Erdle (1984) on aggregation using either raw or standard scores. It was demonstrated that their criticisms were unfounded. (Author)
Descriptors: Correlation, Factor Analysis, Raw Scores, Scores
Peer reviewed Peer reviewed
Penfield, Douglas A.; Koffler, Stephen L. – Educational and Psychological Measurement, 1986
The development of a nonparametric K-sample test for equality of slopes using Puri's generalized L statistic is presented. The test is recommended when the assumptions underlying the parametric model are violated. This procedure replaces original data with either ranks (for data with heavy tails) or normal scores (for data with light tails).…
Descriptors: Mathematical Models, Nonparametric Statistics, Regression (Statistics), Sampling
Peer reviewed Peer reviewed
Secolsky, Charles – Journal of Educational Measurement, 1987
For measuring the face validity of a test, Nevo suggested that test takers and nonprofessional users rate items on a five point scale. This article questions the ability of those raters and the credibility of the aggregated judgment as evidence of the validity of the test. (JAZ)
Descriptors: Content Validity, Measurement Techniques, Rating Scales, Test Items
Peer reviewed Peer reviewed
Woodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores
Pages: 1  |  ...  |  26  |  27  |  28  |  29  |  30  |  31  |  32  |  33  |  34  |  ...  |  78