NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 436 to 450 of 1,166 results Save | Export
Peer reviewed Peer reviewed
Carver, Ronald P. – Journal of Reading Behavior, 1985
Describes a procedure for rescaling objective measures of comprehension so that they reflect amounts of accuracy of comprehension on an absolute scale. Suggests there is adequate empirical evidence supporting the validity of the rauding rescaling procedure. (RS)
Descriptors: Elementary Secondary Education, Predictive Validity, Reading Comprehension, Test Theory
Peer reviewed Peer reviewed
D'Amato, Rik Carl; And Others – Journal of School Psychology, 1988
Investigated the overlap between the Wechsler Intelligence Scale for Children - Revised (WISC-R) and the Halstead-Reitan Neuropsychological Battery (HRNB) in light of their use in diagnosing children's learning problems using scores for children (N=1,181) on the WISC-R and the HRNB. Results showed primary overlap between measures was attributed to…
Descriptors: Adolescents, Children, Intelligence Tests, Test Items
Peer reviewed Peer reviewed
Engelhard, George, Jr. – Educational and Psychological Measurement, 1992
A historical perspective is provided of the concept of invariance in measurement theory, describing sample-invariant item calibration and item-invariant measurement of individuals. Invariance as a key measurement concept is illustrated through the measurement theories of E. L. Thorndike, L. L. Thurstone, and G. Rasch. (SLD)
Descriptors: Behavioral Sciences, Educational History, Measurement Techniques, Psychometrics
Peer reviewed Peer reviewed
Raykov, Tenko – Applied Psychological Measurement, 1998
Examines the relationship between Cronbach's coefficient alpha and the reliability of a composite of a prespecified set of interrelated nonhomogeneous components through simulation. Shows that alpha can over- or underestimate scale reliability at the population level. Illustrates the bias in terms of structural parameters. (SLD)
Descriptors: Reliability, Simulation, Statistical Bias, Structural Equation Models
Peer reviewed Peer reviewed
MacMillan, Peter D. – Journal of Experimental Education, 2000
Compared classical test theory (CTT), generalizability theory (GT), and multifaceted Rasch model (MFRM) approaches to detecting and correcting for rater variability using responses of 4,930 high school students graded by 3 raters on 9 scales. The MFRM approach identified far more raters as different than did the CTT analysis. GT and Rasch…
Descriptors: Generalizability Theory, High School Students, High Schools, Interrater Reliability
Peer reviewed Peer reviewed
Brennan, Robert L. – Educational Measurement: Issues and Practice, 1997
The history of generalizability theory (G theory) is told from the perspective of one researcher's experiences, describing psychometric and scientific perspectives that influenced the development of G theory and its adoption. Work that remains to be done in the field is outlined. (SLD)
Descriptors: Educational Testing, Generalizability Theory, Measurement, Psychometrics
Peer reviewed Peer reviewed
Wollack, James A. – Applied Psychological Measurement, 1997
Introduces a new Item Response Theory (IRT) based statistic for detecting answer copying. Compares this omega statistic with the best classical test theory-based statistic under various conditions, and finds omega superior based on Type I error rate and power. (SLD)
Descriptors: Cheating, Identification, Item Response Theory, Power (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Kupermintz, Haggai – Journal of Educational Measurement, 2004
A decision-theoretic approach to the question of reliability in categorically scored examinations is explored. The concepts of true scores and errors are discussed as they deviate from conventional psychometric definitions and measurement error in categorical scores is cast in terms of misclassifications. A reliability measure based on…
Descriptors: Test Reliability, Error of Measurement, Psychometrics, Test Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Biswas, Ajoy Kumar – Applied Psychological Measurement, 2006
This article studies the ordinal reliability of (total) test scores. This study is based on a classical-type linear model of observed score (X), true score (T), and random error (E). Based on the idea of Kendall's tau-a coefficient, a measure of ordinal reliability for small-examinee populations is developed. This measure is extended to large…
Descriptors: True Scores, Test Theory, Test Reliability, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Borsboom, Denny – Psychometrika, 2006
This paper analyzes the theoretical, pragmatic, and substantive factors that have hampered the integration between psychology and psychometrics. Theoretical factors include the operationalist mode of thinking which is common throughout psychology, the dominance of classical test theory, and the use of "construct validity" as a catch-all category…
Descriptors: Psychometrics, Psychology, Test Theory, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J.; Sotaridona, Leonardo – Journal of Educational and Behavioral Statistics, 2006
A statistical test for detecting answer copying on multiple-choice items is presented. The test is based on the exact null distribution of the number of random matches between two test takers under the assumption that the response process follows a known response model. The null distribution can easily be generalized to the family of distributions…
Descriptors: Test Items, Multiple Choice Tests, Cheating, Responses
Peer reviewed Peer reviewed
Direct linkDirect link
Reid, Christine A.; Kolakowsky-Hayner, Stephanie A.; Lewis, Allen N.; Armstrong, Amy J. – Rehabilitation Counseling Bulletin, 2007
Item response theory (IRT) methodology is introduced as a tool for improving assessment instruments used with people who have disabilities. Need for this approach in rehabilitation is emphasized; differences between IRT and classical test theory are clarified. Concepts essential to understanding IRT are defined, necessary data assumptions are…
Descriptors: Psychometrics, Methods, Item Response Theory, Aptitude Tests
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J.; Breithaupt, Krista; Chuah, Siang Chee; Zhang, Yanwei – Journal of Educational Measurement, 2007
A potential undesirable effect of multistage testing is differential speededness, which happens if some of the test takers run out of time because they receive subtests with items that are more time intensive than others. This article shows how a probabilistic response-time model can be used for estimating differences in time intensities and speed…
Descriptors: Adaptive Testing, Evaluation Methods, Test Items, Reaction Time
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mapuranga, Raymond; Dorans, Neil J.; Middleton, Kyndra – ETS Research Report Series, 2008
In many practical settings, essentially the same differential item functioning (DIF) procedures have been in use since the late 1980s. Since then, examinee populations have become more heterogeneous, and tests have included more polytomously scored items. This paper summarizes and classifies new DIF methods and procedures that have appeared since…
Descriptors: Test Bias, Educational Development, Evaluation Methods, Statistical Analysis
Peer reviewed Peer reviewed
Huynh, Huynh – Psychometrika, 1977
A model for the setting of mastery cut scores is presented. The model, based on the beta-binomial test distribution, allows for hand calculation of cut scores. The model provides a simple way to explore the consequences of selecting a particular cut score. (Author/JKS)
Descriptors: Career Development, Cutting Scores, Mastery Tests, Mathematical Models
Pages: 1  |  ...  |  26  |  27  |  28  |  29  |  30  |  31  |  32  |  33  |  34  |  ...  |  78