NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 496 to 510 of 1,166 results Save | Export
Peer reviewed Peer reviewed
Williams, Richard H.; Zimmerman, Donald W. – Journal of Experimental Education, 1980
It is suggested that error of measurement cannot be routinely incorporated into the "error term" in statistical tests, and that the reliability of test scores does not have the simple relationship to statistical inference that one might expect. (Author/GK)
Descriptors: Error of Measurement, Hypothesis Testing, Mathematical Formulas, Test Reliability
Peer reviewed Peer reviewed
Green, Bert F. – American Psychologist, 1981
Discusses classical test theory, including test construction, administration, and use. Covers basic statistical concepts in measurement, reliability, and validity; principles of sound test construction and item analysis; test administration and scoring; procedures for transforming raw test data into scaled scores; and future prospects in test…
Descriptors: Scores, Statistics, Test Construction, Test Interpretation
Peer reviewed Peer reviewed
Irvine, S.H.; Reuning, H. – Journal of Cross-Cultural Psychology, 1981
Under experimental conditions, Canadian students were subjected to group tests on simple cognitive tasks constructed on the basis of previous theoretical work on perceptual speed. Cross-cultural replications suggest that such tests can be validated within and across cultures. (Author/MJL)
Descriptors: Adolescents, Cognitive Tests, Cross Cultural Studies, Perception
Peer reviewed Peer reviewed
Zumbo, Bruno D.; Pope, Gregory A.; Watson, Jackie E.; Hubley, Anita M. – Educational and Psychological Measurement, 1997
E. Roskam's (1985) conjecture that steeper item characteristic curve (ICC) "a" parameters (slopes) (and higher item total correlations in classical test theory) would be found with more concretely worded test items was tested with results from 925 young adults on the Eysenck Personality Questionnaire (H. Eysenck and S. Eysenck, 1975).…
Descriptors: Correlation, Personality Assessment, Personality Measures, Test Interpretation
Peer reviewed Peer reviewed
Chase, Clint – Mid-Western Educational Researcher, 1996
Classical procedures for calculating the two indices of decision consistency (P and Kappa) for criterion-referenced tests require two testings on each child. Huynh, Peng, and Subkoviak have presented one-testing procedures for these indices. These indices can be estimated without any test administration using Ebel's estimates of the mean, standard…
Descriptors: Criterion Referenced Tests, Educational Research, Educational Testing, Estimation (Mathematics)
Peer reviewed Peer reviewed
Embretson, Susan E. – Applied Psychological Measurement, 1996
Conditions under which interaction effects estimated from classical total scores, rather than item response theory trait scores, can be misleading are discussed with reference to analysis of variance (ANOVA). When no interaction effects exist on the true latent variable, spurious interaction effects can be observed from the total score scale. (SLD)
Descriptors: Analysis of Variance, Interaction, Item Response Theory, Models
Peer reviewed Peer reviewed
Humphreys, Lloyd G. – Applied Psychological Measurement, 1996
The reliability of a gain is determined by the reliabilities of the components, the correlation between them, and their standard deviations. Reliability is not inherently low, but the components of gains in many investigations make low reliability likely and require caution in the use of gain scores. (SLD)
Descriptors: Achievement Gains, Change, Correlation, Error of Measurement
Peer reviewed Peer reviewed
Williams, Richard H.; Zimmerman, Donald W. – Applied Psychological Measurement, 1996
The critiques by L. Collins and L. Humphreys in this issue illustrate problems with the use of gain scores. Collins' examples show that familiar formulas for the reliability of differences do not reflect the precision of measures of change. Additional examples demonstrate flaws in the conventional approach to reliability. (SLD)
Descriptors: Achievement Gains, Change, Correlation, Error of Measurement
Peer reviewed Peer reviewed
Fan, Xitao – Educational and Psychological Measurement, 1998
This study empirically examined the behaviors of item and person statistics derived from item response theory and classical test theory, focusing on item and person statistics and using a large-scale statewide assessment. Findings show that the person and item statistics from the two measurement frameworks are quite comparable. (SLD)
Descriptors: Item Response Theory, State Programs, Statistical Analysis, Test Items
Peer reviewed Peer reviewed
Banerji, Madhabi – Journal of Applied Measurement, 2000
Validated data from a developmental mathematics assessment using classical and three-faceted Rasch measurement methods. Analysis of field test data for 289 elementary school students suggested that a unidimensional construct was being measured, as defined by Rasch criteria. Discusses limitations in confirming content-related validity of the…
Descriptors: Construct Validity, Content Validity, Elementary Education, Elementary School Students
Peer reviewed Peer reviewed
Traub, Ross E. – Educational Measurement: Issues and Practice, 1997
Classical test theory is founded on the proposition that measurement error, a random latent variable, is a component of the observed score random variable. This article traces the history of the development of classical test theory, beginning in the early 20th century. (SLD)
Descriptors: Educational History, Educational Testing, Error of Measurement, Psychometrics
Peer reviewed Peer reviewed
Beland, Anne; Mislevy, Robert J. – Journal of Educational Measurement, 1996
This article addresses issues in model building and statistical inference in the context of student modeling. The use of probability-based reasoning to explicate hypothesized and empirical relationships and to structure inference in the context of proportional reasoning tasks is discussed. Ideas are illustrated with an example concerning…
Descriptors: Cognitive Psychology, Models, Networks, Probability
Peer reviewed Peer reviewed
Direct linkDirect link
MacCann, Robert G. – Psychometrika, 2004
For (0, 1) scored multiple-choice tests, a formula giving test reliability as a function of the number of item options is derived, assuming the "knowledge or random guessing model," the parallelism of the new and old tests (apart from the guessing probability), and the assumptions of classical test theory. It is shown that the formula is a more…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Reliability, Test Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Gest, Scott D.; Davidson, Alice J.; Rulison, Kelly L.; Moody, James; Welsh, Janet A. – New Directions for Child and Adolescent Development, 2007
The near universality of gender segregation in middle childhood and early adolescence has stimulated extensive research on sex differences in peer relationship processes. Recent reviews of the literature suggest that although some claims of two-cultures theory have clear empirical support, such as strong preference for same-sex peers over…
Descriptors: Early Adolescents, Peer Relationship, Friendship, Peer Groups
Peer reviewed Peer reviewed
Direct linkDirect link
Bush, Martin E. – Quality Assurance in Education: An International Perspective, 2006
Purpose: To provide educationalists with an understanding of the key quality issues relating to multiple-choice tests, and a set of guidelines for the quality assurance of such tests. Design/methodology/approach: The discussion of quality issues is structured to reflect the order in which those issues naturally arise. It covers the design of…
Descriptors: Multiple Choice Tests, Test Reliability, Educational Quality, Quality Control
Pages: 1  |  ...  |  30  |  31  |  32  |  33  |  34  |  35  |  36  |  37  |  38  |  ...  |  78