NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Does not meet standards1
Showing 2,866 to 2,880 of 3,316 results Save | Export
Peer reviewed Peer reviewed
Levin, Joel R.; Subkoviak, Michael J. – Applied Psychological Measurement, 1977
Textbook calculations of statistical power or sample size follow from formulas that assume that the variables under consideration are measured without error. However, in the real world of behavioral research, errors of measurement cannot be neglected. The determination of sample size is discussed, and an example illustrates blocking strategy.…
Descriptors: Analysis of Covariance, Analysis of Variance, Error of Measurement, Hypothesis Testing
Peer reviewed Peer reviewed
Bannister, Brendan D.; And Others – Educational and Psychological Measurement, 1987
To control for response bias in student ratings of college teachers, an index of rater error was used that was theoretically independent of actual performance. Partialing out the effects of this extraneous response bias enhanced validity, but partialing out overall effectiveness resulted in reduced convergent and discriminant validities.…
Descriptors: Error of Measurement, Higher Education, Interrater Reliability, Response Style (Tests)
Peer reviewed Peer reviewed
Block, Jack; And Others – Developmental Psychology, 1986
As part of a longitudinal study, Matching Familiar Figures Test (MFFT) performance was assessed at age 11 and related to California Child Q-Sort evaluations obtained both concurrently and at age 14. Offers evidence for a "competence" interpretation rather than a "conceptual tempo" interpretation of the MFFT and strongly…
Descriptors: Adolescents, Children, Conceptual Tempo, Error of Measurement
Peer reviewed Peer reviewed
Meskauskas, John A. – Evaluation and the Health Professions, 1986
Two new indices of stability of content-referenced standard-setting results are presented, relating variability of judges' decisions to the variability of candidate scores and to the reliability of the test. These indices are used to indicate whether scores resulting from a standard-setting study are of sufficient precision. (Author/LMO)
Descriptors: Certification, Credentials, Error of Measurement, Generalizability Theory
Peer reviewed Peer reviewed
Salin, Eric D. – Journal of Chemical Education, 1984
Describes an experiment designed to teach students to apply the same statistical awareness to instrumentation they commonly apply to classical techniques. Uses propagation of error techniques to pinpoint instrumental limitations and breakdowns and to demonstrate capabilities and limitations of volumetric and gravimetric methods. Provides lists of…
Descriptors: Chemistry, College Science, Electronic Equipment, Equipment Standards
Peer reviewed Peer reviewed
Yen, Wendy M. – Journal of Educational Measurement, 1984
A procedure for obtaining maximum likelihood trait estimates from number-correct (NC) scores for the three-parameter logistic model is presented. It produces an NC score to trait estimate conversion table. Analyses in the estimated true score metric confirm the conclusions made in the trait metric. (Author/DWH)
Descriptors: Achievement Tests, Error of Measurement, Estimation (Mathematics), Latent Trait Theory
Basol-Gocmen, Gulsah; Kanyongo, Gibbs Y.; Blankson, Lydia – Online Submission, 2002
The purpose of this paper is to evaluate the use of MC2G program to teach certain topics in statistics education. MC2G is a program written in Pascal Delphi by Gordon Brooks of Ohio University based on Monte Carlo studies. MC2G provides students opportunity to practice important topics in an introductory statistics course, such as power, Type I…
Descriptors: Student Attitudes, Monte Carlo Methods, Computer Software, Effect Size
Karkee, Thakur B.; Wright, Karen R. – Online Submission, 2004
Different item response theory (IRT) models may be employed for item calibration. Change of testing vendors, for example, may result in the adoption of a different model than that previously used with a testing program. To provide scale continuity and preserve cut score integrity, item parameter estimates from the new model must be linked to the…
Descriptors: Measures (Individuals), Evaluation Criteria, Testing, Integrity
Ross, J. Michael – 1996
This paper presents a number of arguments for the increased importance of within-state district-level data in systematic assessments in the organizational structure of schools as educational institutions. The major question is whether the Schools and Staffing Survey (SASS) should shift its focus toward more macro-institutional district level…
Descriptors: Adult Education, Elementary Secondary Education, Error of Measurement, Evaluation Methods
Zwick, Rebecca; And Others – 1994
A simulation study of methods of assessing differential item functioning (DIF) in computer-adaptive tests (CATs) was conducted by Zwick, Thayer, and Wingersky (in press, 1993). Results showed that modified versions of the Mantel-Haenszel and standardization methods work well with CAT data. DIF methods were also investigated for nonadaptive…
Descriptors: Adaptive Testing, Computer Assisted Testing, Error of Measurement, Estimation (Mathematics)
Safarik, John Gerald – California Journal of Educational Research, 1972
Study analyzed the results of a college rule regulating the maximum number of units students were allowed to carry and its effect on academic performance. (Author/RK)
Descriptors: Academic Achievement, Academic Failure, College Students, Credits
Peer reviewed Peer reviewed
Emrick, John A. – Journal of Educational Measurement, 1971
Descriptors: Criterion Referenced Tests, Error of Measurement, Evaluation Methods, Item Analysis
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1972
A reliability coefficient for criterion-referenced tests is developed from the assumptions of classical test theory. The coefficient is based on deviations of scores from the criterion score, rather than from the mean. (Author/CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Mathematical Applications, Norm Referenced Tests
Peer reviewed Peer reviewed
Harris, Chester W. – Journal of Educational Measurement, 1972
An alternative interpretation of Livingston's reliability coefficient (see TM 500 487) is based on the notion of the relation of the size of the reliability coefficient to the range of talent. (Author/CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Mathematical Applications, Norm Referenced Tests
Peer reviewed Peer reviewed
McGaw, Barry; And Others – American Educational Research Journal, 1972
The generalizability theory approach to the estimation of reliability is outlined, and a design is developed in which systematic variations in behavior over differing situations are separated from random fluctuation. Three coefficients of reliability are proposed. (CK)
Descriptors: Analysis of Variance, Behavior Change, Classroom Observation Techniques, Classroom Research
Pages: 1  |  ...  |  188  |  189  |  190  |  191  |  192  |  193  |  194  |  195  |  196  |  ...  |  222