NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 571 to 585 of 1,161 results Save | Export
Peer reviewed Peer reviewed
Stevens, Joseph J.; Aleamoni, Lawrence, M. – Educational and Psychological Measurement, 1986
Prior standardization of scores when an aggregate score is formed has been criticized. This article presents a demonstration of the effects of differential weighting of aggregate components that clarifies the need for prior standardization. The role of standardization in statistics and the use of aggregate scores in research are discussed.…
Descriptors: Correlation, Error of Measurement, Factor Analysis, Raw Scores
Peer reviewed Peer reviewed
Andersen, Erling B. – Psychometrika, 1985
A model for longitudinal latent structure analysis was proposed that combined the values of a latent variable at two time points in a two-dimensional latent density. The correlation coefficient between the two values of the latent variable can then be estimated. (NSF)
Descriptors: Correlation, Latent Trait Theory, Mathematical Models, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Purves, Alan C. – Language Arts, 1986
Explores the three broad thrusts of literature curricula, noting that no test can cover all three. Discusses how the content and objectives of the literature curriculum can be specified and how test questions can be developed for the evaluation of literature comprehension. (HTH)
Descriptors: Curriculum Problems, Elementary Education, Language Arts, Literature Appreciation
Peer reviewed Peer reviewed
Yen, Wendy M. – Psychometrika, 1983
Tau-equivalence means that two tests produce equal true scores for individuals but that the distribution of errors for the tests could be different. This paper examines the effect of performing equipercentile equating techniques on tau-equivalent tests. (JKS)
Descriptors: Equated Scores, Latent Trait Theory, Psychometrics, Scores
Peer reviewed Peer reviewed
Bieliauskas, Vytautas J.; Farragher, John – Journal of Clinical Psychology, 1983
Administered the House-Tree-Person test to male college students (N=24) to examine the effects of varying the size of the drawing form on the scores. Results suggested that use of the drawing sheet did not have a significant influence upon the quantitative aspects of the drawing. (LLL)
Descriptors: College Students, Higher Education, Intelligence Tests, Males
Peer reviewed Peer reviewed
Masters, Geofferey N. – Educational and Psychological Measurement, 1984
DICOT, a computer program for the Rasch analysis of classroom tests, is described. Results are presented in a self-explanatory form. Person ability and item difficulty estimates are expressed in a familiar metric. Person and item fit statistics provide a diagnosis of individual children and identification of problematic items. (Author/DWH)
Descriptors: Classroom Techniques, Foreign Countries, Item Analysis, Latent Trait Theory
Peer reviewed Peer reviewed
Stegelmann, Werner – Psychometrika, 1983
The Rasch model is generalized to a multicomponent model, so that observations of component events are not needed to apply the model. It is shown that the generalized model maintains the property of specific objectivity of the Rasch model. An application to a mathematics test is provided. (Author/JKS)
Descriptors: Estimation (Mathematics), Item Analysis, Latent Trait Theory, Mathematical Models
Peer reviewed Peer reviewed
Williams, Richard H.; Zimmerman, Donald W. – Journal of Experimental Education, 1982
The reliability of simple difference scores is greater than, less than, or equal to that of residualized difference scores, depending on whether the correlation between pretest and posttest scores is greater than, less than, or equal to the ratio of the standard deviations of pretest and posttest scores. (Author)
Descriptors: Achievement Gains, Comparative Analysis, Correlation, Pretests Posttests
Peer reviewed Peer reviewed
Gravett, Sarah – South African Journal of Higher Education, 1996
Argues that student assessment plays a crucial role in the academic life of college students, and assessment arrangements embody the purposes of higher education. Reviews research suggesting learners' perceptions of course testing procedures is the single most important influence on learning. Outlines six guiding principles of test development to…
Descriptors: College Instruction, Educational Objectives, Higher Education, Student Evaluation
Peer reviewed Peer reviewed
Mislevy, Robert J. – Journal of Educational Measurement, 1996
Developments in cognitive and developmental psychology have broadened the inferences researchers want to make about students' learning and the nature and acquisition of knowledge. The principles of inference that led to standard test theory can support inference in the broader context of the cognitive revolution. (SLD)
Descriptors: Cognitive Psychology, Developmental Psychology, Educational Assessment, Educational Research
Peer reviewed Peer reviewed
Klinger, Don A.; Rogers, W. Todd – Alberta Journal of Educational Research, 2003
The estimation accuracy of procedures based on classical test score theory and item response theory (generalized partial credit model) were compared for examinations consisting of multiple-choice and extended-response items. Analysis of British Columbia Scholarship Examination results found an error rate of about 10 percent for both methods, with…
Descriptors: Academic Achievement, Educational Testing, Foreign Countries, High Stakes Tests
Peer reviewed Peer reviewed
Andrich, David – Psychometrika, 1995
This book discusses adapting pencil-and-paper tests to computerized testing. Mention is made of models for graded responses to items and of possibilities beyond pencil-and-paper-tests, but the book is essentially about dichotomously scored test items. Contrasts between item response theory and classical test theory are described. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Response Theory, Scores
Peer reviewed Peer reviewed
Little, Roderick J. A.; Rubin, Donald B. – Journal of Educational and Behavioral Statistics, 1994
Equating a new standard test to an old reference test is considered when samples for equating are not randomly selected from the target population of test takers, identifying two problems from equating from biased samples. An empirical example with data from the Armed Services Vocational Aptitude Battery illustrates the approach. (SLD)
Descriptors: Equated Scores, Military Personnel, Sampling, Statistical Analysis
Peer reviewed Peer reviewed
Reuterberg, Sven-Eric; Gustafsson, Jan-Eric – Educational and Psychological Measurement, 1992
The use of confirmatory factor analysis by the LISREL program is demonstrated as an assumption-testing method when computing reliability coefficients under different model assumptions. Results indicate that reliability estimates are robust against departure from the assumption of parallelism of test items. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Mathematical Models, Robustness (Statistics)
Peer reviewed Peer reviewed
Brown, James Dean – Language Testing, 1999
Explored the relative contributions to Test of English as a Foreign Language (TOEFL) score dependability of various numbers of persons, items, subtests, languages, and their various interactions. Sampled 15,000 test takers, 1000 each from 15 different language backgrounds. (Author/VWL)
Descriptors: English (Second Language), Language Tests, Second Language Learning, Student Characteristics
Pages: 1  |  ...  |  35  |  36  |  37  |  38  |  39  |  40  |  41  |  42  |  43  |  ...  |  78