NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 44 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Shun-Fu Hu; Amery D. Wu; Jake Stone – Journal of Educational Measurement, 2025
Scoring high-dimensional assessments (e.g., > 15 traits) can be a challenging task. This paper introduces the multilabel neural network (MNN) as a scoring method for high-dimensional assessments. Additionally, it demonstrates how MNN can score the same test responses to maximize different performance metrics, such as accuracy, recall, or…
Descriptors: Tests, Testing, Scores, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Kyung Yong; Lee, Won-Chan – Journal of Educational Measurement, 2018
Reporting confidence intervals with test scores helps test users make important decisions about examinees by providing information about the precision of test scores. Although a variety of estimation procedures based on the binomial error model are available for computing intervals for test scores, these procedures assume that items are randomly…
Descriptors: Weighted Scores, Error of Measurement, Test Use, Decision Making
Peer reviewed Peer reviewed
Direct linkDirect link
Trierweiler, Tammy J.; Lewis, Charles; Smith, Robert L. – Journal of Educational Measurement, 2016
In this study, we describe what factors influence the observed score correlation between an (external) anchor test and a total test. We show that the anchor to full-test observed score correlation is based on two components: the true score correlation between the anchor and total test, and the reliability of the anchor test. Findings using an…
Descriptors: Scores, Correlation, Tests, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Mislevy, Robert J. – Journal of Educational Measurement, 2016
Validity is the sine qua non of properties of educational assessment. While a theory of validity and a practical framework for validation has emerged over the past decades, most of the discussion has addressed familiar forms of assessment and psychological framings. Advances in digital technologies and in cognitive and social psychology have…
Descriptors: Test Validity, Technology, Cognitive Psychology, Social Psychology
Peer reviewed Peer reviewed
Direct linkDirect link
Newton, Paul E. – Journal of Educational Measurement, 2013
Kane distinguishes between two kinds of argument: the interpretation/use argument and the validity argument. This commentary considers whether there really are two kinds of argument, two arguments, or just one. It concludes that there is just one argument: the validity argument. (Contains 2 figures and 5 notes.)
Descriptors: Validity, Test Interpretation, Test Use
Peer reviewed Peer reviewed
Direct linkDirect link
Haertel, Edward – Journal of Educational Measurement, 2013
In validating uses of testing, it is helpful to distinguish those that rely directly on the information provided by scores or score distributions ("direct" uses and consequences) versus those that instead capitalize on the motivational effects of testing, or use testing and test reporting to shape public opinion ("indirect" uses and consequences).…
Descriptors: Validity, Testing, Test Results, Test Use
Peer reviewed Peer reviewed
Direct linkDirect link
Sireci, Stephen G. – Journal of Educational Measurement, 2013
Kane (this issue) presents a comprehensive review of validity theory and reminds us that the focus of validation is on test score interpretations and use. In reacting to his article, I support the argument-based approach to validity and all of the major points regarding validation made by Dr. Kane. In addition, I call for a simpler, three-step…
Descriptors: Validity, Theories, Test Interpretation, Test Use
Peer reviewed Peer reviewed
Direct linkDirect link
Borsboom, Denny; Markus, Keith A. – Journal of Educational Measurement, 2013
According to Kane (this issue), "the validity of a proposed interpretation or use depends on how well the evidence supports" the claims being made. Because truth and evidence are distinct, this means that the validity of a test score interpretation could be high even though the interpretation is false. As an illustration, we discuss the case of…
Descriptors: Evidence, Ethics, Validity, Theories
Peer reviewed Peer reviewed
Direct linkDirect link
Brennan, Robert L. – Journal of Educational Measurement, 2013
Kane's paper "Validating the Interpretations and Uses of Test Scores" is the most complete and clearest discussion yet available of the argument-based approach to validation. At its most basic level, validation as formulated by Kane is fundamentally a simply-stated two-step enterprise: (1) specify the claims inherent in a particular interpretation…
Descriptors: Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
This response to the comments contains three main sections, each addressing a subset of the comments. In the first section, I will respond to the comments by Brennan, Haertel, and Moss. All of these comments suggest ways in which my presentation could be extended or improved; I generally agree with their suggestions, so my response to their…
Descriptors: Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Moss, Pamela A. – Journal of Educational Measurement, 2013
Studies of data use illuminate ways in which education professionals have used test scores and other evidence relevant to students' learning--in action in their own contexts of work--to make decisions about their practice. These studies raise instructive challenges for a validity theory that focuses on intended interpretations and uses of test…
Descriptors: Validity, Test Use, Test Interpretation, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
To validate an interpretation or use of test scores is to evaluate the plausibility of the claims based on the scores. An argument-based approach to validation suggests that the claims based on the test scores be outlined as an argument that specifies the inferences and supporting assumptions needed to get from test responses to score-based…
Descriptors: Test Interpretation, Validity, Scores, Test Use
Peer reviewed Peer reviewed
Caruso, John C.; Witkiewitz, Katie – Journal of Educational Measurement, 2002
As an alternative to equally weighted difference scores, examined an orthogonal reliable component analysis (RCA) solution and an oblique principal components analysis (PCA) solution for the standardization sample of the Kaufman Assessment Battery for Children (KABC; A. Kaufman and N. Kaufman, 1983). Discusses the practical implications of the…
Descriptors: Ability, Academic Achievement, Children, Factor Analysis
Peer reviewed Peer reviewed
Cole, Nancy S.; Zieky, Michael J. – Journal of Educational Measurement, 2001
Proposes additional ways for people in the measurement profession to think about the fairness of assessments and about the fairness of the uses of assessments. Suggests that measurement professionals must pay more attention to reducing group differences at the design stage of test development, to providing all examinees an opportunity to…
Descriptors: Educational Testing, Equal Education, Groups, Test Bias
Peer reviewed Peer reviewed
Hosenfeld, Bettina; van den Boom, Dymphna C.; Resing, Wilma C. M. – Journal of Educational Measurement, 1997
Reports the development of a new geometric analogies test for elementary school children and the examination of its use for future longitudinal studies. Three studies involving 361 elementary school students indicate the quality of the scale and support its use. (SLD)
Descriptors: Elementary Education, Elementary School Students, Geometry, Longitudinal Studies
Previous Page | Next Page ยป
Pages: 1  |  2  |  3