NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 8 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Assessment in Education: Principles, Policy & Practice, 2016
How we choose to use a term depends on what we want to do with it. If "validity" is to be used to support a score interpretation, validation would require an analysis of the plausibility of that interpretation. If validity is to be used to support score uses, validation would require an analysis of the appropriateness of the proposed…
Descriptors: Test Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
This response to the comments contains three main sections, each addressing a subset of the comments. In the first section, I will respond to the comments by Brennan, Haertel, and Moss. All of these comments suggest ways in which my presentation could be extended or improved; I generally agree with their suggestions, so my response to their…
Descriptors: Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
To validate an interpretation or use of test scores is to evaluate the plausibility of the claims based on the scores. An argument-based approach to validation suggests that the claims based on the test scores be outlined as an argument that specifies the inferences and supporting assumptions needed to get from test responses to score-based…
Descriptors: Test Interpretation, Validity, Scores, Test Use
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Educational Researcher, 2008
Lissitz and Samuelsen (2007) have proposed an operational definition of "validity" that shifts many of the questions traditionally considered under validity to a separate category associated with the utility of test use. Operational definitions support inferences about how well people perform some kind of task or how they respond to some kind of…
Descriptors: Test Use, Definitions, Validity, Classification
Peer reviewed Peer reviewed
Subkoviak, Michael J.; Kane, Michael T.; Duncan, Patrick H. – Mid-Western Educational Researcher, 2002
Compares Angoff and Nedelsky methods for setting passing scores on tests. Using one of the methods, 84 college students were taught to estimate their probable scores on a vocabulary test. Estimates were compared to their later actual scores. The Nedelsky method was considerably less accurate under certain conditions, and both methods…
Descriptors: Cutting Scores, Difficulty Level, Evaluation Research, Test Construction
Peer reviewed Peer reviewed
Brennan, Robert L.; Kane, Michael T. – Psychometrika, 1977
Using the assumption of randomly parallel tests and concepts from generalizability theory, three signal/noise ratios for domain-referenced tests are developed, discussed, and compared. The three ratios have the same noise but different signals depending upon the kind of decision to be made as a result of measurement. (Author/JKS)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Error of Measurement, Mathematical Models
Kane, Michael T. – 1990
The literature on validity provides much more guidance on how to collect various kinds of validity evidence than it does on which kinds of evidence to collect in specific cases. An argument-based approach to validation redresses the balance by linking the kinds of evidence needed to validate a test-score interpretation to the details of the…
Descriptors: Evaluation Methods, Formative Evaluation, Inferences, Models
Kane, Michael T.; Brennan, Robert L. – 1977
A large number of seemingly diverse coefficients have been proposed as indices of dependability, or reliability, for domain-referenced and/or mastery tests. In this paper, it is shown that most of these indices are special cases of two generalized indices of agreement: one that is corrected for chance, and one that is not. The special cases of…
Descriptors: Bayesian Statistics, Correlation, Criterion Referenced Tests, Cutting Scores