Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 6 |
Since 2006 (last 20 years) | 7 |
Descriptor
Test Interpretation | 24 |
Test Validity | 19 |
Scores | 10 |
Testing Problems | 10 |
Test Use | 7 |
Achievement Tests | 6 |
Educational Assessment | 6 |
Standardized Tests | 6 |
Elementary Secondary Education | 5 |
Evaluation Methods | 5 |
Validity | 5 |
More ▼ |
Source
Educational Measurement:… | 24 |
Author
Krupa, Erin E. | 2 |
Linn, Robert L. | 2 |
Mehrens, William A. | 2 |
An, Lily Shiao | 1 |
Bachman, Lyle F. | 1 |
Bostic, Jonathan | 1 |
Bostic, Jonathan D. | 1 |
Carney, Michele B. | 1 |
Chinen, Starlie | 1 |
Cizek, Gregory J. | 1 |
Cohen, Allan | 1 |
More ▼ |
Publication Type
Journal Articles | 24 |
Opinion Papers | 8 |
Reports - Research | 7 |
Reports - Evaluative | 5 |
Reports - Descriptive | 4 |
Information Analyses | 2 |
Speeches/Meeting Papers | 2 |
Guides - Non-Classroom | 1 |
Tests/Questionnaires | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
What Works Clearinghouse Rating
Folger, Timothy D.; Bostic, Jonathan; Krupa, Erin E. – Educational Measurement: Issues and Practice, 2023
Validity is a fundamental consideration of test development and test evaluation. The purpose of this study is to define and reify three key aspects of validity and validation, namely test-score interpretation, test-score use, and the claims supporting interpretation and use. This study employed a Delphi methodology to explore how experts in…
Descriptors: Test Interpretation, Scores, Test Use, Test Validity
Ing, Marsha; Chinen, Starlie; Jackson, Kara; Smith, Thomas M. – Educational Measurement: Issues and Practice, 2021
Despite the ease of accessing a wide range of measures, little attention is given to validity arguments when considering whether to use the measure for a new purpose or in a different context. Making a validity argument has historically focused on the intended interpretation and use. There has been a press to consider both the intended and actual…
Descriptors: Instructional Improvement, Measures (Individuals), Test Validity, Test Interpretation
An, Lily Shiao; Ho, Andrew Dean; Davis, Laurie Laughlin – Educational Measurement: Issues and Practice, 2022
Technical documentation for educational tests focuses primarily on properties of individual scores at single points in time. Reliability, standard errors of measurement, item parameter estimates, fit statistics, and linking constants are standard technical features that external stakeholders use to evaluate items and individual scale scores.…
Descriptors: Documentation, Scores, Evaluation Methods, Longitudinal Studies
Lavery, Matthew Ryan; Bostic, Jonathan D.; Kruse, Lance; Krupa, Erin E.; Carney, Michele B. – Educational Measurement: Issues and Practice, 2020
Since it was formalized by Kane, the argument-based approach to validation has been promoted as the preferred method for validating interpretations and uses of test scores. Because validation is discussed in terms of arguments, and arguments are both interactive and social, the present review systematically examines the scholarly arguments which…
Descriptors: Persuasive Discourse, Validity, Research Methodology, Peer Evaluation
O'Leary, Timothy M.; Hattie, John A. C.; Griffin, Patrick – Educational Measurement: Issues and Practice, 2017
Validity is the most fundamental consideration in test development. Understandably, much time, effort, and money is spent in its pursuit. Central to the modern conception of validity are the interpretations made, and uses planned, on the basis of test scores. There is, unfortunately, however, evidence that test users have difficulty understanding…
Descriptors: Test Interpretation, Scores, Test Validity, Evidence
Wise, Steven L. – Educational Measurement: Issues and Practice, 2017
The rise of computer-based testing has brought with it the capability to measure more aspects of a test event than simply the answers selected or constructed by the test taker. One behavior that has drawn much research interest is the time test takers spend responding to individual multiple-choice items. In particular, very short response…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Items, Reaction Time

Messick, Samuel – Educational Measurement: Issues and Practice, 1995
Six distinguishable aspects of construct validity are discussed as they apply to performance assessment, emphasizing content, substantive, structural, generalizability, external, and consequential aspects. Taken together, these aspects provide a way to address validity questions in score interpretation and use. (SLD)
Descriptors: Construct Validity, Content Validity, Educational Assessment, Generalization

Kane, Michael; Crooks, Terence; Cohen, Allan – Educational Measurement: Issues and Practice, 1999
Analyzes the three major inferences involved in interpretation of performance assessments: (1) scoring of the observed performances; (2) generalization to a domain of assessment performances like those included in the assessment; and (3) extrapolation to the large performance domain of interest. Suggests ways to improve the validity of performance…
Descriptors: Performance Based Assessment, Performance Factors, Scoring, Test Interpretation

Bachman, Lyle F. – Educational Measurement: Issues and Practice, 2002
Describes an approach to addressing issues of validity of inferences and the extrapolation of inferences to target domains beyond the assessment for alternative assessments. Makes the case that in both language testing and educational assessment the roles of language and content knowledge must be considered, and that the design and development of…
Descriptors: Alternative Assessment, Educational Assessment, Inferences, Performance Based Assessment

Sicoly, Fiore – Educational Measurement: Issues and Practice, 2002
Studied the validity of interpretations derived from school-level assessment scores for different content areas using school-level data from 14 large-scale assessment programs. Results raise suspicions about the value and validity of interpretations based on school-level subject area scores. (SLD)
Descriptors: Accountability, Elementary Secondary Education, Large Scale Assessment, Schools

Cronbach, Lee J. – Educational Measurement: Issues and Practice, 1983
In response to Ebel's article (TM 508 146) the author points out that, in emphasizing what a test is intended to measure, the vital question of what should be measured is set aside. While explicit specifications for a test should be encouraged, questions of worth ought to be prominent in designing and selecting tests. (Author/LC)
Descriptors: Educational Objectives, Measurement Objectives, Predictive Measurement, Test Construction
Cizek, Gregory J.; Crocker, Linda; Frisbie, David A.; Mehrens, William A.; Stiggins, Richard J. – Educational Measurement: Issues and Practice, 2006
The authors describe the significant contributions of Robert Ebel to educational measurement theory and its applications. A biographical sketch details Ebel's roots and professional resume. His influence on classroom assessment views and procedures are explored. Classic publications associated with validity, reliability, and score interpretation…
Descriptors: Test Theory, Educational Assessment, Psychometrics, Test Reliability

Linn, Robert L. – Educational Measurement: Issues and Practice, 1997
It is argued that consequential validity is a concept worth considering. The solution to defining "validity" is not to narrow the concept, but to allow for the differential prediction provided by tests in different circumstances. Consequences of the uses and interpretations of test scores are central to their evaluation. (SLD)
Descriptors: Educational Assessment, Educational Testing, Elementary Secondary Education, Evaluation Methods

Oosterhof, Albert C.; Salisbury, David F. – Educational Measurement: Issues and Practice, 1985
The computer assisted testing (CAT) program at Florida State University's Assessment Resource Center is described. Three measurement issues (test quality, confidence in mastery decisions, and maintenance of test validity) and two instructional design issues (quality of instructional objectives and extended feedback following testing) important for…
Descriptors: Computer Assisted Testing, Educational Objectives, Feedback, Higher Education

Wilson, Sandra Meachan; Hiscox, Michael D. – Educational Measurement: Issues and Practice, 1984
This article presents a model that can be used by local school districts for reanalyzing standardized test results to obtain a more valid assessment of local learning objectives can be used to identify strengths/weaknesses of existing programs as well as individual students. (EGS)
Descriptors: Educational Objectives, Item Analysis, Models, School Districts
Previous Page | Next Page ยป
Pages: 1 | 2