Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 9 |
Descriptor
Models | 38 |
Test Use | 38 |
Test Validity | 28 |
Educational Assessment | 14 |
Evaluation Methods | 12 |
Test Reliability | 12 |
Test Construction | 10 |
Foreign Countries | 9 |
Test Interpretation | 9 |
Test Items | 8 |
Validity | 8 |
More ▼ |
Source
Author
Clark, John L. D. | 2 |
Ackerman, Terry A. | 1 |
Adams, Elizabeth | 1 |
Bartram, Dave | 1 |
Bennett, Randy Elliot | 1 |
Benoit, Joyce | 1 |
Casey, Leo M. | 1 |
Coyne, Iain | 1 |
Crooks, Terry | 1 |
Daniel, Larry G. | 1 |
DeLuca, Christopher | 1 |
More ▼ |
Publication Type
Education Level
Elementary Secondary Education | 3 |
Secondary Education | 2 |
Grade 9 | 1 |
Higher Education | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Postsecondary Education | 1 |
Audience
Practitioners | 2 |
Administrators | 1 |
Researchers | 1 |
Laws, Policies, & Programs
Education Consolidation… | 1 |
Assessments and Surveys
ACTFL Oral Proficiency… | 1 |
Motivated Strategies for… | 1 |
National Assessment of… | 1 |
SAT (College Admission Test) | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Ketterlin-Geller, Leanne R.; Perry, Lindsey; Adams, Elizabeth – Applied Measurement in Education, 2019
Despite the call for an argument-based approach to validity over 25 years ago, few examples exist in the published literature. One possible explanation for this outcome is that the complexity of the argument-based approach makes implementation difficult. To counter this claim, we propose that the Assessment Triangle can serve as the overarching…
Descriptors: Validity, Educational Assessment, Models, Screening Tests
Oliveri, María Elena; Rutkowski, David; Rutkowski, Lesli – ETS Research Report Series, 2018
Fifty years after the first international large-scale assessment (ILSA), participation in these studies continues to grow, with more than 50% of the world's countries participating. Concomitant with growth in ILSAs is an expansion in the diversity of participant countries with respect to languages, cultures, and educational perspectives and goals.…
Descriptors: International Assessment, Test Validity, Test Use, Alignment (Education)
Stone, Elizabeth; Wylie, E. Caroline – ETS Research Report Series, 2019
We describe the summative assessment component within a K-12 assessment program and our development of a validity argument to support its claims with respect to intended uses and interpretations. First, we describe the "Winsight"® assessment program theory of action, a logic model elucidating mechanisms for how use of the assessment…
Descriptors: Summative Evaluation, Educational Assessment, Test Validity, Test Use
Kane, Michael T. – Assessment in Education: Principles, Policy & Practice, 2016
How we choose to use a term depends on what we want to do with it. If "validity" is to be used to support a score interpretation, validation would require an analysis of the plausibility of that interpretation. If validity is to be used to support score uses, validation would require an analysis of the appropriateness of the proposed…
Descriptors: Test Validity, Test Interpretation, Test Use, Scores
Casey, Leo M. – Teachers College Record, 2013
Background/Context: There is a deep and yawning chasm between the world of tests and testing practices as they ought to be and the actual tests and testing practices now imposed on American students, educators, and schools. That chasm of theory and practice is a function of the dominant paradigm of educational reform, with its theory of action…
Descriptors: Educational Change, Commercialization, Models, Test Use
Koch, Martha J.; DeLuca, Christopher – Assessment in Education: Principles, Policy & Practice, 2012
In this article we rethink validation within the complex contexts of high-stakes assessment. We begin by considering the utility of existing models for validation and argue that these models tend to overlook some of the complexities inherent to assessment use, including the multiple interpretations of assessment purposes and the potential…
Descriptors: Foreign Countries, Test Use, Case Studies, Educational Assessment
Hubley, Anita M.; Zumbo, Bruno D. – Social Indicators Research, 2011
The vast majority of measures have, at their core, a purpose of personal and social change. If test developers and users want measures to have personal and social consequences and impact, then it is critical to consider the consequences and side effects of measurement in the validation process itself. The consequential basis of test interpretation…
Descriptors: Construct Validity, Social Change, Measurement, Test Interpretation
McNamara, Tim – Language Assessment Quarterly, 2006
The thought of Samuel Messick has influenced language testing in 2 main ways: in proposing a new understanding of how inferences made based on tests must be challenged, and in drawing attention to the consequences of test use. The former has had a powerful impact on language-testing research, most notably in Bachman's work on validity and the…
Descriptors: Test Use, Testing, Language Tests, Validity

Glutting, Joseph J. – Journal of School Psychology, 1989
Introduces Stanford-Binet Intelligence Scale-Fourth Edition (SB4) as an attempt to revitalize Stanford-Binet by maintaining links with previous editions while simultaneously incorporating more recent developments found in other popular tests of intelligence. Discusses the SB4's theoretical foundation, materials and administration, scaling,…
Descriptors: Intelligence Tests, Models, Test Reliability, Test Use

Coyne, Iain; Bartram, Dave – International Journal of Testing, 2002
Applied a structured framework to the research literature to assess the general effectiveness of integrity tests as used to assess "honesty" in preemployment screening. In terms of acceptability, research has indicated that integrity tests are viewed neither positively nor negatively, but considerable debate remains about false positive rates and…
Descriptors: Honesty, Integrity, Literature Reviews, Models

Wardrop, James L.; And Others – Journal of Educational Measurement, 1982
A structure for describing different approaches to testing is generated by identifying five dimensions along which tests differ: test uses, item generation, item revision, assessment of precision, and validation. These dimensions are used to profile tests of reading comprehension. Only norm-referenced achievement tests had an inference system…
Descriptors: Achievement Tests, Comparative Analysis, Educational Testing, Models

Ackerman, Terry A. – Applied Measurement in Education, 1994
When item response data do not satisfy the unidimensionality assumption, multidimensional item response theory (MIRT) should be used to model the item-examinee interaction. This article presents and discusses MIRT analyses designed to give better insight into what individual items are measuring. (SLD)
Descriptors: Evaluation Methods, Item Response Theory, Measurement Techniques, Models
Grandy, Jerilee – 1992
This study investigated the construct validity and population generalizability of the NTE Core Battery using two different data bases and several analytical designs. Part I of the study used data from the November 1982 administration of the Core Battery at the subscore level for 5,183 examinees. Part II used data from the October 1985…
Descriptors: College Students, Construct Validity, Evaluation Methods, Factor Structure
Leclercq, Dieudonne – Evaluation in Education: An International Review Series, 1982
In a confidence weighting situation, the examinee is asked to indicate the correct answer, and how certain he or she is of the correctness of that answer. This paper reviews the bases for confidence marking, its validity and accuracy in evaluating students, and it's use in research. (BW)
Descriptors: Confidence Testing, Educational Research, Measurement Techniques, Models
Bennett, Randy Elliot – 1990
A new assessment conception is described that integrates constructed-response testing, artificial intelligence, and model-based measurement. The conception incorporates complex constructed-response items for their potential to increase the validity, instructional utility, and credibility of standardized tests. Artificial intelligence methods are…
Descriptors: Artificial Intelligence, Constructed Response, Educational Assessment, Measurement Techniques