Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 23 |
Descriptor
Source
Author
Gallas, Edwin J. | 3 |
Herndon, Enid B. | 2 |
Aiken, Lewis R. | 1 |
Allen, Thomas E. | 1 |
Andrews, Jerrilyn V. | 1 |
Baldwin, Peter | 1 |
Barker, Pierce | 1 |
Bartram, Dave | 1 |
Borsboom, Denny | 1 |
Braun, Henry I. | 1 |
Bridgham, Robert G. | 1 |
More ▼ |
Publication Type
Education Level
Higher Education | 3 |
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Postsecondary Education | 2 |
Secondary Education | 2 |
Grade 5 | 1 |
High Schools | 1 |
Location
Pennsylvania | 3 |
Australia | 2 |
Canada | 1 |
Cyprus | 1 |
Hawaii | 1 |
Massachusetts | 1 |
South Africa | 1 |
South Carolina | 1 |
United Kingdom | 1 |
United States | 1 |
Laws, Policies, & Programs
Elementary and Secondary… | 3 |
No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
What Works Clearinghouse Rating
Baldwin, Peter; Clauser, Brian E. – Journal of Educational Measurement, 2022
While score comparability across test forms typically relies on common (or randomly equivalent) examinees or items, innovations in item formats, test delivery, and efforts to extend the range of score interpretation may require a special data collection before examinees or items can be used in this way--or may be incompatible with common examinee…
Descriptors: Scoring, Testing, Test Items, Test Format
Haertel, Edward H. – Educational Psychologist, 2018
In the service of educational accountability, student achievement tests are being used to measure constructs quite unlike those envisioned by test developers. Scores are compared to cut points to create classifications like "proficient"; scores are combined over time to measure growth; student scores are aggregated to measure the…
Descriptors: Achievement Tests, Scores, Test Validity, Test Interpretation
Dickens, Rachel H.; Meisinger, Elizabeth B.; Tarar, Jessica M. – Canadian Journal of School Psychology, 2015
The Comprehensive Test of Phonological Processing-Second Edition (CTOPP-2; Wagner, Torgesen, Rashotte, & Pearson, 2013) is a norm-referenced test that measures phonological processing skills related to reading for individuals aged 4 to 24. According to its authors, the CTOPP-2 may be used to identify individuals who are markedly below their…
Descriptors: Norm Referenced Tests, Phonology, Test Format, Testing
Sireci, Stephen G. – Journal of Educational Measurement, 2013
Kane (this issue) presents a comprehensive review of validity theory and reminds us that the focus of validation is on test score interpretations and use. In reacting to his article, I support the argument-based approach to validity and all of the major points regarding validation made by Dr. Kane. In addition, I call for a simpler, three-step…
Descriptors: Validity, Theories, Test Interpretation, Test Use
Kopriva, Rebecca J.; Thurlow, Martha L.; Perie, Marianne; Lazarus, Sheryl S.; Clark, Amy – Educational Psychologist, 2016
This article argues that test takers are as integral to determining validity of test scores as defining target content and conditioning inferences on test use. A principled sustained attention to how students interact with assessment opportunities is essential, as is a principled sustained evaluation of evidence confirming the validity or calling…
Descriptors: Tests, Testing, Test Interpretation, Scores
Borsboom, Denny – Measurement: Interdisciplinary Research and Perspectives, 2012
Paul E. Newton provides an insightful and scholarly overview of central issues in validity theory. As he notes, many of the conceptual problems in validity theory derive from the fact that the word "validity" has two meanings. First, it indicates "whether a test measures what it purports to measure." This is a factual claim about the psychometric…
Descriptors: Validity, Psychometrics, Test Interpretation, Scores
Powers, Sonya; Li, Dongmei; Suh, Hongwook; Harris, Deborah J. – ACT, Inc., 2016
ACT reporting categories and ACT Readiness Ranges are new features added to the ACT score reports starting in fall 2016. For each reporting category, the number correct score, the maximum points possible, the percent correct, and the ACT Readiness Range, along with an indicator of whether the reporting category score falls within the Readiness…
Descriptors: Scores, Classification, College Entrance Examinations, Error of Measurement
Kane, Michael – Journal of Educational Measurement, 2011
Errors don't exist in our data, but they serve a vital function. Reality is complicated, but our models need to be simple in order to be manageable. We assume that attributes are invariant over some conditions of observation, and once we do that we need some way of accounting for the variability in observed scores over these conditions of…
Descriptors: Error of Measurement, Scores, Test Interpretation, Testing
Davies, Alan – Language Testing, 2012
In this article, the author begins by discussing four challenges on the concept of validity. These challenges are: (1) the appeal to logic and syllogistic reasoning; (2) the claim of reliability; (3) the local and the universal; and (4) the unitary and the divisible. In language testing validity cannot be achieved directly but only through a…
Descriptors: Language Tests, Test Validity, Test Reliability, Testing
Michaelides, Michalis P. – Assessment in Education: Principles, Policy & Practice, 2014
Student examinees are key stakeholders in large-scale, high-stakes, public examination systems. How they perceive the purpose, comprehend the technical characteristics of testing and how they interpret scores influence their response to the system demands and their preparation for the examinations; this information relates to intended and…
Descriptors: Foreign Countries, National Competency Tests, High Stakes Tests, Student Attitudes
Pommerich, Mary – Educational Measurement: Issues and Practice, 2012
Neil Dorans has made a career of advocating for the examinee. He continues to do so in his NCME career award address, providing a thought-provoking commentary on some current trends in educational measurement that could potentially affect the integrity of test scores. Concerns expressed in the address call attention to a conundrum that faces…
Descriptors: Testing, Scores, Measurement, Test Construction
Hubley, Anita M.; Zumbo, Bruno D. – Social Indicators Research, 2011
The vast majority of measures have, at their core, a purpose of personal and social change. If test developers and users want measures to have personal and social consequences and impact, then it is critical to consider the consequences and side effects of measurement in the validation process itself. The consequential basis of test interpretation…
Descriptors: Construct Validity, Social Change, Measurement, Test Interpretation
Noble, Tracy; Suarez, Catherine; Rosebery, Ann; O'Connor, Mary Catherine; Warren, Beth; Hudicourt-Barnes, Josiane – Journal of Research in Science Teaching, 2012
Education policy in the U.S. in the last two decades has emphasized large-scale assessment of students, with growing consequences for schools, teachers, and students. Given the high stakes of such tests, it is important to understand the relationships between students' answers to test items and their knowledge and skills in the tested content…
Descriptors: Testing, Science Tests, Second Language Learning, Measures (Individuals)
Braun, Henry I.; Mislevy, Robert J. – US Department of Education, 2004
Psychologist Andrea diSessa coined the term "phenomenological primitives", or p-prims, to talk about nonexperts' reasoning about physical situations. P-prims are primitive in the sense that they stand without significant explanatory substructure or explanation. Examples are "Heavy objects fall faster than light objects" and "Continuing force is…
Descriptors: Test Theory, Testing, Evaluation Methods, Scores
Brown, James Dean – Language Assessment Quarterly, 2008
In keeping with the theme of the International Language Testing Association/Language Testing Research Colloquium Conference in 2008, "Focusing on the Core: Justifying the Use of Language Assessments to Stakeholders," I define "stakeholder-friendly tests," "defensible testing," and "testing-context analysis."…
Descriptors: Language Usage, Curriculum Development, Testing, Language Tests