Publication Date
In 2025 | 1 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 11 |
Descriptor
Scores | 23 |
Test Theory | 23 |
Testing | 23 |
Comparative Analysis | 6 |
Test Interpretation | 6 |
Test Reliability | 5 |
Test Validity | 5 |
Achievement Tests | 4 |
Evaluation Methods | 4 |
High Stakes Tests | 4 |
Test Items | 4 |
More ▼ |
Source
Author
Mislevy, Robert J. | 2 |
Bailey, Janelle M. | 1 |
Braun, Henry I. | 1 |
Brooks, Lindsay | 1 |
Burton, Robert S. | 1 |
Cohen, Allan S., Comp. | 1 |
Daniel M. Settlage | 1 |
Darrah, Marjorie | 1 |
Davis, John L. | 1 |
Fuller, Edgar | 1 |
Green, Bert F. | 1 |
More ▼ |
Publication Type
Education Level
Higher Education | 3 |
Postsecondary Education | 1 |
Audience
Practitioners | 2 |
Policymakers | 1 |
Laws, Policies, & Programs
Elementary and Secondary… | 3 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
National Assessment of… | 1 |
What Works Clearinghouse Rating
Kylie Gorney; Sandip Sinharay – Educational and Psychological Measurement, 2025
Test-takers, policymakers, teachers, and institutions are increasingly demanding that testing programs provide more detailed feedback regarding test performance. As a result, there has been a growing interest in the reporting of subscores that potentially provide such detailed feedback. Haberman developed a method based on classical test theory…
Descriptors: Scores, Test Theory, Test Items, Testing
Daniel M. Settlage; Jim R. Wollscheid – Journal of the Scholarship of Teaching and Learning, 2024
The examination of the testing mode effect has received increased attention as higher education has shifted to remote testing during the COVID-19 pandemic. We believe the testing mode effect consists of four components: the ability to physically write on the test, the method of answer recording, the proctoring/testing environment, and the effect…
Descriptors: College Students, Macroeconomics, Tests, Answer Sheets
Soland, James; Kuhfeld, Megan – Educational Assessment, 2019
Considerable research has examined the use of rapid guessing measures to identify disengaged item responses. However, little is known about students who rapidly guess over the course of several tests. In this study, we use achievement test data from six administrations over three years to investigate whether rapid guessing is a stable trait-like…
Descriptors: Testing, Guessing (Tests), Reaction Time, Achievement Tests
Bailey, Janelle M.; Johnson, Bruce; Prather, Edward E.; Slater, Timothy F. – International Journal of Science Education, 2012
Concept inventories (CIs)--typically multiple-choice instruments that focus on a single or small subset of closely related topics--have been used in science education for more than a decade. This paper describes the development and validation of a new CI for astronomy, the "Star Properties Concept Inventory" (SPCI). Questions cover the areas of…
Descriptors: Educational Strategies, Validity, Testing, Astronomy
Mislevy, Robert J. – Educational Measurement: Issues and Practice, 2012
This article presents the author's observations on Neil Dorans's NCME Career Award Address: "The Contestant Perspective on Taking Tests: Emanations from the Statue within." He calls attention to some points that Dr. Dorans made in his address, and offers his thoughts in response.
Descriptors: Testing, Test Reliability, Psychometrics, Scores
Kettler, Ryan J. – Review of Research in Education, 2015
This chapter introduces theory that undergirds the role of testing adaptations in assessment, provides examples of item modifications and testing accommodations, reviews research relevant to each, and introduces a new paradigm that incorporates opportunity to learn (OTL), academic enablers, testing adaptations, and inferences that can be made from…
Descriptors: Meta Analysis, Literature Reviews, Testing, Testing Accommodations
Hubley, Anita M.; Zumbo, Bruno D. – Social Indicators Research, 2011
The vast majority of measures have, at their core, a purpose of personal and social change. If test developers and users want measures to have personal and social consequences and impact, then it is critical to consider the consequences and side effects of measurement in the validation process itself. The consequential basis of test interpretation…
Descriptors: Construct Validity, Social Change, Measurement, Test Interpretation
Kane, Michael – Educational Testing Service, 2010
The 12th annual William H. Angoff Memorial Lecture was presented by Dr. Michael T. Kane, ETS's (Educational Testing Service) Samuel J. Messick Chair in Test Validity and the former Director of Research at the National Conference of Bar Examiners. Dr. Kane argues that it is important for policymakers to recognize the impact of errors of measurement…
Descriptors: Error of Measurement, Scores, Public Policy, Test Theory
Darrah, Marjorie; Fuller, Edgar; Miller, David – Journal of Computers in Mathematics and Science Teaching, 2010
This paper discusses a possible solution to a problem frequently encountered by educators seeking to use computer-based or multiple choice-based exams for mathematics. These assessment methodologies force a discrete grading system on students and do not allow for the possibility of partial credit. The research presented in this paper investigates…
Descriptors: College Students, College Mathematics, Calculus, Computer Assisted Testing
Vannest, Kimberly J.; Parker, Richard I.; Davis, John L.; Soares, Denise A.; Smith, Stacey L. – Behavioral Disorders, 2012
More and more, schools are considering the use of progress monitoring data for high-stakes decisions such as special education eligibility, program changes to more restrictive environments, and major changes in educational goals. Those high-stakes types of data-based decisions will need methodological defensibility. Current practice for…
Descriptors: Decision Making, Educational Change, Regression (Statistics), Field Tests
Brooks, Lindsay – Language Testing, 2009
This study, framed within sociocultural theory, examines the interaction of adult ESL test-takers in two tests of oral proficiency: one in which they interacted with an examiner (the individual format) and one in which they interacted with another student (the paired format). The data for the eight pairs in this study were drawn from a larger…
Descriptors: Testing, Rating Scales, Program Effectiveness, Interaction
Braun, Henry I.; Mislevy, Robert J. – US Department of Education, 2004
Psychologist Andrea diSessa coined the term "phenomenological primitives", or p-prims, to talk about nonexperts' reasoning about physical situations. P-prims are primitive in the sense that they stand without significant explanatory substructure or explanation. Examples are "Heavy objects fall faster than light objects" and "Continuing force is…
Descriptors: Test Theory, Testing, Evaluation Methods, Scores

Wilcox, Rand R. – Educational and Psychological Measurement, 1983
This article provides unbiased estimates of the proportion of items in an item domain that an examinee would answer correctly if every item were attempted, when a closed sequential testing procedure is used. (Author)
Descriptors: Estimation (Mathematics), Psychometrics, Scores, Sequential Approach

Woodruff, David – Journal of Educational Statistics, 1986
The purpose of the present paper is to derive linear equating methods for the common item nonequivalent populations design from explicitly stated congeneric type test score models. The equating methods developed are compared with previously developed methods and applied to five professionally constructed examinations administered to approximately…
Descriptors: Equated Scores, Equations (Mathematics), Mathematical Models, Scores
Burton, Robert S. – New Directions for Testing and Measurement, 1980
Although Model A, the only norm-referenced evaluation procedure in the Title I Evaluation and Reporting System, requires no data other than the test scores themselves, it introduces two sources of bias and involved three test administrations. Roberts' two-test procedure offers the advantages of less bias and less testing. (RL)
Descriptors: Comparative Analysis, Mathematical Formulas, Scores, Statistical Bias
Previous Page | Next Page ยป
Pages: 1 | 2