Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 5 |
Descriptor
Test Interpretation | 31 |
Scoring | 12 |
Educational Assessment | 10 |
Testing Problems | 10 |
Scores | 9 |
Test Use | 9 |
Test Results | 7 |
Interrater Reliability | 6 |
Norm Referenced Tests | 6 |
Test Items | 6 |
Achievement Tests | 5 |
More ▼ |
Source
Educational Measurement:… | 31 |
Author
Hills, John R. | 2 |
Linn, Robert L. | 2 |
Bond, Lloyd | 1 |
Brennan, Robert L. | 1 |
Burton, Elizabeth | 1 |
Clauser, Brian E. | 1 |
Cohen, Allan | 1 |
Connell, Michael L. | 1 |
Crooks, Terence | 1 |
Davenport, Ernest C. | 1 |
Davison, Mark L. | 1 |
More ▼ |
Publication Type
Journal Articles | 31 |
Reports - Evaluative | 31 |
Opinion Papers | 5 |
Speeches/Meeting Papers | 2 |
Tests/Questionnaires | 2 |
Guides - Classroom - Learner | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 3 |
ACT Assessment | 1 |
Graduate Record Examinations | 1 |
Preliminary Scholastic… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
O'Leary, Timothy M.; Hattie, John A. C.; Griffin, Patrick – Educational Measurement: Issues and Practice, 2017
Validity is the most fundamental consideration in test development. Understandably, much time, effort, and money is spent in its pursuit. Central to the modern conception of validity are the interpretations made, and uses planned, on the basis of test scores. There is, unfortunately, however, evidence that test users have difficulty understanding…
Descriptors: Test Interpretation, Scores, Test Validity, Evidence
Davenport, Ernest C.; Davison, Mark L.; Liou, Pey-Yan; Love, Quintin U. – Educational Measurement: Issues and Practice, 2015
This article uses definitions provided by Cronbach in his seminal paper for coefficient a to show the concepts of reliability, dimensionality, and internal consistency are distinct but interrelated. The article begins with a critique of the definition of reliability and then explores mathematical properties of Cronbach's a. Internal consistency…
Descriptors: Reliability, Definitions, Mathematics, Test Interpretation
Sinharay, Sandip – Educational Measurement: Issues and Practice, 2014
Brennan (Brennan, R. L., 2012) noted that users of test scores often want (indeed, demand) that subscores be reported, along with total test scores, for diagnostic purposes. Haberman (Haberman, S. J., 2008) suggested a method based on classical test theory (CTT) to determine if subscores have added value over the total score. According to this…
Descriptors: Scores, Test Theory, Test Interpretation
Wise, Steven L. – Educational Measurement: Issues and Practice, 2017
The rise of computer-based testing has brought with it the capability to measure more aspects of a test event than simply the answers selected or constructed by the test taker. One behavior that has drawn much research interest is the time test takers spend responding to individual multiple-choice items. In particular, very short response…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Items, Reaction Time
Dorans, Neil J. – Educational Measurement: Issues and Practice, 2012
Views on testing--its purpose and uses and how its data are analyzed--are related to one's perspective on test takers. Test takers can be viewed as learners, examinees, or contestants. I briefly discuss the perspective of test takers as learners. I maintain that much of psychometrics views test takers as examinees. I discuss test takers as a…
Descriptors: Testing, Test Theory, Item Response Theory, Test Reliability

Kane, Michael; Crooks, Terence; Cohen, Allan – Educational Measurement: Issues and Practice, 1999
Analyzes the three major inferences involved in interpretation of performance assessments: (1) scoring of the observed performances; (2) generalization to a domain of assessment performances like those included in the assessment; and (3) extrapolation to the large performance domain of interest. Suggests ways to improve the validity of performance…
Descriptors: Performance Based Assessment, Performance Factors, Scoring, Test Interpretation

Frisbie, David A. – Educational Measurement: Issues and Practice, 1988
This National Council on Measurement in Education (NCME) module is designed to help teachers understand test reliability requirements. Focus is on the reliability of scores from teacher-made achievement tests used for interpreting norm-referenced scores. Reliability, factors affecting scores, reliability estimation methods, reliability's relation…
Descriptors: Achievement Tests, Norm Referenced Tests, Teacher Made Tests, Test Interpretation

Clauser, Brian E.; Mazor, Kathleen M. – Educational Measurement: Issues and Practice, 1998
This module prepares the reader to use statistical procedures to detect differentially functioning test items. The Mantel-Haenszel statistic, logistic regression, the SIBTEST procedure, the Standardization procedure, and various item response theory-based procedures are presented. Theoretical frameworks, strengths and weaknesses, and…
Descriptors: Item Bias, Item Response Theory, Statistical Analysis, Teaching Methods

Millman, Jason – Educational Measurement: Issues and Practice, 1994
The unfulfilled promise of criterion-referenced measurement is that it would permit valid inferences about what a student could and could not do. To come closest to achieving all that criterion-referenced testing originally promised, tests of higher item density, with more items per amount of domain, are required. (SLD)
Descriptors: Criterion Referenced Tests, Educational History, Inferences, Norm Referenced Tests

Linn, Robert L. – Educational Measurement: Issues and Practice, 1994
The term "criterion-referenced measurement" has been so frequently used that accumulated surplus meaning has sometimes swamped the fundamental concept. Stripping away extraneous interpretations affirms the contemporary relevance of the concept. Substituting the term "construct-referenced" underscores the need for better…
Descriptors: Behaviorism, Criterion Referenced Tests, Cutting Scores, Learning

Switzer, Deborah M.; Connell, Michael L. – Educational Measurement: Issues and Practice, 1990
Two easy-to-use microcomputer programs, the Student Problem Package and the Test Analysis Package, both by D. L. Harnisch et al. (1985), are described. These programs efficiently analyze test data for teachers. (SLD)
Descriptors: Classroom Techniques, Computer Assisted Testing, Computer Software, Data Analysis

Kolen, Michael J. – Educational Measurement: Issues and Practice, 1988
An instructional module is presented to promote a conceptual understanding of test form equating using traditional methods. Equating is distinguished from scaling. The equating methods described are: (1) mean; (2) linear; and (3) equipercentile. The module includes a self-test. (SLD)
Descriptors: College Entrance Examinations, College Students, Equated Scores, Higher Education

Popham, W. James – Educational Measurement: Issues and Practice, 1994
The instructional impact of criterion-referenced testing during the past three decades is reviewed, and a method of devising criterion-referenced tests is suggested that departs from the usual model. Adding verification along the lines of panel review that newly constructed test items do elicit appropriate information is suggested. (SLD)
Descriptors: Achievement Tests, Criterion Referenced Tests, Educational Assessment, Educational Research

Linn, Robert L.; Burton, Elizabeth – Educational Measurement: Issues and Practice, 1994
Generalizability of performance-based assessment scores across raters and tasks is examined, focusing on implications of generalizability analyses for specific uses and interpretations of assessment results. Although it seems probable that assessment conditions, task characteristics, and interactions with instructional experiences affect the…
Descriptors: Educational Assessment, Educational Experience, Generalizability Theory, Interaction

Lenke, Joanne M.; Keene, John M. – Educational Measurement: Issues and Practice, 1988
In response to J. J. Cannell's (1988) discussion of the reported "above average" achievement of the vast majority of public school systems nationwide, the nature of national norms is discussed. The public must be better informed about the interpretation of test results. (SLD)
Descriptors: Achievement Tests, Elementary Education, National Norms, National Surveys