NotesFAQContact Us
Collection
Advanced
Search Tips
Source
Educational Measurement:…17
Audience
Location
Idaho1
Laws, Policies, & Programs
No Child Left Behind Act 20011
Assessments and Surveys
What Works Clearinghouse Rating
Showing 1 to 15 of 17 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Folger, Timothy D.; Bostic, Jonathan; Krupa, Erin E. – Educational Measurement: Issues and Practice, 2023
Validity is a fundamental consideration of test development and test evaluation. The purpose of this study is to define and reify three key aspects of validity and validation, namely test-score interpretation, test-score use, and the claims supporting interpretation and use. This study employed a Delphi methodology to explore how experts in…
Descriptors: Test Interpretation, Scores, Test Use, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Student, Sanford R.; Gong, Brian – Educational Measurement: Issues and Practice, 2022
We address two persistent challenges in large-scale assessments of the Next Generation Science Standards: (a) the validity of score interpretations that target the standards broadly and (b) how to structure claims for assessments of this complex domain. The NGSS pose a particular challenge for specifying claims about students that evidence from…
Descriptors: Science Tests, Test Validity, Test Items, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
An, Lily Shiao; Ho, Andrew Dean; Davis, Laurie Laughlin – Educational Measurement: Issues and Practice, 2022
Technical documentation for educational tests focuses primarily on properties of individual scores at single points in time. Reliability, standard errors of measurement, item parameter estimates, fit statistics, and linking constants are standard technical features that external stakeholders use to evaluate items and individual scale scores.…
Descriptors: Documentation, Scores, Evaluation Methods, Longitudinal Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Schneider, M. Christina; Agrimson, Jared; Veazey, Mary – Educational Measurement: Issues and Practice, 2022
This paper presents results of a score interpretation study for a computer adaptive mathematics assessment. The study purpose was to test the efficacy of item developers' alignment of items to Range Achievement-Level Descriptors (RALDs; Egan et al.) against the empirical achievement-level alignment of items to investigate the use of RALDs as the…
Descriptors: Computer Assisted Testing, Mathematics Tests, Scores, Grade 3
Peer reviewed Peer reviewed
Direct linkDirect link
O'Leary, Timothy M.; Hattie, John A. C.; Griffin, Patrick – Educational Measurement: Issues and Practice, 2017
Validity is the most fundamental consideration in test development. Understandably, much time, effort, and money is spent in its pursuit. Central to the modern conception of validity are the interpretations made, and uses planned, on the basis of test scores. There is, unfortunately, however, evidence that test users have difficulty understanding…
Descriptors: Test Interpretation, Scores, Test Validity, Evidence
Peer reviewed Peer reviewed
Direct linkDirect link
Johnson, Evelyn S.; Crawford, Angela; Moylan, Laura A.; Zheng, Yuzhu – Educational Measurement: Issues and Practice, 2018
The evidence-centered design framework was used to create a special education teacher observation system, Recognizing Effective Special Education Teachers. Extensive reviews of research informed the domain analysis and modeling stages, and led to the conceptual framework in which effective special education teaching is operationalized as the…
Descriptors: Evidence Based Practice, Special Education Teachers, Observation, Disabilities
Peer reviewed Peer reviewed
Direct linkDirect link
Bejar, Issac I. – Educational Measurement: Issues and Practice, 2012
The scoring process is critical in the validation of tests that rely on constructed responses. Documenting that readers carry out the scoring in ways consistent with the construct and measurement goals is an important aspect of score validity. In this article, rater cognition is approached as a source of support for a validity argument for scores…
Descriptors: Scores, Inferences, Validity, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Nichols, Paul D.; Williams, Natasha – Educational Measurement: Issues and Practice, 2009
This article has three goals. The first goal is to clarify the role that the consequences of test score use play in validity judgments by reviewing the role that modern writers on validity have ascribed for consequences in supporting validity judgments. The second goal is to summarize current views on who is responsible for collecting evidence of…
Descriptors: Tests, Test Validity, Scores, Data Collection
Peer reviewed Peer reviewed
Direct linkDirect link
Lu, Ying; Sireci, Stephen G. – Educational Measurement: Issues and Practice, 2007
Speededness refers to the situation where the time limits on a standardized test do not allow substantial numbers of examinees to fully consider all test items. When tests are not intended to measure speed of responding, speededness introduces a severe threat to the validity of interpretations based on test scores. In this article, we describe…
Descriptors: Test Items, Timed Tests, Standardized Tests, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip; Haberman, Shelby; Puhan, Gautam – Educational Measurement: Issues and Practice, 2007
There is an increasing interest in reporting subscores, both at examinee level and at aggregate levels. However, it is important to ensure reasonable subscore performance in terms of high reliability and validity to minimize incorrect instructional and remediation decisions. This article employs a statistical measure based on classical test theory…
Descriptors: Test Reliability, Test Theory, Test Validity, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Haladyna, Thomas M.; Downing, Steven M. – Educational Measurement: Issues and Practice, 2004
There are many threats to validity in high-stakes achievement testing. One major threat is construct-irrelevant variance (CIV). This article defines CIV in the context of the contemporary, unitary view of validity and presents logical arguments, hypotheses, and documentation for a variety of CIV sources that commonly threaten interpretations of…
Descriptors: Student Evaluation, Evaluation Methods, High Stakes Tests, Construct Validity
Peer reviewed Peer reviewed
Guion, Robert M. – Educational Measurement: Issues and Practice, 1995
This commentary discusses three essential themes in performance assessment and its scoring. First, scores should mean something. Second, performance scores should permit fair and meaningful comparisons. Third, validity-reducing errors should be minimal. Increased attention to performance assessment may overcome these problems. (SLD)
Descriptors: Educational Assessment, Performance Based Assessment, Scores, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Elliott, Stephen N.; Compton, Elizabeth; Roach, Andrew T. – Educational Measurement: Issues and Practice, 2007
The relationships between ratings on the Idaho Alternate Assessment (IAA) for 116 students with significant disabilities and corresponding ratings for the same students on two norm-referenced teacher rating scales were examined to gain evidence about the validity of resulting IAA scores. To contextualize these findings, another group of 54…
Descriptors: Inferences, Disabilities, Rating Scales, Eligibility
Peer reviewed Peer reviewed
Direct linkDirect link
Cizek, Gregory J.; Crocker, Linda; Frisbie, David A.; Mehrens, William A.; Stiggins, Richard J. – Educational Measurement: Issues and Practice, 2006
The authors describe the significant contributions of Robert Ebel to educational measurement theory and its applications. A biographical sketch details Ebel's roots and professional resume. His influence on classroom assessment views and procedures are explored. Classic publications associated with validity, reliability, and score interpretation…
Descriptors: Test Theory, Educational Assessment, Psychometrics, Test Reliability
Peer reviewed Peer reviewed
Linn, Robert L. – Educational Measurement: Issues and Practice, 1997
It is argued that consequential validity is a concept worth considering. The solution to defining "validity" is not to narrow the concept, but to allow for the differential prediction provided by tests in different circumstances. Consequences of the uses and interpretations of test scores are central to their evaluation. (SLD)
Descriptors: Educational Assessment, Educational Testing, Elementary Secondary Education, Evaluation Methods
Previous Page | Next Page ยป
Pages: 1  |  2