Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 12 |
Descriptor
Source
Author
Baker, Eva L. | 32 |
Chung, Gregory K. W. K. | 7 |
Delacruz, Girlie C. | 4 |
Buschang, Rebecca E. | 2 |
Nagashima, Sam O. | 2 |
Abedi, Jamal | 1 |
Berka, Chris | 1 |
Bewley, William L. | 1 |
Brill, David G. | 1 |
Burstein, Leigh | 1 |
Cai, Li | 1 |
More ▼ |
Publication Type
Reports - Evaluative | 14 |
Journal Articles | 12 |
Reports - Research | 8 |
Information Analyses | 6 |
Reports - Descriptive | 4 |
Speeches/Meeting Papers | 4 |
Guides - Non-Classroom | 3 |
Opinion Papers | 3 |
Tests/Questionnaires | 1 |
Education Level
Elementary Secondary Education | 4 |
Grade 6 | 1 |
Grade 7 | 1 |
Grade 8 | 1 |
High Schools | 1 |
Middle Schools | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 3 |
What Works Clearinghouse Rating
Baker, Eva L. – Teachers College Record, 2013
Background/Context: Education policy over the past 40 years has focused on the importance of accountability in school improvement. Although much of the scholarly discourse around testing and assessment is technical and statistical, understanding of validity by a non-specialist audience is essential as long as test results drive our educational…
Descriptors: Validity, Educational Assessment, Accountability, Educational Improvement
Baker, Eva L.; Chung, Gregory K. W. K.; Cai, Li – Review of Research in Education, 2016
This chapter addresses assessment (testing) with an emphasis on the 100-year period since the American Education Research Association was formed. The authors start with definitions and explanations of contemporary tests. They then look backward into the 19th century to significant work by Horace Mann and Herbert Spencer, who engendered two…
Descriptors: Achievement Tests, Educational History, Testing, Educational Assessment
Buschang, Rebecca E.; Chung, Gregory K. W. K.; Delacruz, Girlie C.; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2012
The purpose of this study was to validate inferences about scores of one task designed to measure subject matter knowledge and three tasks designed to measure aspects of pedagogical content knowledge. Evidence for the validity of inferences was based on two expectations. First, if tasks were sensitive to expertise, we would find group differences.…
Descriptors: Validity, Measures (Individuals), Test Interpretation, Algebra
Buschang, Rebecca E.; Chung, Gregory K. W. K.; Delacruz, Girlie C.; Baker, Eva L. – Educational Assessment, 2012
The purpose of this study was to validate inferences about scores of one task designed to measure subject matter knowledge and three tasks designed to measure aspects of pedagogical content knowledge. Evidence for the validity of inferences was based on two expectations. First, if tasks were sensitive to expertise, we would find group differences.…
Descriptors: Algebra, Mathematics Teachers, Teacher Characteristics, Knowledge Base for Teaching
Delacruz, Girlie C.; Chung, Gregory K. W. K.; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2010
This study provides empirical evidence of a highly specific use of games in education--the assessment of the learner. Linear regressions were used to examine the predictive and convergent validity of a math game as assessment of mathematical understanding. Results indicate that prior knowledge significantly predicts game performance. Results also…
Descriptors: Educational Games, Validity, Prior Learning, Scores
Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2010
This report provides an overview of what was known about alternative assessment at the time that the article was written in 1991. Topics include beliefs about assessment reform, overview of alternative assessment including research knowledge, evidence of assessment impact, and critical features of alternative assessment. The author notes that in…
Descriptors: Alternative Assessment, Evaluation Methods, Evaluation Research, Performance Based Assessment
Chung, Gregory K. W. K.; Nagashima, Sam O.; Delacruz, Girlie C.; Lee, John J.; Wainess, Richard; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2011
The UCLA National Center for Research on Evaluation, Standards, and Student Testing (CRESST) is under contract from the Naval Postgraduate School (NPS) to conduct research on assessment models and tools designed to support Marine Corps rifle marksmanship. In this deliverable, we first review the literature on known-distance rifle marksmanship…
Descriptors: Weapons, Psychomotor Skills, Computer Software, Military Personnel
Nagashima, Sam O.; Chung, Gregory K. W. K.; Espinosa, Paul D.; Berka, Chris; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2009
The goal of this report was to test the use of sensor-based skill measures in evaluating performance differences in rifle marksmanship. Ten shots were collected from 30 novices and 9 experts. Three measures for breath control and one for trigger control were used to predict skill classification. The data were fitted with a logistic regression…
Descriptors: Weapons, Classification, Lasers, Models
Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2007
This paper will describe the relationships between research on learning and its application in assessment models and operational systems. These have been topics of research at the National Center for Research on Evaluation, Standards, and Student Testing (CRESST) for more than 20 years and form a significant part of the intellectual foundation of…
Descriptors: Educational Testing, Inferences, Hypothesis Testing, Predictive Validity
Baker, Eva L. – Educational Assessment, 2007
This article describes the history, evidence warrants, and evolution of the Center for Research on Evaluation, Standards, and Student Testing's (CRESST) model-based assessments. It considers alternative interpretations of scientific or practical models and illustrates how model-based assessment addresses both definitions. The components of the…
Descriptors: Educational Testing, Computer Assisted Testing, Validity, Test Construction

Baker, Eva L.; And Others – Journal for the Education of the Gifted, 1994
This article describes performance-based assessment as expounded by its proponents, comments on these conceptions, reviews evidence regarding the technical quality of performance-based assessment, and considers its validity under various policy options. (JDD)
Descriptors: Educational Change, Educational Policy, Elementary Secondary Education, Evaluation Methods
Goldschmidt, Pete; Martinez, Jose Felipe; Niemi, David; Baker, Eva L. – Educational Assessment, 2007
In this article we examine empirical evidence on the criterion, predictive, transfer, and fairness aspects of validity of a large-scale language arts performance assessment, referred to as the Performance Assignment (PA). We use multilevel models to avoid biased inferences that might result from the naturally nested data. Specifically, we examine…
Descriptors: Language Arts, Performance Based Assessment, Academic Achievement, Performance Tests
Baker, Eva L.; Linn Robert L. – 2002
This report analyzes the validity issues that arise in the context of educational accountability systems. The report addresses validity from three interlocking perspectives. The first explores the theory of action underlying accountability provisions. It considers problems emerging from the distance between aspirations for accountability in…
Descriptors: Accountability, Educational Assessment, Educational Change, Educational Testing
Baker, Eva L. – Educational Technology, 1974
Descriptors: Behavioral Objectives, Evaluation Criteria, Evaluation Methods, Instructional Improvement
Polin, Linda; Baker, Eva L. – 1979
This paper presents the interim results of a set of studies undertaken to develop a much needed methodology for establishing content validity in domain-referenced achievement tests. The study results are presented in the context of the larger issue of the improvement of test design. School teachers, administrators and graduate students were…
Descriptors: Achievement Tests, Criterion Referenced Tests, Elementary Secondary Education, Evaluation Methods