NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 14 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Baker, Eva L. – Teachers College Record, 2013
Background/Context: Education policy over the past 40 years has focused on the importance of accountability in school improvement. Although much of the scholarly discourse around testing and assessment is technical and statistical, understanding of validity by a non-specialist audience is essential as long as test results drive our educational…
Descriptors: Validity, Educational Assessment, Accountability, Educational Improvement
Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2010
This report provides an overview of what was known about alternative assessment at the time that the article was written in 1991. Topics include beliefs about assessment reform, overview of alternative assessment including research knowledge, evidence of assessment impact, and critical features of alternative assessment. The author notes that in…
Descriptors: Alternative Assessment, Evaluation Methods, Evaluation Research, Performance Based Assessment
Chung, Gregory K. W. K.; Nagashima, Sam O.; Delacruz, Girlie C.; Lee, John J.; Wainess, Richard; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2011
The UCLA National Center for Research on Evaluation, Standards, and Student Testing (CRESST) is under contract from the Naval Postgraduate School (NPS) to conduct research on assessment models and tools designed to support Marine Corps rifle marksmanship. In this deliverable, we first review the literature on known-distance rifle marksmanship…
Descriptors: Weapons, Psychomotor Skills, Computer Software, Military Personnel
Nagashima, Sam O.; Chung, Gregory K. W. K.; Espinosa, Paul D.; Berka, Chris; Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2009
The goal of this report was to test the use of sensor-based skill measures in evaluating performance differences in rifle marksmanship. Ten shots were collected from 30 novices and 9 experts. Three measures for breath control and one for trigger control were used to predict skill classification. The data were fitted with a logistic regression…
Descriptors: Weapons, Classification, Lasers, Models
Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2007
This paper will describe the relationships between research on learning and its application in assessment models and operational systems. These have been topics of research at the National Center for Research on Evaluation, Standards, and Student Testing (CRESST) for more than 20 years and form a significant part of the intellectual foundation of…
Descriptors: Educational Testing, Inferences, Hypothesis Testing, Predictive Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Goldschmidt, Pete; Martinez, Jose Felipe; Niemi, David; Baker, Eva L. – Educational Assessment, 2007
In this article we examine empirical evidence on the criterion, predictive, transfer, and fairness aspects of validity of a large-scale language arts performance assessment, referred to as the Performance Assignment (PA). We use multilevel models to avoid biased inferences that might result from the naturally nested data. Specifically, we examine…
Descriptors: Language Arts, Performance Based Assessment, Academic Achievement, Performance Tests
Peer reviewed Peer reviewed
Burstein, Leigh; Koretz, Daniel; Linn, Robert; Sugrue, Brenda; Novak, John; Baker, Eva L.; Harris, Elizabeth Lewis – Educational Assessment, 1996
Three studies evaluating the validity of the descriptors and exemplars of the National Assessment of Educational Progress (NAEP) as characterizations of the actual mathematics performance of students at achievement levels are reported. Serious inconsistencies were found between actual performance and descriptors and exemplars. Recommendations for…
Descriptors: Elementary Secondary Education, Mathematics Achievement, Mathematics Tests, National Surveys
Linn, Robert L.; Baker, Eva L. – 1996
During the past 6 years, under a contract from the National Center for Education Statistics, a Technical Review Panel has overseen and conducted a series of research studies addressing a range of validity questions relevant to the various uses and interpretations of the National Assessment of Educational Progress (NAEP). Study topics included: (1)…
Descriptors: Achievement Tests, Comparative Analysis, Data Analysis, Educational Policy
Peer reviewed Peer reviewed
Baker, Eva L. – Educational Leadership, 1994
Teachers must learn to distinguish among performance assessments of different quality and appropriateness. Design criteria (cognitive complexity, linguistic appropriateness, content quality and coverage, and meaningfulness) are judged by examining assessment tasks and scoring rubrics. Effects criteria (transfer, generalizability, instructional…
Descriptors: Context Effect, Elementary Secondary Education, Evaluation Criteria, Guidelines
Peer reviewed Peer reviewed
Direct linkDirect link
Baker, Eva L. – Yearbook of the National Society for the Study of Education, 2005
Over the last 20 years, writing about technology (and more recently about its use in assessment systems) has been a joyful experience, in part because the author was usually right about how fast and how unexpected technology progress would be made. In this article, the author suggests to rethink assessment systems from design perspectives and the…
Descriptors: Student Evaluation, Educational Assessment, Computer Assisted Testing, Educational Technology
Peer reviewed Peer reviewed
Abedi, Jamal; Baker, Eva L. – Educational and Psychological Measurement, 1995
Results from a performance assessment in which 68 high school students wrote essays support the use of latent variable modeling for estimating reliability, concurrent validity, and generalizability of a scoring rubric. The latent variable modeling approach overcomes the limitations of certain conventional statistical techniques in handling…
Descriptors: Criteria, Essays, Estimation (Mathematics), Generalizability Theory
Baker, Eva L.; And Others – 1996
The National Center for Research on Evaluation, Standards, and Student Testing (CRESST) is a partnership of the University of California at Los Angeles, the University of Colorado at Boulder, Stanford University, The RAND Corporation, the University of Pittsburgh, the Educational Testing Service, and the University of California, Santa Barbara.…
Descriptors: Agenda Setting, Educational Assessment, Educational Policy, Educational Practices
O'Neil, Harold F., Jr.; Baker, Eva L. – 1992
A prototype simulation measure of the negotiating skill of interpersonal competence is described, and results of an initial validation study are reported. Of eight negotiating skills identified by the Secretary's Commission on Achieving Necessary Skills of the U.S. Department of Labor, two were the focus of the measure: proposing and examining…
Descriptors: Arbitration, Decision Making, Employer Employee Relationship, Graduate Students
Baker, Eva L. – 1992
Work on the development of history performance assessments is described. So far, six complete sets of assessments in United States history have been developed as part of this project. Students are first assessed on their historical knowledge of the period through a short-answer test. They are then asked to write an essay explaining the positions…
Descriptors: Educational Assessment, Essay Tests, Grade 11, High School Students