NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)1
Since 2006 (last 20 years)15
What Works Clearinghouse Rating
Showing 1 to 15 of 110 results Save | Export
Golan, Shari; Woodbridge, Michelle; Davies-Mercier, Betsy; Pistorino, Carol – Office of Planning, Evaluation and Policy Development, US Department of Education, 2016
States increasingly are incorporating Kindergarten Entry Assessments (KEAs) into their comprehensive assessment systems with the goal of helping educators identify gaps in children's competencies, target instruction to children's individual needs, engage parents to better support their child's learning, and identify needs for expanding and…
Descriptors: Kindergarten, Tests, Testing Programs, Data Use
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sabatini, John; O'Reilly, Tenaha; Deane, Paul – ETS Research Report Series, 2013
This report describes the foundation and rationale for a framework designed to measure reading literacy. The aim of the effort is to build an assessment system that reflects current theoretical conceptions of reading and is developmentally sensitive across a prekindergarten to 12th grade student range. The assessment framework is intended to…
Descriptors: Reading Tests, Literacy, Models, Testing Programs
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Stohlman, Trey – Journal of the Scholarship of Teaching and Learning, 2015
A good assessment plan combines many direct and indirect measures to validate the collected data. One often controversial assessment measure comes in the form of retention exams. Although assessment retention exams may come with faults, others advocate for their inclusion in program assessment. Objective-based tests may offer insight to…
Descriptors: Alternative Assessment, Retention (Psychology), Program Evaluation, Program Effectiveness
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wang, Lin; Qian, Jiahe; Lee, Yi-Hsuan – ETS Research Report Series, 2013
The purpose of this study was to evaluate the combined effects of reduced equating sample size and shortened anchor test length on item response theory (IRT)-based linking and equating results. Data from two independent operational forms of a large-scale testing program were used to establish the baseline results for evaluating the results from…
Descriptors: Test Construction, Item Response Theory, Testing Programs, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Albano, Anthony D. – Journal of Educational Measurement, 2013
In many testing programs it is assumed that the context or position in which an item is administered does not have a differential effect on examinee responses to the item. Violations of this assumption may bias item response theory estimates of item and person parameters. This study examines the potentially biasing effects of item position. A…
Descriptors: Test Items, Item Response Theory, Test Format, Questioning Techniques
Mullis, Ina V. S., Ed.; Martin, Michael O., Ed. – International Association for the Evaluation of Educational Achievement, 2014
It is critical for countries to ensure that capable secondary school students receive further preparation in advanced mathematics and science, so that they are ready to enter challenging university-level studies that prepare them for careers in science, technology, engineering, and mathematics (STEM) fields. This group of students will become the…
Descriptors: Mathematics Tests, Science Tests, Educational Assessment, Secondary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Xinnian; Graesser, Donnasue; Sah, Megha – Advances in Physiology Education, 2015
Laboratory courses serve as important gateways to science, technology, engineering, and mathematics education. One of the challenges in assessing laboratory learning is to conduct meaningful and standardized practical exams, especially for large multisection laboratory courses. Laboratory practical exams in life sciences courses are frequently…
Descriptors: Laboratory Experiments, Standardized Tests, Testing Programs, Testing Problems
Cresswell, John; Schwantner, Ursula; Waters, Charlotte – OECD Publishing, 2015
This report reviews the major international and regional large-scale educational assessments, including international surveys, school-based surveys and household-based surveys. The report compares and contrasts the cognitive and contextual data collection instruments and implementation methods used by the different assessments in order to identify…
Descriptors: International Assessment, Educational Assessment, Data Collection, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E. – Applied Psychological Measurement, 2011
In many practical testing situations, alternate test forms from the same testing program are not strictly parallel to each other and instead the test forms exhibit small psychometric differences. This article investigates the potential practical impact that these small psychometric differences can have on expected classification accuracy. Ten…
Descriptors: Test Format, Test Construction, Testing Programs, Psychometrics
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Moses, Tim; Liu, Jinghua; Tan, Adele; Deng, Weiling; Dorans, Neil J. – ETS Research Report Series, 2013
In this study, differential item functioning (DIF) methods utilizing 14 different matching variables were applied to assess DIF in the constructed-response (CR) items from 6 forms of 3 mixed-format tests. Results suggested that the methods might produce distinct patterns of DIF results for different tests and testing programs, in that the DIF…
Descriptors: Test Construction, Multiple Choice Tests, Test Items, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrara, Steve; Svetina, Dubravka; Skucha, Sylvia; Davidson, Anne H. – Educational Measurement: Issues and Practice, 2011
Items on test score scales located at and below the Proficient cut score define the content area knowledge and skills required to achieve proficiency. Alternately, examinees who perform at the Proficient level on a test can be expected to be able to demonstrate that they have mastered most of the knowledge and skills represented by the items at…
Descriptors: Knowledge Level, Mathematics Tests, Program Effectiveness, Inferences
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rock, Donald A. – ETS Research Report Series, 2012
This paper provides a history of ETS's role in developing assessment instruments and psychometric procedures for measuring change in large-scale national assessments funded by the Longitudinal Studies branch of the National Center for Education Statistics. It documents the innovations developed during more than 30 years of working with…
Descriptors: Models, Educational Change, Longitudinal Studies, Educational Development
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Unger, Darian – American Journal of Business Education, 2010
Although there is significant research on improving college-level teaching practices, most literature in the field assumes an incentive for improvement. The research presented in this paper addresses the issue of poor incentives for improving university-level teaching. Specifically, it proposes instructor-designed common examinations as an…
Descriptors: Educational Innovation, Educational Improvement, Instructional Improvement, Business Administration Education
Dorans, Neil J.; Liu, Jinghua – Educational Testing Service, 2009
The equating process links scores from different editions of the same test. For testing programs that build nearly parallel forms to the same explicit content and statistical specifications and administer forms under the same conditions, the linkings between the forms are expected to be equatings. Score equity assessment (SEA) provides a useful…
Descriptors: Testing Programs, Mathematics Tests, Quality Control, Psychometrics
van der Linden, Wim J.; Veldkamp, Bernard P.; Reese, Lynda M. – 2000
Presented is an integer-programming approach to item pool design that can be used to calculate an optimal blueprint for an item pool to support an existing testing program. The results are optimal in the sense that they minimize the efforts involved in actually producing the items as revealed by current item writing patterns. Also presented is an…
Descriptors: Item Banks, Test Construction, Test Items, Testing Programs
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8