Publication Date
In 2025 | 11 |
Since 2024 | 32 |
Since 2021 (last 5 years) | 95 |
Since 2016 (last 10 years) | 196 |
Since 2006 (last 20 years) | 419 |
Descriptor
Student Evaluation | 727 |
Test Items | 727 |
Test Construction | 231 |
Foreign Countries | 202 |
Evaluation Methods | 158 |
Educational Assessment | 135 |
Multiple Choice Tests | 124 |
Achievement Tests | 113 |
Test Validity | 99 |
Difficulty Level | 91 |
Item Response Theory | 87 |
More ▼ |
Source
Author
Hines, Donna | 7 |
Tindal, Gerald | 7 |
Abedi, Jamal | 6 |
Alonzo, Julie | 6 |
Herman, Joan | 5 |
Liu, Ou Lydia | 5 |
Kao, Jenny C. | 4 |
Mislevy, Robert J. | 4 |
Allen, Nancy L. | 3 |
Avery, Marybell | 3 |
Cai, Li | 3 |
More ▼ |
Publication Type
Education Level
Audience
Teachers | 106 |
Practitioners | 100 |
Students | 50 |
Administrators | 37 |
Researchers | 8 |
Policymakers | 7 |
Parents | 6 |
Community | 3 |
Support Staff | 3 |
Location
Canada | 63 |
Florida | 21 |
Australia | 17 |
Turkey | 10 |
Germany | 9 |
Taiwan | 8 |
United Kingdom (England) | 8 |
United States | 8 |
Indonesia | 7 |
Colorado | 6 |
North Carolina | 6 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Marta Siedlecka; Piotr Litwin; Paulina Szyszka; Boryslaw Paulewicz – European Journal of Psychology of Education, 2025
Students change their responses during tests, and these revisions are often correct. Some studies have suggested that decisions regarding revisions are informed by metacognitive monitoring. We investigated whether assessing and reporting response confidence increases the accuracy of revisions and the final test score, and whether confidence in a…
Descriptors: Student Evaluation, Decision Making, Responses, Achievement Tests
Lyniesha Ward; Fridah Rotich; Jeffrey R. Raker; Regis Komperda; Sachin Nedungadi; Maia Popova – Chemistry Education Research and Practice, 2025
This paper describes the design and evaluation of the Organic chemistry Representational Competence Assessment (ORCA). Grounded in Kozma and Russell's representational competence framework, the ORCA measures the learner's ability to "interpret," "translate," and "use" six commonly used representations of molecular…
Descriptors: Organic Chemistry, Science Tests, Test Construction, Student Evaluation
Goran Trajkovski; Heather Hayes – Digital Education and Learning, 2025
This book explores the transformative role of artificial intelligence in educational assessment, catering to researchers, educators, administrators, policymakers, and technologists involved in shaping the future of education. It delves into the foundations of AI-assisted assessment, innovative question types and formats, data analysis techniques,…
Descriptors: Artificial Intelligence, Educational Assessment, Computer Uses in Education, Test Format
Endang Susantini; Yurizka Melia Sari; Prima Vidya Asteria; Muhammad Ilyas Marzuqi – Journal of Education and Learning (EduLearn), 2025
Assessing preservice' higher order thinking skills (HOTS) in science and mathematics is essential. Teachers' HOTS ability is closely related to their ability to create HOTS-type science and mathematics problems. Among various types of HOTS, one is Bloomian HOTS. To facilitate the preservice teacher to create problems in those subjects, an Android…
Descriptors: Content Validity, Mathematics Instruction, Decision Making, Thinking Skills
Kevin Ackermans; Marjoke Bakker; Pierre Gorissen; Anne-Marieke Loon; Marijke Kral; Gino Camp – Journal of Computer Assisted Learning, 2024
Background: A practical test that measures the information and communication technology (ICT) skills students need for effectively using ICT in primary education has yet to be developed (Oh et al., 2021). This paper reports on the development, validation, and reliability of a test measuring primary school students' ICT skills required for…
Descriptors: Test Construction, Test Validity, Measures (Individuals), Elementary School Students
Mingjia Ma – ProQuest LLC, 2023
Response time is an important research topic in the field of psychometrics. This dissertation tries to explore some response time properties across several item characteristics and examinee characteristics, as well as the interactions between response time and response outcomes, using data from a statewide mathematics assessment in two grades.…
Descriptors: Reaction Time, Mathematics Tests, Standardized Tests, State Standards
Alicia A. Stoltenberg – ProQuest LLC, 2024
Multiple-select multiple-choice items, or multiple-choice items with more than one correct answer, are used to quickly assess content on standardized assessments. Because there are multiple keys to these item types, there are also multiple ways to score student responses to these items. The purpose of this study was to investigate how changing the…
Descriptors: Scoring, Evaluation Methods, Multiple Choice Tests, Standardized Tests
Maristela Petrovic-Dzerdz – Collected Essays on Learning and Teaching, 2024
Large introductory classes, with their expansive curriculum, demand assessment strategies that blend efficiency with reliability, prompting the consideration of multiple-choice (MC) tests as a viable option. Crafting a high-quality MC test, however, necessitates a meticulous process involving reflection on assessment format appropriateness, test…
Descriptors: Multiple Choice Tests, Test Construction, Test Items, Alignment (Education)
Wang, Yu; Chiu, Chia-Yi; Köhn, Hans Friedrich – Journal of Educational and Behavioral Statistics, 2023
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Nonparametric Statistics, Test Format, Educational Assessment
Miranda Kucera; K. Kawena Begay – Communique, 2025
While the field advocates for a diversified and comprehensive professional role (National Association of School Psychologists, 2020), school psychologists have long spent most of their time in assessment-related activities (Farmer et al., 2021), averaging about eight cognitive evaluations monthly (Benson et al., 2020). Assessment practices have…
Descriptors: Equal Education, Student Evaluation, Evaluation Methods, Standardized Tests
Miranda Kucera; K. Kawena Begay – Communique, 2025
In Part 1 of this series, the authors briefly reviewed some challenges inherent in using standardized tools with students who are not well represented in norming data. To help readers clearly conceptualize the framework steps, the authors present two case studies that showcase how a nonstandardized approach to assessment can be individualized to…
Descriptors: Equal Education, Student Evaluation, Evaluation Methods, Standardized Tests
Balta, Nuri; Logman, Paul S. W. M. – Physics Education, 2022
The purpose of this study is to develop a test to assess students' level of counterintuitiveness in basic electric circuits. Data from four samples were gathered and used to develop and validate the counterintuitive basic electric circuit test (CBECT). The initial version of the CBECT was administered to the first sample and data collected from…
Descriptors: Science Tests, Test Construction, Student Evaluation, Intuition
Lance Shultz – ProQuest LLC, 2024
Multiple-true-false (MTF) assessments can provide granular feedback on course materials, which stems from the format of the MTF question and helps to enhance student understanding and illuminates misconceptions that can be hidden with other assessment types (Brassil & Couch, 2019). The purpose of this study was to document how students use and…
Descriptors: Objective Tests, Multiple Choice Tests, Test Items, Student Evaluation
Yue Liu; Zhen Li; Hongyun Liu; Xiaofeng You – Applied Measurement in Education, 2024
Low test-taking effort of examinees has been considered a source of construct-irrelevant variance in item response modeling, leading to serious consequences on parameter estimation. This study aims to investigate how non-effortful response (NER) influences the estimation of item and person parameters in item-pool scale linking (IPSL) and whether…
Descriptors: Item Response Theory, Computation, Simulation, Responses
Shu-Fen Lin; Wan-Chin Shie – International Journal of Science and Mathematics Education, 2024
Teachers lack effective curriculum-based instruments to assess their students' scientific competence that would provide information for modifying their inquiry instruction. The main purpose of this study was to develop and validate a Curriculum-Based Scientific Competence (CBSC) test to assess students' scientific competence in a 1-semester Grade…
Descriptors: Science Curriculum, Validity, Grade 9, Science Tests