Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 10 |
Since 2016 (last 10 years) | 34 |
Since 2006 (last 20 years) | 60 |
Descriptor
Item Response Theory | 64 |
Student Evaluation | 64 |
Test Reliability | 37 |
Test Validity | 30 |
Test Items | 26 |
Test Construction | 23 |
Psychometrics | 21 |
Foreign Countries | 19 |
Reliability | 17 |
Evaluation Methods | 16 |
Interrater Reliability | 14 |
More ▼ |
Source
Author
Petscher, Yaacov | 3 |
Avery, Marybell | 2 |
Dyson, Ben | 2 |
Fisette, Jennifer L. | 2 |
Fox, Connie | 2 |
Franck, Marian | 2 |
Graber, Kim C. | 2 |
Karakaya, Ismail | 2 |
Park, Youngsik | 2 |
Placek, Judith H. | 2 |
Raynes, De | 2 |
More ▼ |
Publication Type
Education Level
Audience
Administrators | 1 |
Researchers | 1 |
Location
Australia | 2 |
Canada | 2 |
Germany | 2 |
New Mexico | 2 |
Turkey | 2 |
Turkey (Ankara) | 2 |
United States | 2 |
California | 1 |
China | 1 |
Florida | 1 |
Indonesia | 1 |
More ▼ |
Laws, Policies, & Programs
Individuals with Disabilities… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Kason Ka Ching Cheung; Jack K. H. Pun; Xuehua Fu – International Journal of Science and Mathematics Education, 2024
Researchers in science education lacks valid and reliable instruments to assess students' "disciplinary" and "epistemic" reading of scientific texts. The main purpose of this study was to develop and validate a Reading in Science Holistic Assessment (RISHA) to assess students' holistic reading of scientific texts. RISHA…
Descriptors: Test Construction, Reading Tests, Science Education, Student Evaluation
Corradi, David – Assessment & Evaluation in Higher Education, 2023
Juries are a high-stake practice in higher education to assess complex competencies. However common, research remains behind in detailing the psychometric qualities of juries, especially when using rubrics or rating scales as an assessment tool. In this study, I analyze a case of a jury assessment (N = 191) of product development where both…
Descriptors: Court Litigation, Educational Practices, Higher Education, Rating Scales
Myers, Aaron J.; Ames, Allison J.; Leventhal, Brian C.; Holzman, Madison A. – Applied Measurement in Education, 2020
When rating performance assessments, raters may ascribe different scores for the same performance when rubric application does not align with the intended application of the scoring criteria. Given performance assessment score interpretation assumes raters apply rubrics as rubric developers intended, misalignment between raters' scoring processes…
Descriptors: Scoring Rubrics, Validity, Item Response Theory, Interrater Reliability
Wang, Yu; Chiu, Chia-Yi; Köhn, Hans Friedrich – Journal of Educational and Behavioral Statistics, 2023
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Nonparametric Statistics, Test Format, Educational Assessment
Fitria Lafifa; Dadan Rosana – Turkish Online Journal of Distance Education, 2024
This research goal to develop a multiple-choice closed-ended test to assessing and evaluate students' digital literacy skills. The sample in this study were students at MTsN 1 Blitar City who were selected using a purposive sampling technique. The test was also validated by experts, namely 2 Doctors of Physics and Science from Yogyakarta State…
Descriptors: Educational Innovation, Student Evaluation, Digital Literacy, Multiple Choice Tests
Geoffrey Converse – ProQuest LLC, 2021
In educational measurement, Item Response Theory (IRT) provides a means of quantifying student knowledge. Specifically, IRT models the probability of a student answering a particular item correctly as a function of the student's continuous-valued latent abilities [theta] (e.g. add, subtract, multiply, divide) and parameters associated with the…
Descriptors: Item Response Theory, Test Validity, Student Evaluation, Computer Assisted Testing
Koçak, Duygu – International Electronic Journal of Elementary Education, 2020
One of the most commonly used methods for measuring higher-order thinking skills such as problem-solving or written expression is open-ended items. Three main approaches are used to evaluate responses to open-ended items: general evaluation, rating scales, and rubrics. In order to measure and improve problem-solving skills of students, firstly, an…
Descriptors: Interrater Reliability, Item Response Theory, Test Items, Rating Scales
Saritas Akyol, Seyhan; Karakaya, Ismail – Eurasian Journal of Educational Research, 2021
Purpose: To assess students' problem-solving skills, this study aims to investigate the consistency between self- and peer-ratings in consideration of the teachers' ratings in the process. Method: This study was a descriptive study which examines the mathematical problem-solving skills with the MFRM model concerning self-, peer- and teachers'…
Descriptors: Problem Solving, Item Response Theory, Self Evaluation (Individuals), Peer Evaluation
Malone, Kathy L.; Boone, William J.; Stammen, Andria; Schuchardt, Anita; Ding, Lin; Sabree, Zakee – EURASIA Journal of Mathematics, Science and Technology Education, 2021
Instruments for assessing secondary students' conceptual understanding of core concepts in biology are needed by educational practitioners and researchers alike. Most instruments available for secondary biology (years 9 to 12) focus only on highly specific biological concepts instead of multiple core concepts. This study describes the development…
Descriptors: Measures (Individuals), Test Construction, Construct Validity, Test Reliability
Scribner, Emily D.; Harris, Sara E. – Journal of Geoscience Education, 2020
The Mineralogy Concept Inventory (MCI) is a statistically validated 18-question assessment that can be used to measure learning gains in introductory mineralogy courses. Development of the MCI was an iterative process involving expert consultation, student interviews, assessment deployment, and statistical analysis. Experts at the two universities…
Descriptors: Undergraduate Students, Mineralogy, Introductory Courses, Science Tests
Castle, Courtney – ProQuest LLC, 2018
The Next Generation Science Standards propose a multidimensional model of science learning, comprised of Core Disciplinary Ideas, Science and Engineering Practices, and Crosscutting Concepts (NGSS Lead States, 2013). Accordingly, there is a need for student assessment aligned with the new standards. Creating assessments that validly and reliably…
Descriptors: Science Education, Student Evaluation, Science Tests, Test Construction
Ziegler, Laura; Garfield, Joan – Statistics Education Research Journal, 2018
The purpose of this study was to develop the Basic Literacy In Statistics (BLIS) assessment for students in an introductory statistics course, at the postsecondary level, that includes, to some extent, simulation-based methods. The definition of statistical literacy used in the development of the assessment was the ability to read, understand, and…
Descriptors: Statistics, Literacy, Introductory Courses, College Students
Tobler, Samuel; Köhler, Katja; Sinha, Tanmay; Hafen, Ernst; Kapur, Manu – CBE - Life Sciences Education, 2022
Undergraduate biology students' molecular-level understanding of stochastic (also referred to as random or noisy) processes found in biological systems is often limited to those examples discussed in class. Therefore, students frequently display little ability to accurately transfer their knowledge to other contexts. Furthermore, elaborate tools…
Descriptors: Undergraduate Students, Biology, Science Instruction, Molecular Biology
Flynn, Carli D.; Davidson, Cliff I.; Dotger, Sharon – Journal of Engineering Education, 2018
Background: Because a fundamental understanding of rate and accumulation principles is important for engineering students across all subdisciplines, a method is needed to assess their understanding of these principles. Purpose/Hypothesis: This article discusses the development of the Rate and Accumulation Concept Inventory (RACI) instrument and…
Descriptors: Engineering Education, Test Construction, Psychometrics, Mastery Learning
A Feasible Guidance for Ordered Multiple-Choice Items in Students' Hierarchical Understanding Levels
Su, King-Dow – Journal of Baltic Science Education, 2019
This research focuses on students' 5 hierarchical levels of Ordered Multiple-Choice (OMC) items for their extensive conceptualized understanding in the particulate nature of matter (PNM) chemistry. The basic framework for OMC items is to link students' conceptual understanding levels with possible cognitive responses. Developed as the substantial…
Descriptors: Multiple Choice Tests, Science Tests, STEM Education, Test Items