Publication Date
In 2025 | 12 |
Descriptor
Test Reliability | 12 |
Scoring | 6 |
Test Validity | 6 |
Foreign Countries | 5 |
Scores | 5 |
Scoring Rubrics | 5 |
Item Response Theory | 4 |
Student Evaluation | 4 |
Evaluation Methods | 3 |
Test Construction | 3 |
Artificial Intelligence | 2 |
More ▼ |
Source
Author
Amirhossein Rasooli | 1 |
Ana Sánchez-Bello | 1 |
Andrea Fernández-Sánchez | 1 |
Arbaiah Inn | 1 |
Benjamin W. Domingue | 1 |
Bryce Hughes | 1 |
Dina Fitria Murad | 1 |
Doris Lee | 1 |
Edd Pitt | 1 |
Esmat Babaii | 1 |
Farshad Effatpanah | 1 |
More ▼ |
Publication Type
Reports - Research | 12 |
Journal Articles | 10 |
Tests/Questionnaires | 1 |
Education Level
Higher Education | 7 |
Postsecondary Education | 7 |
Early Childhood Education | 1 |
Elementary Education | 1 |
Grade 2 | 1 |
Grade 3 | 1 |
Grade 4 | 1 |
Grade 5 | 1 |
High Schools | 1 |
Intermediate Grades | 1 |
Middle Schools | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 1 |
What Works Clearinghouse Rating
Joshua B. Gilbert; James G. Soland; Benjamin W. Domingue – Annenberg Institute for School Reform at Brown University, 2025
Value-Added Models (VAMs) are both common and controversial in education policy and accountability research. While the sensitivity of VAMs to model specification and covariate selection is well documented, the extent to which test scoring methods (e.g., mean scores vs. IRT-based scores) may affect VA estimates is less studied. We examine the…
Descriptors: Value Added Models, Tests, Testing, Scoring
Joseph A. Rios; Jiayi Deng – Educational and Psychological Measurement, 2025
To mitigate the potential damaging consequences of rapid guessing (RG), a form of noneffortful responding, researchers have proposed a number of scoring approaches. The present simulation study examines the robustness of the most popular of these approaches, the unidimensional effort-moderated (EM) scoring procedure, to multidimensional RG (i.e.,…
Descriptors: Scoring, Guessing (Tests), Reaction Time, Item Response Theory
Wallace N. Pinto Jr.; Jinnie Shin – Journal of Educational Measurement, 2025
In recent years, the application of explainability techniques to automated essay scoring and automated short-answer grading (ASAG) models, particularly those based on transformer architectures, has gained significant attention. However, the reliability and consistency of these techniques remain underexplored. This study systematically investigates…
Descriptors: Automation, Grading, Computer Assisted Testing, Scoring
Reuben S. Asempapa; Doris Lee – Discover Education, 2025
Across the world, standards and practices for preparing teachers of mathematics emphasize the importance of math modeling (MM) in developing students' mathematical thinking. The aim of this research study was to develop the Mathematical Modeling Knowledge Scale (MAMKS), capable of determining preservice teachers' (PSTs') knowledge of MM. The study…
Descriptors: Preservice Teachers, Preservice Teacher Education, Mathematics Education, Mathematics Curriculum
Kylie Gorney; Sandip Sinharay – Journal of Educational Measurement, 2025
Although there exists an extensive amount of research on subscores and their properties, limited research has been conducted on categorical subscores and their interpretations. In this paper, we focus on the claim of Feinberg and von Davier that categorical subscores are useful for remediation and instructional purposes. We investigate this claim…
Descriptors: Tests, Scores, Test Interpretation, Alternative Assessment
Maria Blevins; Bryce Hughes; Jennifer Green; Leila Sterman; Shannon Willoughby – Journal of College Science Teaching, 2025
In this work, the authors document an expansion of the Public Speaking Competency Rubric (PSCR). First developed in 2012 by Schreiber, et al., the original rubric has only one item related to non-verbal communication. The authors of this work expanded the rubric to include 10 items related to the non-verbal aspects of public speaking and had it…
Descriptors: Test Construction, Public Speaking, Competence, Scoring Rubrics
Jeff Allen; Ty Cruce – ACT Education Corp., 2025
This report summarizes some of the evidence supporting interpretations of scores from the enhanced ACT, focusing on reliability, concurrent validity, predictive validity, and score comparability. The authors argue that the evidence presented in this report supports the interpretation of scores from the enhanced ACT as measures of high school…
Descriptors: College Entrance Examinations, Testing, Change, Scores
Andrea Fernández-Sánchez; Juan José Lorenzo-Castiñeiras; Ana Sánchez-Bello – European Journal of Education, 2025
The advent of artificial intelligence (AI) technologies heralds a transformative era in education. This study investigates the integration of AI tools in developing educational assessment rubrics within the 'Curriculum Design Development and Evaluation' course at the University of A Coruña during the 2023-2024 academic year. Employing an…
Descriptors: Foreign Countries, Higher Education, Artificial Intelligence, Technology Integration
Nicolas Rochat; Laurent Lima; Pascal Bressoux – Journal of Psychoeducational Assessment, 2025
Inference is considered an important factor in comprehension models and has been described as a causal factor in predicting comprehension. To date, specific tests for inference are rare and often rely on specific thematic texts. This reliance on thematic inference may raise some concerns as inference is related to prior text-specific knowledge.…
Descriptors: Inferences, Reading Comprehension, Reading Tests, Test Reliability
Farshad Effatpanah; Purya Baghaei; Mona Tabatabaee-Yazdi; Esmat Babaii – Language Testing, 2025
This study aimed to propose a new method for scoring C-Tests as measures of general language proficiency. In this approach, the unit of analysis is sentences rather than gaps or passages. That is, the gaps correctly reformulated in each sentence were aggregated as sentence score, and then each sentence was entered into the analysis as a polytomous…
Descriptors: Item Response Theory, Language Tests, Test Items, Test Construction
Dina Fitria Murad; Meta Amalya Dewi; Arbaiah Inn; Silvia Ayunda Murad; Noor Udin; Taufik Darwis – Journal of Educators Online, 2025
This study aims to produce a more personalized recommendation system for online learning using multicriteria in collaborative filtering and data from the Binus Online Learning repository as a knowledge base. The study uses forecasting (regression) and consists of three stages: (1) collecting data on the results of the learning process; (2) adding…
Descriptors: Electronic Learning, Data Collection, Context Effect, Learning Processes
Amirhossein Rasooli; Jim Turner; Tünde Varga-Atkins; Edd Pitt; Shaghayegh Asgari; Will Moindrot – Assessment & Evaluation in Higher Education, 2025
Groupwork is a crucial aspect of work contexts and a key twenty first century skill. Assessment of groupwork provides a persistent challenge for educators in university contexts with students reporting experiences of unfairness from their peers during groupwork. This study developed a novel Peer Assessment Fairness Instrument to explore factors…
Descriptors: Foreign Countries, Undergraduate Students, Student Attitudes, College Faculty