Publication Date
In 2025 | 12 |
Since 2024 | 27 |
Since 2021 (last 5 years) | 94 |
Since 2016 (last 10 years) | 242 |
Since 2006 (last 20 years) | 409 |
Descriptor
Test Reliability | 964 |
Scoring | 675 |
Test Validity | 560 |
Test Construction | 293 |
Scoring Rubrics | 154 |
Testing | 153 |
Scoring Formulas | 146 |
Test Items | 145 |
Test Interpretation | 126 |
Multiple Choice Tests | 109 |
Psychometrics | 107 |
More ▼ |
Source
Author
Publication Type
Education Level
Audience
Practitioners | 37 |
Researchers | 22 |
Teachers | 16 |
Administrators | 10 |
Policymakers | 7 |
Students | 4 |
Counselors | 1 |
Parents | 1 |
Location
New York | 16 |
California | 15 |
Turkey | 15 |
Florida | 11 |
Canada | 10 |
Nebraska | 8 |
Australia | 6 |
Pennsylvania | 6 |
United Kingdom | 6 |
United Kingdom (England) | 6 |
United States | 5 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 1 |
Meets WWC Standards with or without Reservations | 1 |
Jiayi Deng – ProQuest LLC, 2024
Test score comparability in international large-scale assessments (LSA) is of utmost importance in measuring the effectiveness of education systems and understanding the impact of education on economic growth. To effectively compare test scores on an international scale, score linking is widely used to convert raw scores from different linguistic…
Descriptors: Item Response Theory, Scoring Rubrics, Scoring, Error of Measurement
Sophie Litschwartz – Society for Research on Educational Effectiveness, 2021
Background/Context: Pass/fail standardized exams frequently selectively rescore failing exams and retest failing examinees. This practice distorts the test score distribution and can confuse those who do analysis on these distributions. In 2011, the Wall Street Journal showed large discontinuities in the New York City Regent test score…
Descriptors: Standardized Tests, Pass Fail Grading, Scoring Rubrics, Scoring Formulas
Steven Holtzman; Jonathan Steinberg; Jonathan Weeks; Christopher Robertson; Jessica Findley; David Klieger – ETS Research Report Series, 2024
At a time when institutions of higher education are exploring alternatives to traditional admissions testing, institutions are also seeking to better support students and prepare them for academic success. Under such an engaged model, one may seek to measure not just the accumulated knowledge and skills that students would bring to a new academic…
Descriptors: Law Schools, College Applicants, Legal Education (Professions), College Entrance Examinations
M. Arda Atakaya; Ugur Sak; M. Bahadir Ayas – Creativity Research Journal, 2024
Scoring in creativity research has been a central problem since creativity became an important issue in psychology and education in the 1950s. The current study examined the psychometric properties of 27 creativity indices derived from summed and averaged scores using 15 scoring methods. Participants included 2802 middle-school students. Data…
Descriptors: Psychometrics, Creativity, Creativity Tests, Scoring
National Institute for Excellence in Teaching, 2023
Aspiring teachers must develop an in-depth understanding of high-quality instructional practices. In order to prepare, instruct, and coach aspiring teachers, the National Institute for Excellence in Teaching (NIET) has developed a the NIET Aspiring Teacher Rubric (ATR) based on principles of excellence in instruction. This research brief…
Descriptors: Scoring Rubrics, Preservice Teachers, Test Construction, Test Validity
Joshua B. Gilbert; James G. Soland; Benjamin W. Domingue – Annenberg Institute for School Reform at Brown University, 2025
Value-Added Models (VAMs) are both common and controversial in education policy and accountability research. While the sensitivity of VAMs to model specification and covariate selection is well documented, the extent to which test scoring methods (e.g., mean scores vs. IRT-based scores) may affect VA estimates is less studied. We examine the…
Descriptors: Value Added Models, Tests, Testing, Scoring
Joseph A. Rios; Jiayi Deng – Educational and Psychological Measurement, 2025
To mitigate the potential damaging consequences of rapid guessing (RG), a form of noneffortful responding, researchers have proposed a number of scoring approaches. The present simulation study examines the robustness of the most popular of these approaches, the unidimensional effort-moderated (EM) scoring procedure, to multidimensional RG (i.e.,…
Descriptors: Scoring, Guessing (Tests), Reaction Time, Item Response Theory
Kroc, Edward; Olvera Astivia, Oscar L. – Educational and Psychological Measurement, 2022
Setting cutoff scores is one of the most common practices when using scales to aid in classification purposes. This process is usually done univariately where each optimal cutoff value is decided sequentially, subscale by subscale. While it is widely known that this process necessarily reduces the probability of "passing" such a test,…
Descriptors: Multivariate Analysis, Cutting Scores, Classification, Measurement
Herwin, Herwin; Pristiwaluyo, Triyanto; Ruslan, Ruslan; Dahalan, Shakila Che – Cypriot Journal of Educational Sciences, 2022
The application of multiple-choice tests often does not consider the scoring technique and the number of choices. The study aims at describing the effect of the scoring technique and numerous options towards the reliability of multiple-choice objective tests on social subjects in elementary school. The study is quantitative research with…
Descriptors: Scoring, Multiple Choice Tests, Test Reliability, Elementary School Students
Sievers, Matt; Reemts, Connor; Dickinson, Katherine J.; Mukerji, Joya; Beltran, Ismael Barreras; Theobald, Elli J.; Velasco, Vicente; Freeman, Scott – Biochemistry and Molecular Biology Education, 2023
Researchers have called for undergraduate courses to update teaching frameworks based on the Modern Synthesis with insights from molecular biology, by stressing the molecular underpinnings of variation and adaptation. To support this goal, we developed a modified version of the widely used Assessing Conceptual Reasoning of Natural Selection…
Descriptors: Student Evaluation, Knowledge Level, Molecular Biology, Evolution
Wallace N. Pinto Jr.; Jinnie Shin – Journal of Educational Measurement, 2025
In recent years, the application of explainability techniques to automated essay scoring and automated short-answer grading (ASAG) models, particularly those based on transformer architectures, has gained significant attention. However, the reliability and consistency of these techniques remain underexplored. This study systematically investigates…
Descriptors: Automation, Grading, Computer Assisted Testing, Scoring
Marcos Jiménez; María Zapata-Cáceres; Marcos Román-González; Gregorio Robles; Jesús Moreno-León; Estefanía Martín-Barroso – Journal of Science Education and Technology, 2024
Computational thinking (CT) is a multidimensional term that encompasses a wide variety of problem-solving skills related to the field of computer science. Unfortunately, standardized, valid, and reliable methods to assess CT skills in preschool children are lacking, compromising the reliability of the results reported in CT interventions. To…
Descriptors: Computation, Thinking Skills, Student Evaluation, Preschool Children
Susan K. Johnsen – Gifted Child Today, 2024
The author provides a checklist for educators who are selecting technically adequate tests for identifying and referring students for gifted education services and programs. The checklist includes questions related to how the test was normed, reliability and validity studies as well as questions related to types of scores, administration, and…
Descriptors: Test Selection, Academically Gifted, Gifted Education, Test Validity
Marcelo Fernando Rauber; Christiane Gresse von Wangenheim; Pedro Alberto Barbetta; Adriano Ferreti Borgatto; Ramon Mayor Martins; Jean Carlo Rossa Hauck – Informatics in Education, 2024
The insertion of Machine Learning (ML) in everyday life demonstrates the importance of popularizing an understanding of ML already in school. Accompanying this trend arises the need to assess the students' learning. Yet, so far, few assessments have been proposed, most lacking an evaluation. Therefore, we evaluate the reliability and validity of…
Descriptors: Artificial Intelligence, Measures (Individuals), Test Reliability, Test Validity
Mohammad Hmoud; Hadeel Swaity; Eman Anjass; Eva María Aguaded-Ramírez – Electronic Journal of e-Learning, 2024
This research aimed to develop and validate a rubric to assess Artificial Intelligence (AI) chatbots' effectiveness in accomplishing tasks, particularly within educational contexts. Given the rapidly growing integration of AI in various sectors, including education, a systematic and robust tool for evaluating AI chatbot performance is essential.…
Descriptors: Artificial Intelligence, Man Machine Systems, Natural Language Processing, Test Construction