Publication Date
In 2025 | 0 |
Since 2024 | 3 |
Since 2021 (last 5 years) | 23 |
Since 2016 (last 10 years) | 52 |
Since 2006 (last 20 years) | 92 |
Descriptor
Multiple Choice Tests | 435 |
Test Items | 149 |
Test Construction | 137 |
Higher Education | 127 |
Test Format | 98 |
Comparative Analysis | 67 |
Difficulty Level | 66 |
Test Validity | 64 |
Test Reliability | 63 |
Foreign Countries | 61 |
Item Analysis | 56 |
More ▼ |
Source
Author
Publication Type
Education Level
Audience
Researchers | 39 |
Practitioners | 10 |
Teachers | 5 |
Administrators | 1 |
Counselors | 1 |
Location
Australia | 8 |
Canada | 5 |
Netherlands | 5 |
Turkey | 5 |
Ireland | 3 |
Spain | 3 |
Albania | 2 |
Austria | 2 |
Connecticut | 2 |
Florida | 2 |
Illinois | 2 |
More ▼ |
Laws, Policies, & Programs
Civil Rights Act 1964 Title VI | 1 |
Civil Rights Act 1964 Title… | 1 |
Goals 2000 | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Andreea Dutulescu; Stefan Ruseti; Denis Iorga; Mihai Dascalu; Danielle S. McNamara – Grantee Submission, 2024
The process of generating challenging and appropriate distractors for multiple-choice questions is a complex and time-consuming task. Existing methods for an automated generation have limitations in proposing challenging distractors, or they fail to effectively filter out incorrect choices that closely resemble the correct answer, share synonymous…
Descriptors: Multiple Choice Tests, Artificial Intelligence, Attention, Natural Language Processing
Olney, Andrew M. – Grantee Submission, 2022
Multi-angle question answering models have recently been proposed that promise to perform related tasks like question generation. However, performance on related tasks has not been thoroughly studied. We investigate a leading model called Macaw on the task of multiple choice question generation and evaluate its performance on three angles that…
Descriptors: Test Construction, Multiple Choice Tests, Test Items, Models
Dave, Neisarg; Bakes, Riley; Pursel, Barton; Giles, C. Lee – International Educational Data Mining Society, 2021
We investigate encoder-decoder GRU networks with attention mechanism for solving a diverse array of elementary math problems with mathematical symbolic structures. We quantitatively measure performances of recurrent models on a given question type using a test set of unseen problems with a binary scoring and partial credit system. From our…
Descriptors: Multiple Choice Tests, Mathematics Tests, Problem Solving, Attention
Andrew M. Olney – Grantee Submission, 2023
Multiple choice questions are traditionally expensive to produce. Recent advances in large language models (LLMs) have led to fine-tuned LLMs that generate questions competitive with human-authored questions. However, the relative capabilities of ChatGPT-family models have not yet been established for this task. We present a carefully-controlled…
Descriptors: Test Construction, Multiple Choice Tests, Test Items, Algorithms
Sibic, Okan; Sesen, Burcin Acar – International Journal of Assessment Tools in Education, 2022
One of the main goals of science education is to make students gain science process skills. Thus, it is significant to measure whether students gain those skills or not. For this purpose, various tests have been produced and used in various studies. This study aims to examine science process skills tests which have been used in the theses produced…
Descriptors: Foreign Countries, Science Education, Science Process Skills, Masters Theses
Herrmann-Abell, Cari F.; Hardcastle, Joseph; DeBoer, George E. – Grantee Submission, 2022
As implementation of the "Next Generation Science Standards" moves forward, there is a need for new assessments that can measure students' integrated three-dimensional science learning. The National Research Council has suggested that these assessments be multicomponent tasks that utilize a combination of item formats including…
Descriptors: Multiple Choice Tests, Conditioning, Test Items, Item Response Theory
Ashish Gurung; Kirk Vanacore; Andrew A. McReynolds; Korinn S. Ostrow; Eamon S. Worden; Adam C. Sales; Neil T. Heffernan – Grantee Submission, 2024
Learning experience designers consistently balance the trade-off between open and close-ended activities. The growth and scalability of Computer Based Learning Platforms (CBLPs) have only magnified the importance of these design trade-offs. CBLPs often utilize close-ended activities (i.e. Multiple-Choice Questions [MCQs]) due to feasibility…
Descriptors: Multiple Choice Tests, Testing, Test Format, Computer Assisted Testing
Cari F. Herrmann-Abell; George E. DeBoer – Grantee Submission, 2023
This study describes the role that Rasch measurement played in the development of assessments aligned to the "Next Generation Science Standards," tasks that require students to use the three dimensions of science practices, disciplinary core ideas and cross-cutting concepts to make sense of energy-related phenomena. A set of 27…
Descriptors: Item Response Theory, Computer Simulation, Science Tests, Energy
Josef Guggemos; Roman Rietsche; Stephan Aier; Jannis Strecker; Simon Mayer – International Association for Development of the Information Society, 2024
Technological advancements, particularly in artificial intelligence, significantly transform our society and work practices. Computational thinking (CT) has emerged as a crucial 21st-century skill, enabling individuals to solve problems more effectively through an automation-oriented perspective and fundamental concepts of computer science. To…
Descriptors: Computation, Thinking Skills, 21st Century Skills, Test Construction
Walter M. Stroup; Anthony Petrosino; Corey Brady; Karen Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Tests of statistical significance often play a decisive role in establishing the empirical warrant of evidence-based research in education. The results from pattern-based assessment items, as introduced in this paper, are categorical and multimodal and do not immediately support the use of measures of central tendency as typically related to…
Descriptors: Statistical Significance, Comparative Analysis, Research Methodology, Evaluation Methods
Tomkowicz, Joanna; Kim, Dong-In; Wan, Ping – Online Submission, 2022
In this study we evaluated the stability of item parameters and student scores, using the pre-equated (pre-pandemic) parameters from Spring 2019 and post-equated (post-pandemic) parameters from Spring 2021 in two calibration and equating designs related to item parameter treatment: re-estimating all anchor parameters (Design 1) and holding the…
Descriptors: Equated Scores, Test Items, Evaluation Methods, Pandemics
Joe Olsen; Amy Adair; Janice Gobert; Michael Sao Pedro; Mariel O'Brien – Grantee Submission, 2022
Many national science frameworks (e.g., Next Generation Science Standards) argue that developing mathematical modeling competencies is critical for students' deep understanding of science. However, science teachers may be unprepared to assess these competencies. We are addressing this need by developing virtual lab performance assessments that…
Descriptors: Mathematical Models, Intelligent Tutoring Systems, Performance Based Assessment, Data Collection
Nur Ainil Sulaiman; Nor Azwahanum Nor Shaid; Chua Pei Chia – International Society for Technology, Education, and Science, 2023
Nowadays, Malaysian ESL students struggle to read English texts, particularly narrative texts. Lacking in utilising and applying the suitable reading strategies are among factors that hinder reading comprehension. This study aimed to investigate the effectiveness of Question-Answer Relationship (QAR) strategy on ESL pupils' narrative reading and…
Descriptors: Foreign Countries, Second Language Learning, English (Second Language), Reading Strategies
PaaBen, Benjamin; Dywel, Malwina; Fleckenstein, Melanie; Pinkwart, Niels – International Educational Data Mining Society, 2022
Item response theory (IRT) is a popular method to infer student abilities and item difficulties from observed test responses. However, IRT struggles with two challenges: How to map items to skills if multiple skills are present? And how to infer the ability of new students that have not been part of the training data? Inspired by recent advances…
Descriptors: Item Response Theory, Test Items, Item Analysis, Inferences
Herrmann-Abell, Cari F.; Hardcastle, Joseph; DeBoer, George E. – Grantee Submission, 2019
The "Next Generation Science Standards" calls for new assessments that measure students' integrated three-dimensional science learning. The National Research Council has suggested that these assessments utilize a combination of item formats including constructed-response and multiple-choice. In this study, students were randomly assigned…
Descriptors: Science Tests, Multiple Choice Tests, Test Format, Test Items