NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20251
Since 20242
Since 2021 (last 5 years)12
Since 2016 (last 10 years)15
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jila Niknejad; Margaret Bayer – International Journal of Mathematical Education in Science and Technology, 2025
In Spring 2020, the need for redesigning online assessments to preserve integrity became a priority to many educators. Many of us found methods to proctor examinations using Zoom and proctoring software. Such examinations pose their own issues. To reduce the technical difficulties and cost, many Zoom proctored examination sessions were shortened;…
Descriptors: Mathematics Instruction, Mathematics Tests, Computer Assisted Testing, Computer Software
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kilic, Abdullah Faruk; Uysal, Ibrahim – International Journal of Assessment Tools in Education, 2022
Most researchers investigate the corrected item-total correlation of items when analyzing item discrimination in multi-dimensional structures under the Classical Test Theory, which might lead to underestimating item discrimination, thereby removing items from the test. Researchers might investigate the corrected item-total correlation with the…
Descriptors: Item Analysis, Correlation, Item Response Theory, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Peabody, Michael R. – Measurement: Interdisciplinary Research and Perspectives, 2023
Many organizations utilize some form of automation in the test assembly process; either fully algorithmic or heuristically constructed. However, one issue with heuristic models is that when the test assembly problem changes the entire model may need to be re-conceptualized and recoded. In contrast, mixed-integer programming (MIP) is a mathematical…
Descriptors: Programming Languages, Algorithms, Heuristics, Mathematical Models
Peer reviewed Peer reviewed
Direct linkDirect link
Emery-Wetherell, Meaghan; Wang, Ruoyao – Assessment & Evaluation in Higher Education, 2023
Over four semesters of a large introductory statistics course the authors found students were engaging in contract cheating on Chegg.com during multiple choice examinations. In this paper we describe our methodology for identifying, addressing and eventually eliminating cheating. We successfully identified 23 out of 25 students using a combination…
Descriptors: Computer Assisted Testing, Multiple Choice Tests, Cheating, Identification
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Partchev, Ivailo – Journal of Intelligence, 2020
We analyze a 12-item version of Raven's Standard Progressive Matrices test, traditionally scored with the sum score. We discuss some important differences between assessment in practice and psychometric modelling. We demonstrate some advanced diagnostic tools in the freely available R package, dexter. We find that the first item in the test…
Descriptors: Intelligence Tests, Scores, Psychometrics, Diagnostic Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Musa Adekunle Ayanwale; Jamiu Oluwadamilare Amusa; Adekunle Ibrahim Oladejo; Funmilayo Ayedun – Interchange: A Quarterly Review of Education, 2024
The study focuses on assessing the proficiency levels of higher education students, specifically the physics achievement test (PHY 101) at the National Open University of Nigeria (NOUN). This test, like others, evaluates various aspects of knowledge and skills simultaneously. However, relying on traditional models for such tests can result in…
Descriptors: Item Response Theory, Difficulty Level, Item Analysis, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Saatcioglu, Fatima Munevver; Atar, Hakan Yavuz – International Journal of Assessment Tools in Education, 2022
This study aims to examine the effects of mixture item response theory (IRT) models on item parameter estimation and classification accuracy under different conditions. The manipulated variables of the simulation study are set as mixture IRT models (Rasch, 2PL, 3PL); sample size (600, 1000); the number of items (10, 30); the number of latent…
Descriptors: Accuracy, Classification, Item Response Theory, Programming Languages
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ayanwale, Musa Adekunle; Ndlovu, Mdutshekelwa – Education Sciences, 2021
This study investigated the scalability of a cognitive multiple-choice test through the Mokken package in the R programming language for statistical computing. A 2019 mathematics West African Examinations Council (WAEC) instrument was used to gather data from randomly drawn K-12 participants (N = 2866; Male = 1232; Female = 1634; Mean age = 16.5…
Descriptors: Cognitive Tests, Multiple Choice Tests, Scaling, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Becker, Benjamin; van Rijn, Peter; Molenaar, Dylan; Debeer, Dries – Assessment & Evaluation in Higher Education, 2022
A common approach to increase test security in higher educational high-stakes testing is the use of different test forms with identical items but different item orders. The effects of such varied item orders are relatively well studied, but findings have generally been mixed. When multiple test forms with different item orders are used, we argue…
Descriptors: Information Security, High Stakes Tests, Computer Security, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Harring, Jeffrey R.; Johnson, Tessa L. – Educational Measurement: Issues and Practice, 2020
In this digital ITEMS module, Dr. Jeffrey Harring and Ms. Tessa Johnson introduce the linear mixed effects (LME) model as a flexible general framework for simultaneously modeling continuous repeated measures data with a scientifically defensible function that adequately summarizes both individual change as well as the average response. The module…
Descriptors: Educational Assessment, Data Analysis, Longitudinal Studies, Case Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Lu, Owen H. T.; Huang, Anna Y. Q.; Tsai, Danny C. L.; Yang, Stephen J. H. – Educational Technology & Society, 2021
Human-guided machine learning can improve computing intelligence, and it can accurately assist humans in various tasks. In education research, artificial intelligence (AI) is applicable in many situations, such as predicting students' learning paths and strategies. In this study, we explore the benefits of repetitive practice of short-answer…
Descriptors: Test Items, Artificial Intelligence, Test Construction, Student Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Shahmirzadi, Niloufar – International Journal of Language Testing, 2023
The documentation of test takers' achievements has been accomplished through large-scale assessments to find general information about students' language ability. To remove subjectivity, Cognitive Diagnostic Assessment (CDA) has recently played a crucial role in perceiving candidates' latent attribute patterns to find multi-diagnostic information…
Descriptors: Placement Tests, Test Validity, Programming Languages, Diagnostic Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ehara, Yo – International Educational Data Mining Society, 2022
Language learners are underserved if there are unlearned meanings of a word that they think they have already learned. For example, "circle" as a noun is well known, whereas its use as a verb is not. For artificial-intelligence-based support systems for learning vocabulary, assessing each learner's knowledge of such atypical but common…
Descriptors: Language Tests, Vocabulary Development, Second Language Learning, Second Language Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Gregg, Nikole; Leventhal, Brian C. – Educational Measurement: Issues and Practice, 2020
In this digital ITEMS module, Nikole Gregg and Dr. Brian Leventhal discuss strategies to ensure data visualizations achieve graphical excellence. Data visualizations are commonly used by measurement professionals to communicate results to examinees, the public, educators, and other stakeholders. To do so effectively, it is important that these…
Descriptors: Data Analysis, Evidence Based Practice, Visualization, Test Results
Peer reviewed Peer reviewed
Direct linkDirect link
Nicklin, Christopher; Vitta, Joseph P. – Language Testing, 2022
Instrument measurement conducted with Rasch analysis is a common process in language assessment research. A recent systematic review of 215 studies involving Rasch analysis in language testing and applied linguistics research reported that 23 different software packages had been utilized. However, none of the analyses were conducted with one of…
Descriptors: Programming Languages, Vocabulary Development, Language Tests, Computer Software