NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 85 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hayat, Bahrul – Cogent Education, 2022
The purpose of this study comprises (1) calibrating the Basic Statistics Test for Indonesian undergraduate psychology students using the Rasch model, (2) testing the impact of adjustment for guessing on item parameters, person parameters, test reliability, and distribution of item difficulty and person ability, and (3) comparing person scores…
Descriptors: Guessing (Tests), Statistics Education, Undergraduate Students, Psychology
Peer reviewed Peer reviewed
Direct linkDirect link
Dimitrov, Dimiter M.; Atanasov, Dimitar V.; Luo, Yong – Measurement: Interdisciplinary Research and Perspectives, 2020
This study examines and compares four person-fit statistics (PFSs) in the framework of the "D"- scoring method (DSM): (a) van der Flier's "U3" statistic; (b) "Ud" statistic, as a modification of "U3" under the DSM; (c) "Zd" statistic, as a modification of the "Z3 (l[subscript z])"…
Descriptors: Goodness of Fit, Item Analysis, Item Response Theory, Scoring
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Coniam, David; Lee, Tony; Lampropoulou, Leda – English Language Teaching, 2021
This article explores the issue of identifying guessers -- with a specific focus on multiple-choice tests. Guessing has long been considered a problem due to the fact that it compromises validity. A test taker scoring higher than they should through guessing does not provide a picture of their actual ability. After an initial description of issues…
Descriptors: Language Tests, Guessing (Tests), English (Second Language), Second Language Learning
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Sora; Bolt, Daniel M. – Journal of Educational Measurement, 2018
Both the statistical and interpretational shortcomings of the three-parameter logistic (3PL) model in accommodating guessing effects on multiple-choice items are well documented. We consider the use of a residual heteroscedasticity (RH) model as an alternative, and compare its performance to the 3PL with real test data sets and through simulation…
Descriptors: Statistical Analysis, Models, Guessing (Tests), Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Deribo, Tobias; Goldhammer, Frank; Kroehne, Ulf – Educational and Psychological Measurement, 2023
As researchers in the social sciences, we are often interested in studying not directly observable constructs through assessments and questionnaires. But even in a well-designed and well-implemented study, rapid-guessing behavior may occur. Under rapid-guessing behavior, a task is skimmed shortly but not read and engaged with in-depth. Hence, a…
Descriptors: Reaction Time, Guessing (Tests), Behavior Patterns, Bias
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Aborisade, Olatunbosun James; Fajobi, Olutoyin Olufunke – Educational Research and Reviews, 2020
West Africa Examination Council (WAEC) and National Examination Council (NECO) are the two major examination bodies saddled with the responsibility of awarding Senior Secondary School Certificate in Nigeria. This study examined the comparability of the psychometric properties of the items constructed by the two examination bodies using Item…
Descriptors: Foreign Countries, Mathematics Tests, Psychometrics, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Lúcio, Patrícia Silva; Vandekerckhove, Joachim; Polanczyk, Guilherme V.; Cogo-Moreira, Hugo – Journal of Psychoeducational Assessment, 2021
The present study compares the fit of two- and three-parameter logistic (2PL and 3PL) models of item response theory in the performance of preschool children on the Raven's Colored Progressive Matrices. The test of Raven is widely used for evaluating nonverbal intelligence of factor g. Studies comparing models with real data are scarce on the…
Descriptors: Guessing (Tests), Item Response Theory, Test Validity, Preschool Children
Peer reviewed Peer reviewed
Direct linkDirect link
Abulela, Mohammed A. A.; Rios, Joseph A. – Applied Measurement in Education, 2022
When there are no personal consequences associated with test performance for examinees, rapid guessing (RG) is a concern and can differ between subgroups. To date, the impact of differential RG on item-level measurement invariance has received minimal attention. To that end, a simulation study was conducted to examine the robustness of the…
Descriptors: Comparative Analysis, Robustness (Statistics), Nonparametric Statistics, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chu, Wei; Pavlik, Philip I., Jr. – International Educational Data Mining Society, 2023
In adaptive learning systems, various models are employed to obtain the optimal learning schedule and review for a specific learner. Models of learning are used to estimate the learner's current recall probability by incorporating features or predictors proposed by psychological theory or empirically relevant to learners' performance. Logistic…
Descriptors: Reaction Time, Accuracy, Models, Predictor Variables
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Asiret, Semih; Sünbül, Seçil Ömür – Educational Sciences: Theory and Practice, 2016
In this study, equating methods for random group design using small samples through factors such as sample size, difference in difficulty between forms, and guessing parameter was aimed for comparison. Moreover, which method gives better results under which conditions was also investigated. In this study, 5,000 dichotomous simulated data…
Descriptors: Equated Scores, Sample Size, Difficulty Level, Guessing (Tests)
Peer reviewed Peer reviewed
Direct linkDirect link
Goncher, Andrea M.; Boles, Wageeh – European Journal of Engineering Education, 2019
Concept inventories (CIs) are assessment instruments designed to measure students' conceptual understanding of fundamental concepts in particular fields. CIs utilise multiple-choice questions (MCQs), and specifically designed response selections, to help identify misconceptions. One shortcoming of this assessment instrument is that it fails to…
Descriptors: Engineering Education, Misconceptions, Concept Formation, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Ames, Allison; Smith, Elizabeth – Journal of Educational Measurement, 2018
Bayesian methods incorporate model parameter information prior to data collection. Eliciting information from content experts is an option, but has seen little implementation in Bayesian item response theory (IRT) modeling. This study aims to use ethical reasoning content experts to elicit prior information and incorporate this information into…
Descriptors: Item Response Theory, Bayesian Statistics, Ethics, Specialists
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Storme, Martin; Myszkowski, Nils; Baron, Simon; Bernard, David – Journal of Intelligence, 2019
Assessing job applicants' general mental ability online poses psychometric challenges due to the necessity of having brief but accurate tests. Recent research (Myszkowski & Storme, 2018) suggests that recovering distractor information through Nested Logit Models (NLM; Suh & Bolt, 2010) increases the reliability of ability estimates in…
Descriptors: Intelligence Tests, Item Response Theory, Comparative Analysis, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Marshall, Francisca B.; Marshall, Justin – Journal of Interactive Learning Research, 2021
The goal of this study was to explore how knowledge-check questions in video lectures affected learning. In a quasi-experimental study, six courses (n=84) were assigned to one of three groups: a control group and two treatment groups. The three groups saw the same video and knowledge-check questions. The three groups were evaluated with different…
Descriptors: Teaching Methods, Video Technology, Knowledge Level, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Bush, Martin – Assessment & Evaluation in Higher Education, 2015
The humble multiple-choice test is very widely used within education at all levels, but its susceptibility to guesswork makes it a suboptimal assessment tool. The reliability of a multiple-choice test is partly governed by the number of items it contains; however, longer tests are more time consuming to take, and for some subject areas, it can be…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Format, Test Reliability
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6