NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 202510
Since 202415
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Aiman Mohammad Freihat; Omar Saleh Bani Yassin – Educational Process: International Journal, 2025
Background/purpose: This study aimed to reveal the accuracy of estimation of multiple-choice test items parameters following the models of the item-response theory in measurement. Materials/methods: The researchers depended on the measurement accuracy indicators, which express the absolute difference between the estimated and actual values of the…
Descriptors: Accuracy, Computation, Multiple Choice Tests, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Neda Kianinezhad; Mohsen Kianinezhad – Language Education & Assessment, 2025
This study presents a comparative analysis of classical reliability measures, including Cronbach's alpha, test-retest, and parallel forms reliability, alongside modern psychometric methods such as the Rasch model and Mokken scaling, to evaluate the reliability of C-tests in language proficiency assessment. Utilizing data from 150 participants…
Descriptors: Psychometrics, Test Reliability, Language Proficiency, Language Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Katrin Schuessler; Vanessa Fischer; Maik Walpuski – Instructional Science: An International Journal of the Learning Sciences, 2025
Cognitive load studies are mostly centered on information on perceived cognitive load. Single-item subjective rating scales are the dominant measurement practice to investigate overall cognitive load. Usually, either invested mental effort or perceived task difficulty is used as an overall cognitive load measure. However, the extent to which the…
Descriptors: Cognitive Processes, Difficulty Level, Rating Scales, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Stephen Humphry; Paul Montuoro; Carolyn Maxwell – Journal of Psychoeducational Assessment, 2024
This article builds upon a proiminent definition of construct validity that focuses on variation in attributes causing variation in measurement outcomes. This article synthesizes the defintion and uses Rasch measurement modeling to explicate a modified conceptualization of construct validity for assessments of developmental attributes. If…
Descriptors: Construct Validity, Measurement Techniques, Developmental Stages, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Apichat Khamboonruang – Language Testing in Asia, 2025
Chulalongkorn University Language Institute (CULI) test was developed as a local standardised test of English for professional and international communication. To ensure that the CULI test fulfils its intended purposes, this study employed Kane's argument-based validation and Rasch measurement approaches to construct the validity argument for the…
Descriptors: Universities, Second Language Learning, Second Language Instruction, Language Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Tomás Hyland; Niall Seery; Jeffrey Buckley – European Journal of Engineering Education, 2024
An association between spatial ability, particularly visualisation, and science, technology, engineering, and mathematics (STEM) education has been repeatedly demonstrated over the past four decades. Understanding this relationship is of critical importance as substantial evidence illustrates the malleability of spatial ability and that the…
Descriptors: Spatial Ability, Predictor Variables, Retention (Psychology), Visualization
Peer reviewed Peer reviewed
Direct linkDirect link
Sophie Langhorne; Nora Uglik-Marucha; Charlotte Broadhurst; Elena Lieven; Amelia Pearson; Silia Vitoratou; Kathy Leadbitter – Journal of Autism and Developmental Disorders, 2025
Tools to measure autism knowledge are needed to assess levels of understanding within particular groups of people and to evaluate whether awareness-raising campaigns or interventions lead to improvements in understanding. Several such measures are in circulation, but, to our knowledge, there are no psychometrically-validated questionnaires that…
Descriptors: Foreign Countries, Autism Spectrum Disorders, Questionnaires, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Cui-Yan Hoe; Chieh-Yu Chen; Ching-I Chen – Infants and Young Children, 2025
The Ages and Stages Questionnaires: Social-Emotional, Second Edition (ASQ:SE-2) has been translated into Traditional Chinese (ASQ:SE-2-TC) in Taiwan. This study investigated whether the ASQ:SE-2-TC is also suitable for use in Malaysian Chinese families, and if any cultural differences are presented in ASQ:SE-2-TC items. This study analyzed the…
Descriptors: Social Emotional Learning, Child Development, Screening Tests, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Rodrigo Moreta-Herrera; Xavier Oriol-Granado; Mònica González; Jose A. Rodas – Infant and Child Development, 2025
This study evaluates the Children's Worlds Psychological Well-Being Scale (CW-PSWBS) within a diverse international cohort of children aged 10 and 12, utilising Classical Test Theory (CTT) and Item Response Theory (IRT) methodologies. Through a detailed psychometric analysis, this research assesses the CW-PSWBS's structural integrity, focusing on…
Descriptors: Well Being, Rating Scales, Children, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mimi Ismail; Ahmed Al - Badri; Said Al - Senaidi – Journal of Education and e-Learning Research, 2025
This study aimed to reveal the differences in individuals' abilities, their standard errors, and the psychometric properties of the test according to the two methods of applying the test (electronic and paper). The descriptive approach was used to achieve the study's objectives. The study sample consisted of 74 male and female students at the…
Descriptors: Achievement Tests, Computer Assisted Testing, Psychometrics, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
David R. Moore; Li Lin; Ritu Bhalerao; Jody Caldwell-Kurtzman; Lisa L. Hunter – Journal of Speech, Language, and Hearing Research, 2025
Purpose: Listening difficulty (LiD), often classified as auditory processing disorder (APD), has been studied in both research and clinic settings. The aim of this study was to examine the predictive relation between these two settings. In our SICLiD (Sensitive Indicators of Childhood Listening Difficulties) research study, children with normal…
Descriptors: Interdisciplinary Approach, Clinical Diagnosis, Educational Diagnosis, Intervention
Peer reviewed Peer reviewed
Direct linkDirect link
Joshua B. Gilbert; Luke W. Miratrix; Mridul Joshi; Benjamin W. Domingue – Journal of Educational and Behavioral Statistics, 2025
Analyzing heterogeneous treatment effects (HTEs) plays a crucial role in understanding the impacts of educational interventions. A standard practice for HTE analysis is to examine interactions between treatment status and preintervention participant characteristics, such as pretest scores, to identify how different groups respond to treatment.…
Descriptors: Causal Models, Item Response Theory, Statistical Inference, Psychometrics
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Isolda Margarita Castillo-Martínez; Davis Velarde-Camaqui; María Soledad Ramírez-Montoya; Jorge Sanabria-Z – Journal of Social Studies Education Research, 2024
Reasoning for complexity is a fundamental competency in these complex times for solutions to social problems and decision-making. The purpose of this paper is to demonstrate the validity and reliability of the eComplexity instrument by presenting its psychometric properties. The instrument consists of a Likert-type scale questionnaire designed to…
Descriptors: Psychometrics, Test Validity, Test Reliability, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Qi Huang; Daniel M. Bolt; Weicong Lyu – Large-scale Assessments in Education, 2024
Large scale international assessments depend on invariance of measurement across countries. An important consideration when observing cross-national differential item functioning (DIF) is whether the DIF actually reflects a source of bias, or might instead be a methodological artifact reflecting item response theory (IRT) model misspecification.…
Descriptors: Test Items, Item Response Theory, Test Bias, Test Validity
Joshua B. Gilbert; Luke W. Miratrix; Mridul Joshi; Benjamin W. Domingue – Annenberg Institute for School Reform at Brown University, 2024
Analyzing heterogeneous treatment effects (HTE) plays a crucial role in understanding the impacts of educational interventions. A standard practice for HTE analysis is to examine interactions between treatment status and pre-intervention participant characteristics, such as pretest scores, to identify how different groups respond to treatment.…
Descriptors: Causal Models, Item Response Theory, Statistical Inference, Psychometrics