NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
No Child Left Behind Act 20013
What Works Clearinghouse Rating
Showing 1 to 15 of 114 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Nan Xie; Zhengxu Li; Haipeng Lu; Wei Pang; Jiayin Song; Beier Lu – IEEE Transactions on Learning Technologies, 2025
Classroom engagement is a critical factor for evaluating students' learning outcomes and teachers' instructional strategies. Traditional methods for detecting classroom engagement, such as coding and questionnaires, are often limited by delays, subjectivity, and external interference. While some neural network models have been proposed to detect…
Descriptors: Learner Engagement, Artificial Intelligence, Technology Uses in Education, Educational Technology
Peer reviewed Peer reviewed
Direct linkDirect link
Benjamin R. Shear; Derek C. Briggs – Asia Pacific Education Review, 2024
Research in the social and behavioral sciences relies on a wide range of experimental and quasi-experimental designs to estimate the causal effects of specific programs, policies, and events. In this paper we highlight measurement issues relevant to evaluating the validity of causal estimation and generalization. These issues impact all four…
Descriptors: Measurement Techniques, Inferences, COVID-19, Pandemics
Peer reviewed Peer reviewed
Direct linkDirect link
Lyrica Lucas; Anum Khushal; Robert Mayes; Brian A. Couch; Joseph Dauer – International Journal of Science Education, 2025
Educational reform priorities such as emphasis on quantitative modelling (QM) have positioned undergraduate biology instructors as designers of QM experiences to engage students in authentic science practices that support the development of data-driven and evidence-based reasoning. Yet, little is known about how biology instructors adapt to the…
Descriptors: Undergraduate Students, College Science, Biology, Classroom Observation Techniques
Paul T. von Hippel; Brendan A. Schuetze – Annenberg Institute for School Reform at Brown University, 2025
Researchers across many fields have called for greater attention to heterogeneity of treatment effects--shifting focus from the average effect to variation in effects between different treatments, studies, or subgroups. True heterogeneity is important, but many reports of heterogeneity have proved to be false, non-replicable, or exaggerated. In…
Descriptors: Educational Research, Replication (Evaluation), Generalizability Theory, Inferences
Peer reviewed Peer reviewed
Direct linkDirect link
Hopster-den Otter, Dorien; Wools, Saskia; Eggen, Theo J. H. M.; Veldkamp, Bernard P. – Journal of Educational Measurement, 2019
In educational practice, test results are used for several purposes. However, validity research is especially focused on the validity of summative assessment. This article aimed to provide a general framework for validating formative assessment. The authors applied the argument-based approach to validation to the context of formative assessment.…
Descriptors: Formative Evaluation, Test Validity, Scores, Inferences
Peer reviewed Peer reviewed
Direct linkDirect link
Nicolas Rochat; Laurent Lima; Pascal Bressoux – Journal of Psychoeducational Assessment, 2025
Inference is considered an important factor in comprehension models and has been described as a causal factor in predicting comprehension. To date, specific tests for inference are rare and often rely on specific thematic texts. This reliance on thematic inference may raise some concerns as inference is related to prior text-specific knowledge.…
Descriptors: Inferences, Reading Comprehension, Reading Tests, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
An, Lily Shiao; Ho, Andrew Dean; Davis, Laurie Laughlin – Educational Measurement: Issues and Practice, 2022
Technical documentation for educational tests focuses primarily on properties of individual scores at single points in time. Reliability, standard errors of measurement, item parameter estimates, fit statistics, and linking constants are standard technical features that external stakeholders use to evaluate items and individual scale scores.…
Descriptors: Documentation, Scores, Evaluation Methods, Longitudinal Studies
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Seyedeh Azadeh Ghiasian; Fatemeh Hemmati; Seyyed Mohammad Alavi; Afsar Rouhi – International Journal of Language Testing, 2025
A critical component of cognitive diagnostic models (CDMs) is a Q-matrix that stipulates associations between items of a test and their required attributes. The present study aims to develop and empirically validate a Q-matrix for the listening comprehension section of the International English Language Testing System (IELTS). To this end, a…
Descriptors: Test Items, Listening Comprehension Tests, English (Second Language), Language Tests
Maddox, Bryan – OECD Publishing, 2023
The digital transition in educational testing has introduced many new opportunities for technology to enhance large-scale assessments. These include the potential to collect and use log data on test-taker response processes routinely, and on a large scale. Process data has long been recognised as a valuable source of validation evidence in…
Descriptors: Measurement, Inferences, Test Reliability, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Yan, Xun; Staples, Shelley – Language Testing, 2020
The argument-based approach to validity (Kane, 2013) focuses on two steps: (1) making claims about the proposed interpretation and use of test scores as a coherent, interpretive argument; and (2) evaluating those claims based on theoretical and empirical evidence related to test performances and scores. This paper discusses the role of…
Descriptors: Writing Tests, Language Tests, Language Proficiency, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Yaneva, Victoria; Clauser, Brian E.; Morales, Amy; Paniagua, Miguel – Journal of Educational Measurement, 2021
Eye-tracking technology can create a record of the location and duration of visual fixations as a test-taker reads test questions. Although the cognitive process the test-taker is using cannot be directly observed, eye-tracking data can support inferences about these unobserved cognitive processes. This type of information has the potential to…
Descriptors: Eye Movements, Test Validity, Multiple Choice Tests, Cognitive Processes
Peer reviewed Peer reviewed
Direct linkDirect link
Knoch, Ute; Chapelle, Carol A. – Language Testing, 2018
Argument-based validation requires test developers and researchers to specify what is entailed in test interpretation and use. Doing so has been shown to yield advantages (Chapelle, Enright, & Jamieson, 2010), but it also requires an analysis of how the concerns of language testers can be conceptualized in the terms used to construct a…
Descriptors: Test Validity, Language Tests, Evaluation Research, Rating Scales
Peer reviewed Peer reviewed
Direct linkDirect link
Shear, Benjamin R. – Journal of Educational Measurement, 2023
Large-scale standardized tests are regularly used to measure student achievement overall and for student subgroups. These uses assume tests provide comparable measures of outcomes across student subgroups, but prior research suggests score comparisons across gender groups may be complicated by the type of test items used. This paper presents…
Descriptors: Gender Bias, Item Analysis, Test Items, Achievement Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rodríguez-Vásquez, Flor Monserrat; Ariza-Hernandez, Francisco J. – EURASIA Journal of Mathematics, Science and Technology Education, 2021
The evaluation of learning in mathematics is a worldwide problem, therefore, new methods are required to assess the understanding of mathematical concepts. In this paper, we propose to use the Item Response Theory to analyze the understanding level of undergraduate students about the real function mathematical concept. The Bayesian approach was…
Descriptors: Bayesian Statistics, Mathematics Education, Item Response Theory, Undergraduate Students
Peer reviewed Peer reviewed
Direct linkDirect link
Jacobson, Erik; Svetina, Dubravka – Applied Measurement in Education, 2019
Contingent argument-based approaches to validity require a unique argument for each use, in contrast to more prescriptive approaches that identify the common kinds of validity evidence researchers should consider for every use. In this article, we evaluate our use of an approach that is both prescriptive "and" argument-based to develop a…
Descriptors: Test Validity, Test Items, Test Construction, Test Interpretation
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8