NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 52 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Youmi Suk – Asia Pacific Education Review, 2024
Regression discontinuity (RD) designs have gained significant popularity as a quasi-experimental device for evaluating education programs and policies. In this paper, we present a comprehensive review of RD designs, focusing on the continuity-based framework, the most widely adopted RD framework. We first review the fundamental aspects of RD…
Descriptors: Educational Research, Preschool Education, Regression (Statistics), Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Philipp Sterner; Kim De Roover; David Goretzko – Structural Equation Modeling: A Multidisciplinary Journal, 2025
When comparing relations and means of latent variables, it is important to establish measurement invariance (MI). Most methods to assess MI are based on confirmatory factor analysis (CFA). Recently, new methods have been developed based on exploratory factor analysis (EFA); most notably, as extensions of multi-group EFA, researchers introduced…
Descriptors: Error of Measurement, Measurement Techniques, Factor Analysis, Structural Equation Models
Peer reviewed Peer reviewed
Direct linkDirect link
Sternberg, Robert J. – Journal of Creative Behavior, 2020
Creativity testing as it is now done is often based on a defective assumption that different kinds of creativity can be compressed into a single unidimensional scale. There is no reason to believe that the different kinds of creativity represent, simply, different amounts of a single unidimensional construct. The article shows how three different…
Descriptors: Creativity Tests, Test Validity, Misconceptions, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Yan Jin; Jason Fan – Language Assessment Quarterly, 2023
In language assessment, AI technology has been incorporated in task design, assessment delivery, automated scoring of performance-based tasks, score reporting, and provision of feedback. AI technology is also used for collecting and analyzing performance data in language assessment validation. Research has been conducted to investigate the…
Descriptors: Language Tests, Artificial Intelligence, Computer Assisted Testing, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Andrew P. Jaciw – American Journal of Evaluation, 2025
By design, randomized experiments (XPs) rule out bias from confounded selection of participants into conditions. Quasi-experiments (QEs) are often considered second-best because they do not share this benefit. However, when results from XPs are used to generalize causal impacts, the benefit from unconfounded selection into conditions may be offset…
Descriptors: Elementary School Students, Elementary School Teachers, Generalization, Test Bias
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ketabi, Somaye; Alavi, Seyyed Mohammed; Ravand, Hamdollah – International Journal of Language Testing, 2021
Although Diagnostic Classification Models (DCMs) were introduced to education system decades ago, it seems that these models were not employed for the original aims upon which they had been designed. Using DCMs has been mostly common in analyzing large-scale non-diagnostic tests and these models have been rarely used in developing Cognitive…
Descriptors: Diagnostic Tests, Test Construction, Goodness of Fit, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Lewis, Todd F. – Measurement and Evaluation in Counseling and Development, 2017
American Educational Research Association (AERA) standards stipulate that researchers show evidence of the internal structure of instruments. Confirmatory factor analysis (CFA) is one structural equation modeling procedure designed to assess construct validity of assessments that has broad applicability for counselors interested in instrument…
Descriptors: Educational Research, Factor Analysis, Structural Equation Models, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Dickison, Philip; Luo, Xiao; Kim, Doyoung; Woo, Ada; Muntean, William; Bergstrom, Betty – Journal of Applied Testing Technology, 2016
Designing a theory-based assessment with sound psychometric qualities to measure a higher-order cognitive construct is a highly desired yet challenging task for many practitioners. This paper proposes a framework for designing a theory-based assessment to measure a higher-order cognitive construct. This framework results in a modularized yet…
Descriptors: Thinking Skills, Cognitive Tests, Test Construction, Nursing
Peer reviewed Peer reviewed
Direct linkDirect link
von Davier, Matthias; Khorramdel, Lale; He, Qiwei; Shin, Hyo Jeong; Chen, Haiwen – Journal of Educational and Behavioral Statistics, 2019
International large-scale assessments (ILSAs) transitioned from paper-based assessments to computer-based assessments (CBAs) facilitating the use of new item types and more effective data collection tools. This allows implementation of more complex test designs and to collect process and response time (RT) data. These new data types can be used to…
Descriptors: International Assessment, Computer Assisted Testing, Psychometrics, Item Response Theory
Carlson, Tiffany; Crepeau-Hobson, Franci – Communique, 2021
When the coronavirus pandemic was declared a public health crisis in March 2020, school psychologists were forced into situations where face-to-face interaction with their students was discouraged and in some cases, prohibited. Consequently, the traditional practice of school psychology abruptly ended. Individualized Education Plans (IEP) and…
Descriptors: Cognitive Tests, Ethics, Decision Making, Models
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Kaufman, Alan S. – Journal of Intelligence, 2021
U.S. Supreme Court justices and other federal judges are, effectively, appointed for life, with no built-in check on their cognitive functioning as they approach old age. There is about a century of research on aging and intelligence that shows the vulnerability of processing speed, fluid reasoning, visual-spatial processing, and working memory to…
Descriptors: Judges, Federal Government, Aging (Individuals), Decision Making
Peer reviewed Peer reviewed
Direct linkDirect link
Bao, Lei; Koenig, Kathleen; Xiao, Yang; Fritchman, Joseph; Zhou, Shaona; Chen, Cheng – Physical Review Physics Education Research, 2022
Abilities in scientific thinking and reasoning have been emphasized as core areas of initiatives, such as the Next Generation Science Standards or the College Board Standards for College Success in Science, which focus on the skills the future will demand of today's students. Although there is rich literature on studies of how these abilities…
Descriptors: Physics, Science Instruction, Teaching Methods, Thinking Skills
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Brijmohan, Amanda; Khan, Gulam A.; Orpwood, Graham; Brown, Emily Sandford; Childs, Ruth A. – Canadian Journal of Education, 2018
Developing a new assessment requires the expertise of both content experts and assessment specialists. Using the example of an assessment developed for Ontario's Colleges Mathematics Assessment Program (CMAP), this article (1) describes the decisions that must be made in developing a new assessment, (2) explores the complementary contributions of…
Descriptors: Expertise, Mathematics Instruction, College Mathematics, College Students
Hutchins, Shaun D. – Online Submission, 2019
The purpose of this Professional Pathways for Teachers (PPfT) evaluation was to examine the measurement validity and reliability of PPfT appraisal data from the 2017-2018 school year in the Austin Independent School District. The PPfT appraisal is a multi-measure system that covers three areas: instructional practices (IP), professional growth and…
Descriptors: Test Validity, Test Reliability, School Districts, Teacher Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Baghaei, Purya; Kubinger, Klaus D. – Practical Assessment, Research & Evaluation, 2015
The present paper gives a general introduction to the linear logistic test model (Fischer, 1973), an extension of the Rasch model with linear constraints on item parameters, along with eRm (an R package to estimate different types of Rasch models; Mair, Hatzinger, & Mair, 2014) functions to estimate the model and interpret its parameters. The…
Descriptors: Item Response Theory, Models, Test Validity, Hypothesis Testing
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4