NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Policymakers1
Laws, Policies, & Programs
No Child Left Behind Act 20013
What Works Clearinghouse Rating
Showing 1 to 15 of 33 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
He, Yinhong – Journal of Educational Measurement, 2023
Back random responding (BRR) behavior is one of the commonly observed careless response behaviors. Accurately detecting BRR behavior can improve test validities. Yu and Cheng (2019) showed that the change point analysis (CPA) procedure based on weighted residual (CPA-WR) performed well in detecting BRR. Compared with the CPA procedure, the…
Descriptors: Test Validity, Item Response Theory, Measurement, Monte Carlo Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Stephen M. Leach; Jason C. Immekus; Jeffrey C. Valentine; Prathiba Batley; Dena Dossett; Tamara Lewis; Thomas Reece – Assessment for Effective Intervention, 2025
Educators commonly use school climate survey scores to inform and evaluate interventions for equitably improving learning and reducing educational disparities. Unfortunately, validity evidence to support these (and other) score uses often falls short. In response, Whitehouse et al. proposed a collaborative, two-part validity testing framework for…
Descriptors: School Surveys, Measurement, Hierarchical Linear Modeling, Educational Environment
Peer reviewed Peer reviewed
Direct linkDirect link
Leder, Gilah C.; Forgasz, Helen J. – ZDM: The International Journal on Mathematics Education, 2018
Assessment in mathematics is assumed to provide credible and important information about what students know and can do. In this paper we focus on large scale tests and question whether mathematics assessment is essentially gender neutral. We consider aspects of test validity and discuss issues of terminology related to gender and mathematics. In…
Descriptors: Mathematics Education, Evaluation Methods, Gender Bias, Test Content
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Keng-Lin; Tsai, Shih-Li; Chiu, Yu-Ting; Ho, Ming-Jung – Advances in Health Sciences Education, 2016
Measurement invariance is a prerequisite for comparing measurement scores from different groups. In medical education, multi-source feedback (MSF) is utilized to assess core competencies, including the professionalism. However, little attention has been paid to the measurement invariance of assessment instruments; that is, whether an instrument…
Descriptors: Measurement, Scores, Medical Education, Competence
Peer reviewed Peer reviewed
Direct linkDirect link
Algozzine, Bob; Horner, Robert H.; Todd, Anne W.; Newton, J. Stephen; Algozzine, Kate; Cusumano, Dale – Journal of Psychoeducational Assessment, 2016
Although there is a strong legislative base and perceived efficacy for multidisciplinary team decision making, limited evidence supports its effectiveness or consistency of implementation in practice. In recent research, we used the Decision Observation, Recording, and Analysis (DORA) tool to document activities and adult behaviors during positive…
Descriptors: Problem Solving, Participative Decision Making, Positive Behavior Supports, Meetings
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Literat, Ioana – Journal of Media Literacy Education, 2014
This study assesses the psychometric properties of a newly tested self-report assessment tool for media literacy, based on the twelve new media literacy skills (NMLs) developed by Jenkins et al. (2006). The sample (N = 327) consisted of normal volunteers who completed a comprehensive online survey that measured their NML skills, media exposure,…
Descriptors: Media Literacy, Measurement, Measurement Techniques, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Sandilands, Debra; Oliveri, Maria Elena; Zumbo, Bruno D.; Ercikan, Kadriye – International Journal of Testing, 2013
International large-scale assessments of achievement often have a large degree of differential item functioning (DIF) between countries, which can threaten score equivalence and reduce the validity of inferences based on comparisons of group performances. It is important to understand potential sources of DIF to improve the validity of future…
Descriptors: Validity, Measures (Individuals), International Studies, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Pellegrino, James W. – Journal of Research in Science Teaching, 2012
Beginning with a reference to living in a time of both uncertainty and opportunity, this article presents a discussion of key areas where shared understanding is needed if we are to successfully realize the design and use of high quality, valid assessments of science. The key areas discussed are: (1) assessment purpose and use, (2) the nature of…
Descriptors: Science Education, Science and Society, Academic Standards, State Standards
Peer reviewed Peer reviewed
Direct linkDirect link
Juni, Samuel; Trobliger, Robert – Canadian Journal of School Psychology, 2009
The analysis of response inconsistency is a crucial aspect of intellectual and clinical psychological assessment. Erratic patterns of failures and successes across and within particular domains qualify the measurement of intellectual potential and functioning. Although the interpretation of intertest scatter (inconsistencies between subtest…
Descriptors: Intelligence, Psychological Evaluation, Codification, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Browne, Jeremy – Journal of Technology and Teacher Education, 2011
The need for rigorously developed measures of preservice teacher traits regarding technology integration training has been acknowledged (Kay 2006), but such instruments are still extremely rare. The Technology Integration Confidence Scale (TICS) represents one such measure, but past analyses of its functioning have been limited by sample size and…
Descriptors: Educational Technology, Technology Integration, Preservice Teachers, Self Efficacy
Peer reviewed Peer reviewed
Direct linkDirect link
Nickerson, Amanda B.; Fishman, Callen – School Psychology Quarterly, 2009
This study assessed the convergent and divergent validity of the Devereux Student Strengths Assessment (DESSA; LeBuffe, Shapiro, & Naglieri, 2008). A total of 227 (n = 94 teachers, n = 133 parents) raters completed the DESSA, in addition to the Behavioral and Emotional Rating Scales-2 (BERS-2; Epstein, 2004) and/or the Behavior Assessment…
Descriptors: Test Validity, Rating Scales, Correlation, Interrater Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Lockwood, J. R.; McCaffrey, Daniel F. – Education Finance and Policy, 2009
This article develops a model for longitudinal student achievement data designed to estimate heterogeneity in teacher effects across students of different achievement levels. The model specifies interactions between teacher effects and students' predicted scores on a test, estimating both average effects of individual teachers and interaction…
Descriptors: Models, Achievement, Test Validity, Educational Policy
Russell, Michael; Kavanaugh, Maureen – IAP - Information Age Publishing, Inc., 2011
The importance of student assessment, particularly for summative purposes, has increased greatly over the past thirty years. At the same time, emphasis on including all students in assessment programs has also increased. Assessment programs, whether they are large-scale, district-based, or teacher developed, have traditionally attempted to assess…
Descriptors: Testing Accommodations, Testing Programs, Educational Assessment, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Moss, Pamela A. – Educational Researcher, 2007
In response to Lissitz and Samuelsen (2007), the author reconstructs the historical arguments for the more comprehensive unitary concept of validity and the principles of scientific inquiry underlying it. Her response is organized in terms of four questions: (a) How did validity in educational measurement come to be conceptualized as unitary, and…
Descriptors: Evaluators, Construct Validity, Test Validity, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Camara, Wayne – Journal of Applied Testing Technology, 2009
The five papers in this special issue of the "Journal of Applied Testing Technology" address fundamental issues of validity when tests are modified or accommodations are provided to English Language Learners (ELL) or students with disabilities. Three papers employed differential item functioning (DIF) and factor analysis and found the…
Descriptors: Second Language Learning, Factor Analysis, English (Second Language), Cognitive Ability
Previous Page | Next Page ยป
Pages: 1  |  2  |  3