NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)6
Since 2006 (last 20 years)10
Audience
Researchers1
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Kyung Yong; Lee, Won-Chan – Journal of Educational Measurement, 2018
Reporting confidence intervals with test scores helps test users make important decisions about examinees by providing information about the precision of test scores. Although a variety of estimation procedures based on the binomial error model are available for computing intervals for test scores, these procedures assume that items are randomly…
Descriptors: Weighted Scores, Error of Measurement, Test Use, Decision Making
Peer reviewed Peer reviewed
Direct linkDirect link
Hopster-den Otter, Dorien; Muilenburg, Selia N.; Wools, Saskia; Veldkamp, Bernard P.; Eggen, Theo J. H. M. – Assessment in Education: Principles, Policy & Practice, 2019
This study investigated (1) the extent to which presentations of measurement error in score reports influence teachers' decisions and (2) teachers' preferences in relation to these presentations. Three presentation formats of measurement error (blur, colour value and error bar) were compared to a presentation format that omitted measurement error.…
Descriptors: Error of Measurement, Scores, Decision Making, Teacher Attitudes
Peer reviewed Peer reviewed
PDF on ERIC Download full text
McCaffrey, Daniel F.; Oliveri, Maria Elena; Holtzman, Steven – ETS Research Report Series, 2018
Scores from noncognitive measures are increasingly valued for their utility in helping to inform postsecondary admissions decisions. However, their use has presented challenges because of faking, response biases, or subjectivity, which standardized third-party evaluations (TPEs) can help minimize. Analysts and researchers using TPEs, however, need…
Descriptors: Generalizability Theory, Scores, College Admission, Admission Criteria
Peer reviewed Peer reviewed
Direct linkDirect link
Burns, Matthew K.; Taylor, Crystal N.; Warmbold-Brann, Kristy L.; Preast, June L.; Hosp, John L.; Ford, Jeremy W. – Psychology in the Schools, 2017
Intervention researchers often use curriculum-based measurement of reading fluency (CBM-R) with a brief experimental analysis (BEA) to identify an effective intervention for individual students. The current study synthesized data from 22 studies that used CBM-R data within a BEA by computing the standard error of measure (SEM) for the median data…
Descriptors: Error of Measurement, Decision Making, Reading Fluency, Curriculum Based Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Van Norman, Ethan R.; Christ, Theodore J.; Newell, Kirsten W. – School Psychology Review, 2017
Research regarding the technical adequacy of growth estimates from curriculum-based measurement of reading progress monitoring data suggests that current decision-making frameworks are likely to yield inaccurate recommendations unless data are collected for extensive periods of time. Instances where data may not need to be collected for long…
Descriptors: Progress Monitoring, Curriculum Based Assessment, Goal Orientation, Decision Making
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sekercioglu, Güçlü – International Online Journal of Education and Teaching, 2018
An empirical evidence for independent samples of a population regarding measurement invariance implies that factor structure of a measurement tool is equal across these samples; in other words, it measures the intended psychological trait within the same structure. In this case, the evidence of construct validity would be strengthened within the…
Descriptors: Factor Analysis, Error of Measurement, Factor Structure, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Severo, Milton; Silva-Pereira, Fernanda; Ferreira, Maria Amelia – Anatomical Sciences Education, 2013
Several studies have shown that the standard error of measurement (SEM) can be used as an additional “safety net” to reduce the frequency of false-positive or false-negative student grading classifications. Practical examinations in clinical anatomy are often used as diagnostic tests to admit students to course final examinations. The aim of this…
Descriptors: Anatomy, Medical Education, Decision Making, Pass Fail Grading
Peer reviewed Peer reviewed
Direct linkDirect link
Kruyen, Peter M.; Emons, Wilco H. M.; Sijtsma, Klaas – International Journal of Testing, 2012
Personnel selection shows an enduring need for short stand-alone tests consisting of, say, 5 to 15 items. Despite their efficiency, short tests are more vulnerable to measurement error than longer test versions. Consequently, the question arises to what extent reducing test length deteriorates decision quality due to increased impact of…
Descriptors: Measurement, Personnel Selection, Decision Making, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Salbach, Nancy M.; Jaglal, Susan B.; Williams, Jack I. – Journal of Continuing Education in the Health Professions, 2013
Introduction: The reliability, minimal detectable change (MDC), and construct validity of the evidence-based practice confidence (EPIC) scale were evaluated among physical therapists (PTs) in clinical practice. Methods: A longitudinal mail survey was conducted. Internal consistency and test-retest reliability were estimated using Cronbach's alpha…
Descriptors: Evidence, Error of Measurement, Measures (Individuals), Physical Therapy
Peer reviewed Peer reviewed
Direct linkDirect link
Parker, Richard I.; Vannest, Kimberly J.; Davis, John L.; Clemens, Nathan H. – Journal of Special Education, 2012
Within a response to intervention model, educators increasingly use progress monitoring (PM) to support medium- to high-stakes decisions for individual students. For PM to serve these more demanding decisions requires more careful consideration of measurement error. That error should be calculated within a fixed linear regression model rather than…
Descriptors: Measurement, Computation, Response to Intervention, Regression (Statistics)
Thompson, Bruce; Crowley, Susan – 1994
Most training programs in education and psychology focus on classical test theory techniques for assessing score dependability. This paper discusses generalizability theory and explores its concepts using a small heuristic data set. Generalizability theory subsumes and extends classical test score theory. It is able to estimate the magnitude of…
Descriptors: Analysis of Variance, Cutting Scores, Decision Making, Error of Measurement
Kifer, Edward; Bramble, William – 1974
A latent trait model, the Rasch, was fitted to a criterion-referenced test. Approximately 90 percent of the items fit the model. Those items which fit the model were then calibrated. Based on the item calibration, individual ability estimates and the standard errors of those estimates were calculated. Using the ability estimates, it was possible,…
Descriptors: Academic Ability, Achievement Tests, Criterion Referenced Tests, Decision Making
Lord, Frederic M. – 1983
If a loss function is available specifying the social cost of an error of measurement in the score on a unidimensional test, an asymptotic method, based on item response theory, is developed for optimal test design for a specified target population of examinees. Since in the real world such loss functions are not available, it is more useful to…
Descriptors: Cutting Scores, Decision Making, Error of Measurement, Estimation (Mathematics)
Wilde, Elizabeth Ty; Hollister, Robinson – Institute for Research on Poverty, 2002
In this study we test the performance of some nonexperimental estimators of impacts applied to an educational intervention--reduction in class size--where achievement test scores were the outcome. We compare the nonexperimental estimates of the impacts to "true impact" estimates provided by a random-assignment design used to assess the…
Descriptors: Computation, Outcome Measures, Achievement Tests, Scores
Fruen, Mary – NCME Measurement in Education, 1978
There are both strengths and weaknesses of using standardized test scores as a criterion for admission to institutions of higher education. The relative importance of scores is dependent on the institution's degree of selectivity. In general, decision processes and admissions criteria are not well defined. Advantages of test scores include: use of…
Descriptors: Admission Criteria, College Admission, College Entrance Examinations, Competitive Selection