NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers2
Location
Texas1
Laws, Policies, & Programs
Assessments and Surveys
State of Texas Assessments of…1
What Works Clearinghouse Rating
Showing 1 to 15 of 41 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Walter M. Stroup; Anthony Petrosino; Corey Brady; Karen Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Tests of statistical significance often play a decisive role in establishing the empirical warrant of evidence-based research in education. The results from pattern-based assessment items, as introduced in this paper, are categorical and multimodal and do not immediately support the use of measures of central tendency as typically related to…
Descriptors: Statistical Significance, Comparative Analysis, Research Methodology, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Slavin, Robert E.; Cheung, Alan C. K. – Journal of Education for Students Placed at Risk, 2017
Large-scale randomized studies provide the best means of evaluating practical, replicable approaches to improving educational outcomes. This article discusses the advantages, problems, and pitfalls of these evaluations, focusing on alternative methods of randomization, recruitment, ensuring high-quality implementation, dealing with attrition, and…
Descriptors: Randomized Controlled Trials, Evaluation Methods, Recruitment, Attrition (Research Studies)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Steiner, Peter M.; Wong, Vivian – Society for Research on Educational Effectiveness, 2016
Despite recent emphasis on the use of randomized control trials (RCTs) for evaluating education interventions, in most areas of education research, observational methods remain the dominant approach for assessing program effects. Over the last three decades, the within-study comparison (WSC) design has emerged as a method for evaluating the…
Descriptors: Randomized Controlled Trials, Comparative Analysis, Research Design, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Gelman, Andrew; Hill, Jennifer; Yajima, Masanao – Journal of Research on Educational Effectiveness, 2012
Applied researchers often find themselves making statistical inferences in settings that would seem to require multiple comparisons adjustments. We challenge the Type I error paradigm that underlies these corrections. Moreover we posit that the problem of multiple comparisons can disappear entirely when viewed from a hierarchical Bayesian…
Descriptors: Intervals, Comparative Analysis, Inferences, Error Patterns
Newman, Denis; Jaciw, Andrew P. – Empirical Education Inc., 2012
The motivation for this paper is the authors' recent work on several randomized control trials in which they found the primary result, which averaged across subgroups or sites, to be moderated by demographic or site characteristics. They are led to examine a distinction that the Institute of Education Sciences (IES) makes between "confirmatory"…
Descriptors: Educational Research, Research Methodology, Research Design, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Christie, Christina A.; Fleischer, Dreolin Nesbitt – American Journal of Evaluation, 2010
To describe the recent practice of evaluation, specifically method and design choices, the authors performed a content analysis on 117 evaluation studies published in eight North American evaluation-focused journals for a 3-year period (2004-2006). The authors chose this time span because it follows the scientifically based research (SBR)…
Descriptors: Content Analysis, Periodicals, Qualitative Research, Research Design
Peer reviewed Peer reviewed
Direct linkDirect link
Breton, Theodore R. – Economics of Education Review, 2011
This paper challenges Hanushek and Woessmann's (2008) contention that the quality and not the quantity of schooling determines a nation's rate of economic growth. I first show that their statistical analysis is flawed. I then show that when a nation's average test scores and average schooling attainment are included in a national income model,…
Descriptors: Economic Progress, Income, Statistical Significance, Educational Quality
Peer reviewed Peer reviewed
Direct linkDirect link
Maraun, Michael; Gabriel, Stephanie – Psychological Methods, 2010
In his article, "An Alternative to Null-Hypothesis Significance Tests," Killeen (2005) urged the discipline to abandon the practice of "p[subscript obs]"-based null hypothesis testing and to quantify the signal-to-noise characteristics of experimental outcomes with replication probabilities. He described the coefficient that he…
Descriptors: Hypothesis Testing, Statistical Inference, Probability, Statistical Significance
Peer reviewed Peer reviewed
Direct linkDirect link
Mutlu, Akmer; Krosschell, Kristin; Spira, Deborah Gaebler – Developmental Medicine & Child Neurology, 2009
OKAim: The aim of this systematic review was to examine the literature on the effects of partial body-weight support treadmill training (PBWSTT) in children with cerebral palsy (CP) on functional outcomes and attainment of ambulation. Method: We searched the relevant literature from 1950 to July 2007. We found eight studies on the use of PWSBTT on…
Descriptors: Cerebral Palsy, Statistical Significance, Classification, Psychomotor Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Serlin, Ronald C. – Psychological Methods, 2010
The sense that replicability is an important aspect of empirical science led Killeen (2005a) to define "p[subscript rep]," the probability that a replication will result in an outcome in the same direction as that found in a current experiment. Since then, several authors have praised and criticized 'p[subscript rep]," culminating…
Descriptors: Epistemology, Effect Size, Replication (Evaluation), Measurement Techniques
Peer reviewed Peer reviewed
Direct linkDirect link
Cumming, Geoff – Psychological Methods, 2010
This comment offers three descriptions of "p[subscript rep]" that start with a frequentist account of confidence intervals, draw on R. A. Fisher's fiducial argument, and do not make Bayesian assumptions. Links are described among "p[subscript rep]," "p" values, and the probability a confidence interval will capture…
Descriptors: Replication (Evaluation), Measurement Techniques, Research Methodology, Validity
Peer reviewed Peer reviewed
PDF on ERIC Download full text
What Works Clearinghouse, 2011
With its critical assessments of scientific evidence on the effectiveness of education programs, policies, and practices (referred to as "interventions"), and a range of products summarizing this evidence, the What Works Clearinghouse (WWC) is an important part of the Institute of Education Sciences' strategy to use rigorous and relevant…
Descriptors: Standards, Access to Information, Information Management, Guides
Peer reviewed Peer reviewed
Direct linkDirect link
Thompson, Bruce – Psychology in the Schools, 2007
The present article provides a primer on (a) effect sizes, (b) confidence intervals, and (c) confidence intervals for effect sizes. Additionally, various admonitions for reformed statistical practice are presented. For example, a very important implication of the realization that there are dozens of effect size statistics is that "authors must…
Descriptors: Intervals, Effect Size, Statistical Analysis, Statistical Significance
Peer reviewed Peer reviewed
Direct linkDirect link
Erceg-Hurn, David M.; Mirosevich, Vikki M. – American Psychologist, 2008
Classic parametric statistical significance tests, such as analysis of variance and least squares regression, are widely used by researchers in many disciplines, including psychology. For classic parametric tests to produce accurate results, the assumptions underlying them (e.g., normality and homoscedasticity) must be satisfied. These assumptions…
Descriptors: Statistical Significance, Least Squares Statistics, Effect Size, Statistical Studies
Peer reviewed Peer reviewed
Cahan, Sorel – Educational Researcher, 2000
Shows why the two-step approach proposed by D. Robinson and J. Levine (1997) is inappropriate for the evaluation of empirical results and reiterates the preferred approach of increased sample size and the computation of confidence intervals. (SLD)
Descriptors: Effect Size, Evaluation Methods, Research Methodology, Sample Size
Previous Page | Next Page ยป
Pages: 1  |  2  |  3