NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 16 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
John R. Donoghue; Carol Eckerly – Applied Measurement in Education, 2024
Trend scoring constructed response items (i.e. rescoring Time A responses at Time B) gives rise to two-way data that follow a product multinomial distribution rather than the multinomial distribution that is usually assumed. Recent work has shown that the difference in sampling model can have profound negative effects on statistics usually used to…
Descriptors: Scoring, Error of Measurement, Reliability, Scoring Rubrics
Shear, Benjamin R.; Reardon, Sean F. – Stanford Center for Education Policy Analysis, 2019
This paper describes a method for pooling grouped, ordered-categorical data across multiple waves to improve small-sample heteroskedastic ordered probit (HETOP) estimates of latent distributional parameters. We illustrate the method with aggregate proficiency data reporting the number of students in schools or districts scoring in each of a small…
Descriptors: Computation, Scores, Statistical Distributions, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Burns, Matthew K.; Taylor, Crystal N.; Warmbold-Brann, Kristy L.; Preast, June L.; Hosp, John L.; Ford, Jeremy W. – Psychology in the Schools, 2017
Intervention researchers often use curriculum-based measurement of reading fluency (CBM-R) with a brief experimental analysis (BEA) to identify an effective intervention for individual students. The current study synthesized data from 22 studies that used CBM-R data within a BEA by computing the standard error of measure (SEM) for the median data…
Descriptors: Error of Measurement, Decision Making, Reading Fluency, Curriculum Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Lin, Chih-Kai – Language Testing, 2017
Sparse-rated data are common in operational performance-based language tests, as an inevitable result of assigning examinee responses to a fraction of available raters. The current study investigates the precision of two generalizability-theory methods (i.e., the rating method and the subdividing method) specifically designed to accommodate the…
Descriptors: Data Analysis, Language Tests, Generalizability Theory, Accuracy
Custer, Michael – Online Submission, 2015
This study examines the relationship between sample size and item parameter estimation precision when utilizing the one-parameter model. Item parameter estimates are examined relative to "true" values by evaluating the decline in root mean squared deviation (RMSD) and the number of outliers as sample size increases. This occurs across…
Descriptors: Sample Size, Item Response Theory, Computation, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Keller-Margulis, Milena A.; Mercer, Sterett H.; Thomas, Erin L. – School Psychology Quarterly, 2016
The purpose of this study was to examine the reliability of written expression curriculum-based measurement (WE-CBM) in the context of universal screening from a generalizability theory framework. Students in second through fifth grade (n = 145) participated in the study. The sample included 54% female students, 49% White students, 23% African…
Descriptors: Generalizability Theory, Reliability, Written Language, Curriculum Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Reardon, Sean F.; Ho, Andrew D. – Journal of Educational and Behavioral Statistics, 2015
In an earlier paper, we presented methods for estimating achievement gaps when test scores are coarsened into a small number of ordered categories, preventing fine-grained distinctions between individual scores. We demonstrated that gaps can nonetheless be estimated with minimal bias across a broad range of simulated and real coarsened data…
Descriptors: Achievement Gap, Performance Factors, Educational Practices, Scores
Reardon, Sean F.; Ho, Andrew D. – Grantee Submission, 2015
Ho and Reardon (2012) present methods for estimating achievement gaps when test scores are coarsened into a small number of ordered categories, preventing fine-grained distinctions between individual scores. They demonstrate that gaps can nonetheless be estimated with minimal bias across a broad range of simulated and real coarsened data…
Descriptors: Achievement Gap, Performance Factors, Educational Practices, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Browne, Dillon T.; Leckie, George; Prime, Heather; Perlman, Michal; Jenkins, Jennifer M. – Developmental Psychology, 2016
The present study sought to investigate the family, individual, and dyad-specific contributions to observed cognitive sensitivity during family interactions. Moreover, the influence of cumulative risk on sensitivity at the aforementioned levels of the family was examined. Mothers and 2 children per family were observed interacting in a round robin…
Descriptors: Family Relationship, Family (Sociological Unit), Sibling Relationship, Siblings
Peer reviewed Peer reviewed
Direct linkDirect link
Burt, Keith B.; Obradovic, Jelena – Developmental Review, 2013
The purpose of this paper is to review major statistical and psychometric issues impacting the study of psychophysiological reactivity and discuss their implications for applied developmental researchers. We first cover traditional approaches such as the observed difference score (DS) and the observed residual score (RS), including a review of…
Descriptors: Measurement Techniques, Psychometrics, Data Analysis, Researchers
Moses, Tim; Liu, Jinghua – Educational Testing Service, 2011
In equating research and practice, equating functions that are smooth are typically assumed to be more accurate than equating functions with irregularities. This assumption presumes that population test score distributions are relatively smooth. In this study, two examples were used to reconsider common beliefs about smoothing and equating. The…
Descriptors: Equated Scores, Data Analysis, Scores, Methods
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Tourangeau, Karen; Nord, Christine; Lê, Thanh; Wallner-Allen, Kathleen; Vaden-Kiernan, Nancy; Blaker, Lisa; Najarian, Michelle – National Center for Education Statistics, 2018
This manual provides guidance and documentation for users of the longitudinal kindergarten-fourth grade (K-4) public-use data file of the Early Childhood Longitudinal Study, Kindergarten Class of 2010-11 (ECLS-K:2011), which includes the first release of the public version of the third-grade data. This manual mainly provides information specific…
Descriptors: Longitudinal Studies, Children, Surveys, Kindergarten
Peer reviewed Peer reviewed
Direct linkDirect link
Zimmerman, Donald W. – Educational and Psychological Measurement, 2007
Properties of the Spearman correction for attenuation were investigated using Monte Carlo methods, under conditions where correlations between error scores exist as a population parameter and also where correlated errors arise by chance in random sampling. Equations allowing for all possible dependence among true and error scores on two tests at…
Descriptors: Monte Carlo Methods, Correlation, Sampling, Data Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Lu, Irene R. R.; Thomas, D. Roland – Structural Equation Modeling: A Multidisciplinary Journal, 2008
This article considers models involving a single structural equation with latent explanatory and/or latent dependent variables where discrete items are used to measure the latent variables. Our primary focus is the use of scores as proxies for the latent variables and carrying out ordinary least squares (OLS) regression on such scores to estimate…
Descriptors: Least Squares Statistics, Computation, Item Response Theory, Structural Equation Models
Peer reviewed Peer reviewed
Blair, R. Clifford; And Others – Educational and Psychological Measurement, 1983
Sampling experiments were used to assess the Type I error rates of the t test in situations where classes were randomly assigned to groups but analyses were carried out on individual student scores. Even small amounts of between-class variation caused large inflations in the Type I error rate of the t test. (Author/BW)
Descriptors: Academic Achievement, Data Analysis, Elementary Secondary Education, Error of Measurement
Previous Page | Next Page »
Pages: 1  |  2