Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 8 |
Descriptor
Sampling | 55 |
Mathematical Models | 12 |
Statistical Analysis | 11 |
Research Design | 10 |
Tables (Data) | 10 |
Equated Scores | 9 |
Error of Measurement | 8 |
Measurement | 7 |
Test Construction | 7 |
Test Items | 7 |
Test Reliability | 7 |
More ▼ |
Source
Journal of Educational… | 55 |
Author
Publication Type
Journal Articles | 36 |
Reports - Research | 19 |
Reports - Evaluative | 12 |
Reports - Descriptive | 4 |
Speeches/Meeting Papers | 3 |
Information Analyses | 2 |
Book/Product Reviews | 1 |
Education Level
Audience
Researchers | 1 |
Location
United Kingdom (Scotland) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Combs, Adam – Journal of Educational Measurement, 2023
A common method of checking person-fit in Bayesian item response theory (IRT) is the posterior-predictive (PP) method. In recent years, more powerful approaches have been proposed that are based on resampling methods using the popular L*[subscript z] statistic. There has also been proposed a new Bayesian model checking method based on pivotal…
Descriptors: Bayesian Statistics, Goodness of Fit, Evaluation Methods, Monte Carlo Methods
Castellano, Katherine E.; McCaffrey, Daniel F.; Lockwood, J. R. – Journal of Educational Measurement, 2023
The simple average of student growth scores is often used in accountability systems, but it can be problematic for decision making. When computed using a small/moderate number of students, it can be sensitive to the sample, resulting in inaccurate representations of growth of the students, low year-to-year stability, and inequities for…
Descriptors: Academic Achievement, Accountability, Decision Making, Computation
Kim, Hyung Jin; Brennan, Robert L.; Lee, Won-Chan – Journal of Educational Measurement, 2020
In equating, smoothing techniques are frequently used to diminish sampling error. There are typically two types of smoothing: presmoothing and postsmoothing. For polynomial log-linear presmoothing, an optimum smoothing degree can be determined statistically based on the Akaike information criterion or Chi-square difference criterion. For…
Descriptors: Equated Scores, Sampling, Error of Measurement, Statistical Analysis
Sinharay, Sandip – Journal of Educational Measurement, 2016
De la Torre and Deng suggested a resampling-based approach for person-fit assessment (PFA). The approach involves the use of the [math equation unavailable] statistic, a corrected expected a posteriori estimate of the examinee ability, and the Monte Carlo (MC) resampling method. The Type I error rate of the approach was closer to the nominal level…
Descriptors: Sampling, Research Methodology, Error Patterns, Monte Carlo Methods
Albano, Anthony D. – Journal of Educational Measurement, 2015
Research on equating with small samples has shown that methods with stronger assumptions and fewer statistical estimates can lead to decreased error in the estimated equating function. This article introduces a new approach to linear observed-score equating, one which provides flexible control over how form difficulty is assumed versus estimated…
Descriptors: Equated Scores, Sample Size, Sampling, Statistical Inference
Li, Deping; Jiang, Yanlin; von Davier, Alina A. – Journal of Educational Measurement, 2012
This study investigates a sequence of item response theory (IRT) true score equatings based on various scale transformation approaches and evaluates equating accuracy and consistency over time. The results show that the biases and sample variances for the IRT true score equating (both direct and indirect) are quite small (except for the mean/sigma…
Descriptors: True Scores, Equated Scores, Item Response Theory, Accuracy
Baldwin, Peter – Journal of Educational Measurement, 2011
Growing interest in fully Bayesian item response models begs the question: To what extent can model parameter posterior draws enhance existing practices? One practice that has traditionally relied on model parameter point estimates but may be improved by using posterior draws is the development of a common metric for two independently calibrated…
Descriptors: Item Response Theory, Bayesian Statistics, Computation, Sampling
Kane, Michael T. – Journal of Educational Measurement, 2013
To validate an interpretation or use of test scores is to evaluate the plausibility of the claims based on the scores. An argument-based approach to validation suggests that the claims based on the test scores be outlined as an argument that specifies the inferences and supporting assumptions needed to get from test responses to score-based…
Descriptors: Test Interpretation, Validity, Scores, Test Use

Sirotnik, Kenneth; Wellington, Roger – Journal of Educational Measurement, 1977
A single conceptual and theoretical framework for sampling any configuration of data from one or more population matrices is presented, integrating past designs and discussing implications for more general designs. The theory is based upon a generalization of the generalized symmetric mean approach for single matrix samples. (Author/CTM)
Descriptors: Analysis of Variance, Data Analysis, Item Sampling, Mathematical Models

Smith, Philip L. – Journal of Educational Measurement, 1981
This study explores a strategy for improving the stability of variance component estimates when only small samples are available, using a series of small, less complex generalizability (G) study designs as a surrogate for a single large design. (Author/BW)
Descriptors: Models, Reliability, Research Design, Sampling

Shavelson, Richard J.; Ruiz-Primo, Maria Araceli; Wiley, Edward W. – Journal of Educational Measurement, 1999
Reports a reanalysis of data collected in a person x task x occasion rater or method G-study design (M. Ruiz-Primo and others, 1993), and brings this reanalysis to bear on the interpretation of task-sampling variability and the convergence of different performance-assessment methods. (SLD)
Descriptors: Performance Based Assessment, Sampling, Sciences

Scrams, David J.; McLeod, Lori D. – Journal of Educational Measurement, 2000
Presents an approach to graphical differential item functioning (DIF) based on a sampling-theory approach to expected response functions. Applied the approach to a set of pretest items and compared results to traditional Mantel Haenszel DIF statistics. Discusses implications of the method as a complement to the approach of P. Pashley (1992). (SLD)
Descriptors: Item Bias, Pretests Posttests, Sampling

Kane, Michael – Journal of Educational Measurement, 2002
Reviews the criticisms of sampling assumptions in generalizability theory (and in reliability theory) and examines the feasibility of using representative sampling, stratification, homogeneity assumptions, and replications to address these criticisms. Suggests some general outlines for the conduct of generalizability theory studies. (SLD)
Descriptors: Generalizability Theory, Reliability, Research Methodology, Sampling

Sabers, Darrell L.; Klausmeier, Richard D. – Journal of Educational Measurement, 1971
Descriptors: Measurement Techniques, Raw Scores, Sampling, Statistical Analysis

Lee, Guemin; Fitzpatrick, Anne R. – Journal of Educational Measurement, 2003
Studied three procedures for estimating the standard errors of school passing rates using a generalizability theory model and considered the effects of student sample size. Results show that procedures differ in terms of assumptions about the populations from which students were sampled, and student sample size was found to have a large effect on…
Descriptors: Error of Measurement, Estimation (Mathematics), Generalizability Theory, Sampling