NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 67 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Cassiday, Kristina R.; Cho, Youngmi; Harring, Jeffrey R. – Educational and Psychological Measurement, 2021
Simulation studies involving mixture models inevitably aggregate parameter estimates and other output across numerous replications. A primary issue that arises in these methodological investigations is label switching. The current study compares several label switching corrections that are commonly used when dealing with mixture models. A growth…
Descriptors: Probability, Models, Simulation, Mathematics
Daniel McNeish; Jeffrey R. Harring; Daniel J. Bauer – Grantee Submission, 2022
Growth mixture models (GMMs) are a popular method to identify latent classes of growth trajectories. One shortcoming of GMMs is nonconvergence, which often leads researchers to apply covariance equality constraints to simplify estimation, though this may be a dubious assumption. Alternative model specifications have been proposed to reduce…
Descriptors: Growth Models, Classification, Accuracy, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Bashir, Rabia; Dunn, Adam G.; Surian, Didi – Research Synthesis Methods, 2021
Few data-driven approaches are available to estimate the risk of conclusion change in systematic review updates. We developed a rule-based approach to automatically extract information from reviews and updates to be used as features for modelling conclusion change risk. Rules were developed to extract relevant information from published Cochrane…
Descriptors: Literature Reviews, Data, Automation, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Alahmadi, Sarah; Jones, Andrew T.; Barry, Carol L.; Ibáñez, Beatriz – Applied Measurement in Education, 2023
Rasch common-item equating is often used in high-stakes testing to maintain equivalent passing standards across test administrations. If unaddressed, item parameter drift poses a major threat to the accuracy of Rasch common-item equating. We compared the performance of well-established and newly developed drift detection methods in small and large…
Descriptors: Equated Scores, Item Response Theory, Sample Size, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Nájera, Pablo; Sorrel, Miguel A.; Abad, Francisco José – Educational and Psychological Measurement, 2019
Cognitive diagnosis models (CDMs) are latent class multidimensional statistical models that help classify people accurately by using a set of discrete latent variables, commonly referred to as attributes. These models require a Q-matrix that indicates the attributes involved in each item. A potential problem is that the Q-matrix construction…
Descriptors: Matrices, Statistical Analysis, Models, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Klingbeil, David A.; Van Norman, Ethan R.; Nelson, Peter M. – Assessment for Effective Intervention, 2021
This direct replication study compared the use of dichotomized likelihood ratios and interval likelihood ratios, derived using a prior sample of students, for predicting math risk in middle school. Data from the prior year state test and the Measures of Academic Progress were analyzed to evaluate differences in the efficiency and diagnostic…
Descriptors: Achievement Tests, Grade 6, Grade 7, At Risk Students
Peer reviewed Peer reviewed
Direct linkDirect link
No, Unkyung; Hong, Sehee – Educational and Psychological Measurement, 2018
The purpose of the present study is to compare performances of mixture modeling approaches (i.e., one-step approach, three-step maximum-likelihood approach, three-step BCH approach, and LTB approach) based on diverse sample size conditions. To carry out this research, two simulation studies were conducted with two different models, a latent class…
Descriptors: Sample Size, Classification, Comparative Analysis, Statistical Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sünbül, Seçil Ömür – International Journal of Evaluation and Research in Education, 2018
In this study, it was aimed to investigate the impact of different missing data handling methods on DINA model parameter estimation and classification accuracy. In the study, simulated data were used and the data were generated by manipulating the number of items and sample size. In the generated data, two different missing data mechanisms…
Descriptors: Data, Test Items, Sample Size, Statistical Analysis
McNeish, Daniel; Harring, Jeffrey – Grantee Submission, 2019
Growth mixture models (GMMs) are prevalent for modeling unknown population heterogeneity via distinct latent classes. However, GMMs are riddled with convergence issues, often requiring researchers to atheoretically alter the model with cross-class constraints to obtain convergence. We discuss how within-class random effects in GMMs exacerbate…
Descriptors: Structural Equation Models, Classification, Computation, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Conger, Anthony J. – Educational and Psychological Measurement, 2017
Drawing parallels to classical test theory, this article clarifies the difference between rater accuracy and reliability and demonstrates how category marginal frequencies affect rater agreement and Cohen's kappa. Category assignment paradigms are developed: comparing raters to a standard (index) versus comparing two raters to one another…
Descriptors: Interrater Reliability, Evaluators, Accuracy, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Cousineau, Denis; Laurencelle, Louis – Educational and Psychological Measurement, 2017
Assessing global interrater agreement is difficult as most published indices are affected by the presence of mixtures of agreements and disagreements. A previously proposed method was shown to be specifically sensitive to global agreement, excluding mixtures, but also negatively biased. Here, we propose two alternatives in an attempt to find what…
Descriptors: Interrater Reliability, Evaluation Methods, Statistical Bias, Accuracy
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bashkov, Bozhidar M.; Clauser, Jerome C. – Practical Assessment, Research & Evaluation, 2019
Successful testing programs rely on high-quality test items to produce reliable scores and defensible exams. However, determining what statistical screening criteria are most appropriate to support these goals can be daunting. This study describes and demonstrates cost-benefit analysis as an empirical approach to determining appropriate screening…
Descriptors: Test Items, Test Reliability, Evaluation Criteria, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Miciak, Jeremy; Taylor, W. Pat; Stuebing, Karla K.; Fletcher, Jack M. – Journal of Psychoeducational Assessment, 2018
We investigated the classification accuracy of learning disability (LD) identification methods premised on the identification of an intraindividual pattern of processing strengths and weaknesses (PSW) method using multiple indicators for all latent constructs. Known LD status was derived from latent scores; values at the observed level identified…
Descriptors: Accuracy, Learning Disabilities, Classification, Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Vogel, Tobias; Carr, Evan W.; Davis, Tyler; Winkielman, Piotr – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2018
Stimuli that capture the central tendency of presented exemplars are often preferred--a phenomenon also known as the classic beauty-in-averageness effect. However, recent studies have shown that this effect can reverse under certain conditions. We propose that a key variable for such ugliness-in-averageness effects is the category structure of the…
Descriptors: Interpersonal Attraction, Preferences, Stimuli, Experiments
Peer reviewed Peer reviewed
Direct linkDirect link
Mahar, Matthew T.; Welk, Gregory J.; Rowe, David A. – Measurement in Physical Education and Exercise Science, 2018
Purpose: To develop models to estimate aerobic fitness (VO[subscript 2]max) from PACER performance in 10- to 18-year-old youth, with and without body mass index (BMI) as a predictor. Method: Youth (N = 280) completed the PACER and a maximal treadmill test to assess VO[subscript 2]max. Validation and cross-validation groups were randomly formed to…
Descriptors: Exercise, Physical Fitness, Preadolescents, Adolescents
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5