NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20252
Since 2022 (last 5 years)7
Since 2017 (last 10 years)17
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 17 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Ricardo Colaço; Pedro Freitas; Luis Catela Nunes; Ana Balcão Reis – Large-scale Assessments in Education, 2025
We analyse the PISA-reported convergence in the performance of private and public schools in Portugal. When PISA sampling weights are used, the number of students enrolled in those types of schools and specific grades/tracks of study differs significantly from official population figures. To account for those differences, we apply a…
Descriptors: Foreign Countries, Achievement Tests, International Assessment, Secondary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Paul A. Jewsbury; Matthew S. Johnson – Large-scale Assessments in Education, 2025
The standard methodology for many large-scale assessments in education involves regressing latent variables on numerous contextual variables to estimate proficiency distributions. To reduce the number of contextual variables used in the regression and improve estimation, we propose and evaluate principal component analysis on the covariance matrix…
Descriptors: Factor Analysis, Matrices, Regression (Statistics), Educational Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ayse Bilicioglu Gunes; Bayram Bicak – International Journal of Assessment Tools in Education, 2023
The main purpose of this study is to examine the Type I error and statistical power ratios of Differential Item Functioning (DIF) techniques based on different theories under different conditions. For this purpose, a simulation study was conducted by using Mantel-Haenszel (MH), Logistic Regression (LR), Lord's [chi-squared], and Raju's Areas…
Descriptors: Test Items, Item Response Theory, Error of Measurement, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Hyung Jin; Lee, Won-Chan – Journal of Educational Measurement, 2022
Orlando and Thissen (2000) introduced the "S - X[superscript 2]" item-fit index for testing goodness-of-fit with dichotomous item response theory (IRT) models. This study considers and evaluates an alternative approach for computing "S - X[superscript 2]" values and other factors associated with collapsing tables of observed…
Descriptors: Goodness of Fit, Test Items, Item Response Theory, Computation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Basman, Munevver – International Journal of Assessment Tools in Education, 2023
To ensure the validity of the tests is to check that all items have similar results across different groups of individuals. However, differential item functioning (DIF) occurs when the results of individuals with equal ability levels from different groups differ from each other on the same test item. Based on Item Response Theory and Classic Test…
Descriptors: Test Bias, Test Items, Test Validity, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Wolkowitz, Amanda A.; Wright, Keith D. – Journal of Educational Measurement, 2019
This article explores the amount of equating error at a passing score when equating scores from exams with small samples sizes. This article focuses on equating using classical test theory methods of Tucker linear, Levine linear, frequency estimation, and chained equipercentile equating. Both simulation and real data studies were used in the…
Descriptors: Error Patterns, Sample Size, Test Theory, Test Bias
Abdalla, Widad – ProQuest LLC, 2019
Trend scoring is often used in large-scale assessments to monitor for rater drift when the same constructed response items are administered in multiple test administrations. In trend scoring, a set of responses from Time "A" are rescored by raters at Time "B." The purpose of this study is to examine the ability of…
Descriptors: Scoring, Interrater Reliability, Test Items, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Rickard, Timothy C.; Pan, Steven C.; Gupta, Mohan W. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2022
We explored the possibility of publication bias in the sleep and explicit motor sequence learning literature by applying precision effect test (PET) and precision effect test with standard errors (PEESE) weighted regression analyses to the 88 effect sizes from a recent comprehensive literature review (Pan & Rickard, 2015). Basic PET analysis…
Descriptors: Publications, Bias, Sleep, Psychomotor Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Berrío, Ángela I.; Herrera, Aura N.; Gómez-Benito, Juana – Journal of Experimental Education, 2019
This study examined the effect of sample size ratio and model misfit on the Type I error rates and power of the Difficulty Parameter Differences procedure using Winsteps. A unidimensional 30-item test with responses from 130,000 examinees was simulated and four independent variables were manipulated: sample size ratio (20/100/250/500/1000); model…
Descriptors: Sample Size, Test Bias, Goodness of Fit, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Park, Sunyoung; Beretvas, S. Natasha – Journal of Experimental Education, 2019
The log-odds ratio (ln[OR]) is commonly used to quantify treatments' effects on dichotomous outcomes and then pooled across studies using inverse-variance (1/v) weights. Calculation of the ln[OR]'s variance requires four cell frequencies for two groups crossed with values for dichotomous outcomes. While primary studies report the total sample size…
Descriptors: Sample Size, Meta Analysis, Statistical Analysis, Efficiency
Peer reviewed Peer reviewed
Direct linkDirect link
Hood, Audrey V. B.; Whillock, Summer R.; Meade, Michelle L.; Hutchison, Keith A. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2023
Collaborative inhibition (reduced recall in collaborative vs. nominal groups) is a robust phenomenon. However, it is possible that not everyone is as susceptible to collaborative inhibition, such as those higher in working memory capacity (WMC). In the current study, we examined the relationship between WMC and collaborative inhibition.…
Descriptors: Short Term Memory, Recall (Psychology), Task Analysis, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Zhang, Zhonghua; Zhao, Mingren – Journal of Educational Measurement, 2019
The present study evaluated the multiple imputation method, a procedure that is similar to the one suggested by Li and Lissitz (2004), and compared the performance of this method with that of the bootstrap method and the delta method in obtaining the standard errors for the estimates of the parameter scale transformation coefficients in item…
Descriptors: Item Response Theory, Error Patterns, Item Analysis, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Doleman, Brett; Freeman, Suzanne C.; Lund, Jonathan N.; Williams, John P.; Sutton, Alex J. – Research Synthesis Methods, 2020
This study aimed to determine for continuous outcomes dependent on baseline risk, whether funnel plot asymmetry may be due to statistical artefact rather than publication bias and evaluate a novel test to resolve this. Firstly, we conducted assessment for publication bias in nine meta-analyses of postoperative analgesics (344 trials with 25 348…
Descriptors: Outcomes of Treatment, Risk, Publications, Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Cormier, Damien C.; Van Norman, Ethan R.; Cheong, Clarissa; Kennedy, Kathleen E.; Bulut, Okan; Mrazik, Martin – Canadian Journal of School Psychology, 2019
This study aims to systematically evaluate the scoring errors made by psychologists in training, in the hopes of providing strong, empirically based guidelines to training programs. Survival analysis was used to determine the number of attempts required for graduate students to achieve proficiency in scoring standardized record forms from the…
Descriptors: Cognitive Measurement, Assessment Literacy, Scoring, Psychologists
Cai, Zhiqiang; Siebert-Evenstone, Amanda; Eagan, Brendan; Shaffer, David Williamson; Hu, Xiangen; Graesser, Arthur C. – Grantee Submission, 2019
Coding is a process of assigning meaning to a given piece of evidence. Evidence may be found in a variety of data types, including documents, research interviews, posts from social media, conversations from learning platforms, or any source of data that may provide insights for the questions under qualitative study. In this study, we focus on text…
Descriptors: Semantics, Computational Linguistics, Evidence, Coding
Previous Page | Next Page »
Pages: 1  |  2