NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)1
Since 2006 (last 20 years)9
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 13 results Save | Export
Herman, Joan L.; La Torre, Deborah; Epstein, Scott; Wang, Jia – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2016
This report presents the results of expert panels' item-by-item analysis of the 2015 PISA Reading Literacy and Mathematics Literacy assessments and compares study findings on PISA's representation of deeper learning with that of other related studies. Results indicate that about 11% to 14% of PISA's total raw score value for reading and…
Descriptors: Achievement Tests, International Assessment, Foreign Countries, Secondary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Debelak, Rudolf; Arendasy, Martin – Educational and Psychological Measurement, 2012
A new approach to identify item clusters fitting the Rasch model is described and evaluated using simulated and real data. The proposed method is based on hierarchical cluster analysis and constructs clusters of items that show a good fit to the Rasch model. It thus gives an estimate of the number of independent scales satisfying the postulates of…
Descriptors: Test Items, Factor Analysis, Evaluation Methods, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Eun Sook; Yoon, Myeongsun; Lee, Taehun – Educational and Psychological Measurement, 2012
Multiple-indicators multiple-causes (MIMIC) modeling is often used to test a latent group mean difference while assuming the equivalence of factor loadings and intercepts over groups. However, this study demonstrated that MIMIC was insensitive to the presence of factor loading noninvariance, which implies that factor loading invariance should be…
Descriptors: Test Items, Simulation, Testing, Statistical Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Padilla, Jose Luis; Hidalgo, M. Dolores; Benitez, Isabel; Gomez-Benito, Juana – Psicologica: International Journal of Methodology and Experimental Psychology, 2012
The analysis of differential item functioning (DIF) examines whether item responses differ according to characteristics such as language and ethnicity, when people with matching ability levels respond differently to the items. This analysis can be performed by calculating various statistics, one of the most important being the Mantel-Haenszel,…
Descriptors: Foreign Countries, Test Bias, Computer Software, Computer Software Evaluation
Tian, Feng – ProQuest LLC, 2011
There has been a steady increase in the use of mixed-format tests, that is, tests consisting of both multiple-choice items and constructed-response items in both classroom and large-scale assessments. This calls for appropriate equating methods for such tests. As Item Response Theory (IRT) has rapidly become mainstream as the theoretical basis for…
Descriptors: Item Response Theory, Comparative Analysis, Equated Scores, Statistical Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Usener, Claus A.; Majchrzak, Tim A.; Kuchen, Herbert – Interactive Technology and Smart Education, 2012
Purpose: To overcome the high manual effort of assessments for teaching personnel, e-assessment systems are used to assess students using information systems (IS). The purpose of this paper is to propose an extension of EASy, a system for e-assessment of exercises that require higher-order cognitive skills. The latest module allows assessing…
Descriptors: Foreign Countries, Computer Software, Computer Software Evaluation, Computer Assisted Testing
ACT, Inc., 2013
This manual contains information about the American College Test (ACT) PlanĀ® program. The principal focus of this manual is to document the Plan program's technical adequacy in light of its intended purposes. This manual supersedes the 2011 edition. The content of this manual responds to requirements of the testing industry as established in the…
Descriptors: College Entrance Examinations, Formative Evaluation, Evaluation Research, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
van Ginkel, Joost R.; van der Ark, L. Andries – Applied Psychological Measurement, 2005
A well-known problem in the analysis of test and questionnaire data is that some item scores may be missing. Advanced methods for the imputation of missing data are available, such as multiple imputation under the multivariate normal model and imputation under the saturated logistic model (Schafer, 1997). Accompanying software was made available…
Descriptors: Syntax, Statistical Analysis, Test Items, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Cominole, Melissa; Wheeless, Sara; Dudley, Kristin; Franklin, Jeff; Wine, Jennifer – National Center for Education Statistics, 2007
The "2004/06 Beginning Postsecondary Students Longitudinal Study (BPS:04/06)" is sponsored by the U.S. Department of Education to respond to the need for a national, comprehensive database concerning issues students may face in enrollment, persistence, progress, and attainment in postsecondary education and in consequent early rates of…
Descriptors: Postsecondary Education, Stopouts, Research Methodology, Data Collection
Peer reviewed Peer reviewed
Direct linkDirect link
Monahan, Patrick O.; Ankenmann, Robert D. – Journal of Educational Measurement, 2005
Empirical studies demonstrated Type-I error (TIE) inflation (especially for highly discriminating easy items) of the Mantel-Haenszel chi-square test for differential item functioning (DIF), when data conformed to item response theory (IRT) models more complex than Rasch, and when IRT proficiency distributions differed only in means. However, no…
Descriptors: Sample Size, Item Response Theory, Test Items, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Hidalgo, M. Dolores; Lopez-Pina, Jose Antonio – Educational and Psychological Measurement, 2004
This article compares several procedures in their efficacy for detecting differential item functioning (DIF): logistic regression analysis, the Mantel-Haenszel (MH) procedure, and the modified Mantel-Haenszel procedure by Mazor, Clauser, and Hambleton. It also compares the effect size measures that these procedures provide. In this study,…
Descriptors: Effect Size, Test Bias, Regression (Statistics), Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Al-A'ali, Mansoor – Educational Technology & Society, 2007
Computer adaptive testing is the study of scoring tests and questions based on assumptions concerning the mathematical relationship between examinees' ability and the examinees' responses. Adaptive student tests, which are based on item response theory (IRT), have many advantages over conventional tests. We use the least square method, a…
Descriptors: Educational Testing, Higher Education, Elementary Secondary Education, Student Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wine, Jennifer S.; Heuer, Ruth E.; Wheeless, Sara C.; Francis, Talbric L.; Franklin, Jeff W.; Dudley, Kristin M. – National Center for Education Statistics, 2002
This report describes the methods and procedures used for the Beginning Postsecondary Students Longitudinal Study: 1996-2001 (BPS:1996/2001). These students, who started their postsecondary education during the 1995-96 academic year, were first interviewed in 1996 as part of the National Postsecondary Student Aid Study (NPSAS:96). A follow-up…
Descriptors: Longitudinal Studies, Postsecondary Education, Research Methodology, Interviews