NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 826 to 840 of 3,711 results Save | Export
Tan, Xuan; Xiang, Bihua; Dorans, Neil J.; Qu, Yanxuan – Educational Testing Service, 2010
The nature of the matching criterion (usually the total score) in the study of differential item functioning (DIF) has been shown to impact the accuracy of different DIF detection procedures. One of the topics related to the nature of the matching criterion is whether the studied item should be included. Although many studies exist that suggest…
Descriptors: Test Bias, Test Items, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip; Dorans, Neil J. – Journal of Educational and Behavioral Statistics, 2010
The Mantel-Haenszel (MH) procedure (Mantel and Haenszel) is a popular method for estimating and testing a common two-factor association parameter in a 2 x 2 x K table. Holland and Holland and Thayer described how to use the procedure to detect differential item functioning (DIF) for tests with dichotomously scored items. Wang, Bradlow, Wainer, and…
Descriptors: Test Bias, Statistical Analysis, Computation, Bayesian Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Fidalgo, Angel M.; Bartram, Dave – Applied Psychological Measurement, 2010
The main objective of this study was to establish the relative efficacy of the generalized Mantel-Haenszel test (GMH) and the Mantel test for detecting large numbers of differential item functioning (DIF) patterns. To this end this study considered a topic not dealt with in the literature to date: the possible differential effect of type of scores…
Descriptors: Test Bias, Statistics, Scoring, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Penfield, Randall D. – Applied Psychological Measurement, 2010
Crossing, or intersecting, differential item functioning (DIF) is a form of nonuniform DIF that exists when the sign of the between-group difference in expected item performance changes across the latent trait continuum. The presence of crossing DIF presents a problem for many statistics developed for evaluating DIF because positive and negative…
Descriptors: Test Bias, Test Items, Statistics, Test Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Shaw, Stuart; Imam, Helen – Language Assessment Quarterly, 2013
International assessments in a wide range of subjects are being prepared for and delivered through the medium of English in a variety of educational contexts. These assessments are taken by many candidates whose first language is not necessarily English. This raises important issues relating to assessment validity and fairness. This study…
Descriptors: English (Second Language), Test Validity, Test Bias, High Stakes Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Albano, Anthony D.; Rodriguez, Michael C. – Educational and Psychological Measurement, 2013
Although a substantial amount of research has been conducted on differential item functioning in testing, studies have focused on detecting differential item functioning rather than on explaining how or why it may occur. Some recent work has explored sources of differential functioning using explanatory and multilevel item response models. This…
Descriptors: Test Bias, Hierarchical Linear Modeling, Gender Differences, Educational Opportunities
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, W. Holmes; Hernández Finch, Maria E. – Educational and Psychological Measurement, 2013
The assessment of test data for the presence of differential item functioning (DIF) is a key component of instrument development and validation. Among the many methods that have been used successfully in such analyses is the mixture modeling approach. Using this approach to identify the presence of DIF has been touted as potentially superior for…
Descriptors: Learning Disabilities, Testing Accommodations, Test Bias, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Arjoon, Janelle A.; Xu, Xiaoying; Lewis, Jennifer E. – Journal of Chemical Education, 2013
Many of the instruments developed for research use by the chemistry education community are relatively new. Because psychometric evidence dictates the validity of interpretations made from test scores, gathering and reporting validity and reliability evidence is of utmost importance. Therefore, the purpose of this study was to investigate what…
Descriptors: Science Instruction, Measurement Techniques, Psychometrics, Evidence
Peer reviewed Peer reviewed
Direct linkDirect link
Wayne Au; Karen Gourd – English Journal, 2013
High-stakes standardized testing has become ubiquitous in US education. In this article, the authors argue that not only is such testing bad for education, but it also contradicts curriculum and instruction aligned with professional standards promoted by NCTE and innovative educators. Additionally, contradictory to the concept of "no child…
Descriptors: High Stakes Tests, Standardized Tests, Educational Assessment, English Teachers
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sabatini, John; Bruce, Kelly; Steinberg, Jonathan; Weeks, Jonathan – ETS Research Report Series, 2015
This technical report describes the conceptual foundation and measurement properties of the Reading Inventory and Scholastic Evaluation (RISE). The RISE is a 6-subtest, Web-administered reading skills components battery. The theoretical and empirical foundations of each subtest in the battery are reviewed, as well as item designs. The results…
Descriptors: Reading Tests, Reading Skills, Test Construction, Research Reports
Peer reviewed Peer reviewed
Direct linkDirect link
Moreno, Gerardo; Wong-Lo, Mickie; Short, Maureen; Bullock, Lyndal M. – Emotional & Behavioural Difficulties, 2014
As the US student population continues to become increasingly diverse, educators have encountered difficulties in distinguishing between cultural differences and genuine disability indicators. This concern is clearly evident in assisting students from diverse backgrounds who demonstrate chronic challenging behaviours. Past practices (e.g.…
Descriptors: Student Diversity, Ethnic Diversity, Functional Behavioral Assessment, Cultural Relevance
Peer reviewed Peer reviewed
Direct linkDirect link
Kan, Adnan; Bulut, Okan – International Journal of Testing, 2014
This study investigated whether the linguistic complexity of items leads to gender differential item functioning (DIF) on mathematics assessments. Two forms of a mathematics test were developed. The first form consisted of algebra items based on mathematical expressions, terms, and equations. In the second form, the same items were written as word…
Descriptors: Gender Differences, Test Bias, Difficulty Level, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Noble, Tracy; Kachchaf, Rachel; Rosebery, Ann; Warren, Beth; O'Connor, Mary Catherine; Wang, Yang – Grantee Submission, 2014
Little research has examined individual linguistic features that influence English language learners (ELLs) test performance. Furthermore, research has yet to explore the relationship between the science strand of test items and the types of linguistic features the items include. Utilizing Differential Item Functioning, this study examines ELL…
Descriptors: Science Tests, English Language Learners, Linguistics, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Gao, Yong; Mack, Mick G.; Ragan, Moira A.; Ragan, Brian – Measurement in Physical Education and Exercise Science, 2012
In this study the authors used differential item functioning analysis to examine if there were items in the Mental, Emotional, and Bodily Toughness Inventory functioning differently across gender and athletic membership. A total of 444 male (56.3%) and female (43.7%) participants (30.9% athletes and 69.1% non-athletes) responded to the Mental,…
Descriptors: Test Bias, Physical Activities, Athletes, Measures (Individuals)
Peer reviewed Peer reviewed
Direct linkDirect link
Walker, Cindy M.; Zhang, Bo; Banks, Kathleen; Cappaert, Kevin – Educational and Psychological Measurement, 2012
The purpose of this simulation study was to establish general effect size guidelines for interpreting the results of differential bundle functioning (DBF) analyses using simultaneous item bias test (SIBTEST). Three factors were manipulated: number of items in a bundle, test length, and magnitude of uniform differential item functioning (DIF)…
Descriptors: Test Bias, Test Length, Simulation, Guidelines
Pages: 1  |  ...  |  52  |  53  |  54  |  55  |  56  |  57  |  58  |  59  |  60  |  ...  |  248