Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
PEPNet-West, 2010
Public policy making relies on accurate information, but standardized and other mandated tests may not accurately evaluate the abilities and knowledge of individuals who are deaf or hard of hearing. Many individuals find tests difficult, but individuals who are deaf or hard of hearing may find them especially so. Reports from the 2008 Test Equity…
Descriptors: Partial Hearing, Public Policy, Deafness, Standardized Tests
Fidalgo, Angel M.; Scalon, Joao D. – Journal of Psychoeducational Assessment, 2010
In spite of the growing interest in cross-cultural research and assessment, there is little research on statistical procedures that can be used to simultaneously assess the differential item functioning (DIF) across multiple groups. The chief objective of this work is to show a unified framework for the analysis of DIF in multiple groups using one…
Descriptors: Test Bias, Statistics, Evaluation, Item Response Theory
Ercikan, Kadriye; Arim, Rubab; Law, Danielle; Domene, Jose; Gagnon, France; Lacroix, Serge – Educational Measurement: Issues and Practice, 2010
This paper demonstrates and discusses the use of think aloud protocols (TAPs) as an approach for examining and confirming sources of differential item functioning (DIF). The TAPs are used to investigate to what extent surface characteristics of the items that are identified by expert reviews as sources of DIF are supported by empirical evidence…
Descriptors: Test Bias, Protocol Analysis, Cognitive Processes, Expertise
Lee, John Chi-kin; Yin, Hongbiao; Zhang, Zhonghua – International Journal of Testing, 2010
This article reports the adaptation and analysis of Pintrich's Motivated Strategies for Learning Questionnaire (MSLQ) in Hong Kong. First, this study examined the psychometric qualities of the existing Chinese version of MSLQ (MSLQ-CV). Based on this examination, this study developed a revised Chinese version of MSLQ (MSLQ-RCV) for junior…
Descriptors: Foreign Countries, Questionnaires, Psychometrics, Secondary School Students
Stone, Elizabeth; Cook, Linda; Cahalan-Laitusis, Cara; Cline, Frederick – Applied Measurement in Education, 2010
This validity study examined differential item functioning (DIF) results on large-scale state standards-based English-language arts assessments at grades 4 and 8 for students without disabilities taking the test under standard conditions and students who are blind or visually impaired taking the test with either a large print or braille form.…
Descriptors: Test Bias, Large Type Materials, Testing Accommodations, Language Arts
Drummond, Todd W. – ProQuest LLC, 2011
Cross-lingual tests are assessment instruments created in one language and adapted for use with another language group. Practitioners and researchers use cross-lingual tests for various descriptive, analytical and selection purposes both in comparative studies across nations and within countries marked by linguistic diversity (Hambleton, 2005).…
Descriptors: Foreign Countries, Test Bias, High Stakes Tests, Test Items
Puhan, Gautam – International Journal of Testing, 2011
This study examined the effect of including or excluding repeaters on the equating process and results. New forms of two tests were equated to their respective old forms using either all examinees or only the first timer examinees in the new form sample. Results showed that for both tests used in this study, including or excluding repeaters in the…
Descriptors: Equated Scores, Educational Testing, Student Evaluation, Sample Size
Imus, Anna; Schmitt, Neal; Kim, Brian; Oswald, Frederick L.; Merritt, Stephanie; Wrestring, Alyssa Friede – Applied Measurement in Education, 2011
Investigations of differential item functioning (DIF) have been conducted mostly on ability tests and have found little evidence of easily interpretable differences across various demographic subgroups. In this study, we examined the degree to which DIF in biographical data items referencing academically relevant background, experiences, and…
Descriptors: Test Bias, Gender Differences, Racial Differences, Biographical Inventories
Rasch Analysis of the Assessment of Children's Hand Skills in Children with and without Disabilities
Chien, Chi-Wen; Brown, Ted; McDonald, Rachael – Research in Developmental Disabilities: A Multidisciplinary Journal, 2011
The Assessment of Children's Hand Skills (ACHS) is a new assessment tool that utilizes a naturalistic observational method to capture children's real-life hand skill performance when engaging in various types of activities. The ACHS also intends to be used with both typically developing children and those presenting with disabilities. The purpose…
Descriptors: Test Items, Construct Validity, Test Bias, Disabilities
Shih, Ching-Lin; Wang, Wen-Chung – Applied Psychological Measurement, 2009
The multiple indicators, multiple causes (MIMIC) method with a pure short anchor was proposed to detect differential item functioning (DIF). A simulation study showed that the MIMIC method with an anchor of 1, 2, 4, or 10 DIF-free items yielded a well-controlled Type I error rate even when such tests contained as many as 40% DIF items. In general,…
Descriptors: Test Bias, Simulation, Methods, Factor Analysis
Heilig, Julian Vasquez; Brown, Keffrelyn D.; Brown, Anthony L. – Harvard Educational Review, 2012
In this article, Julian Vasquez Heilig, Keffrelyn Brown, and Anthony Brown offer findings from a close textual analysis of how the Texas social studies standards address race, racism, and communities of color. Using the lens of critical race theory, the authors uncover the sometimes subtle ways that the standards can appear to adequately address…
Descriptors: State Standards, Critical Theory, Social Theories, Racial Factors
Nixon, Lisa – ProQuest LLC, 2013
The purpose of this mixed methods study was to determine the key implementation issues of a standards-based teacher evaluation system as perceived by campus administrators. The 80 campus administrators that participated in this study were from six public school districts located in southeastern Texas that serve students in grades Kindergarten…
Descriptors: Teacher Evaluation, State Standards, Program Implementation, Barriers
Kroopnick, Marc Howard – ProQuest LLC, 2010
When Item Response Theory (IRT) is operationally applied for large scale assessments, unidimensionality is typically assumed. This assumption requires that the test measures a single latent trait. Furthermore, when tests are vertically scaled using IRT, the assumption of unidimensionality would require that the battery of tests across grades…
Descriptors: Simulation, Scaling, Standard Setting, Item Response Theory
College Board, 2010
This is the College Board's response to a research article by Drs. Maria Veronica Santelices and Mark Wilson in the Harvard Educational Review, entitled "Unfair Treatment? The Case of Freedle, the SAT, and the Standardization Approach to Differential Item Functioning" (see EJ930622).
Descriptors: Test Bias, College Entrance Examinations, Standardized Tests, Test Items
Arendasy, Martin E.; Sommer, Markus – Intelligence, 2010
In complex three-dimensional mental rotation tasks males have been reported to score up to one standard deviation higher than females. However, this effect size estimate could be compromised by the presence of gender bias at the item level, which calls the validity of purely quantitative performance comparisons into question. We hypothesized that…
Descriptors: Effect Size, Psychometrics, Gender Differences, Gender Bias

Direct link
Peer reviewed
