Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Thurman, Carol – ProQuest LLC, 2009
The increased use of polytomous item formats has led assessment developers to pay greater attention to the detection of differential item functioning (DIF) in these items. DIF occurs when an item performs differently for two contrasting groups of respondents (e.g., males versus females) after controlling for differences in the abilities of the…
Descriptors: Test Items, Monte Carlo Methods, Test Bias, Educational Testing
Wells, Craig S.; Baldwin, Su; Hambleton, Ronald K.; Sireci, Stephen G.; Karatonis, Ana; Jirka, Stephen – Applied Measurement in Education, 2009
Score equity assessment is an important analysis to ensure inferences drawn from test scores are comparable across subgroups of examinees. The purpose of the present evaluation was to assess the extent to which the Grade 8 NAEP Math and Reading assessments for 2005 were equivalent across selected states. More specifically, the present study…
Descriptors: National Competency Tests, Test Bias, Equated Scores, Grade 8
Monnot, Matthew J.; Quirk, Stuart W.; Hoerger, Michael; Brewer, Linda – Psychological Assessment, 2009
An assessment of predictive bias was conducted on numerous scales of the Minnesota Multiphasic Personality Inventory-2 (MMPI-2; J. N. Butcher, W. G. Dahlstrom, J. R. Graham, A. Tellegen, & B. Kaemmer, 1989), including the Restructured Clinical (RC) scales, in the prediction of clinical diagnostic status for African American and Caucasian male…
Descriptors: Racial Bias, Personality Measures, African Americans, Whites
Wang, Wen-Chung; Shih, Ching-Lin; Yang, Chih-Chien – Educational and Psychological Measurement, 2009
This study implements a scale purification procedure onto the standard MIMIC method for differential item functioning (DIF) detection and assesses its performance through a series of simulations. It is found that the MIMIC method with scale purification (denoted as M-SP) outperforms the standard MIMIC method (denoted as M-ST) in controlling…
Descriptors: Test Items, Measures (Individuals), Test Bias, Evaluation Research
Banks, Kathleen – Educational Assessment, 2009
The purpose of this article is to describe and demonstrate a three-step process of using differential distractor functioning (DDF) in a post hoc analysis to understand sources of differential item functioning (DIF) in multiple-choice testing. The process is demonstrated on two multiple-choice tests that used complex alternatives (e.g., "No…
Descriptors: Test Bias, Multiple Choice Tests, Testing, Gender Differences
Miyazaki, Kei; Hoshino, Takahiro; Mayekawa, Shin-ichi; Shigemasu, Kazuo – Psychometrika, 2009
This study proposes a new item parameter linking method for the common-item nonequivalent groups design in item response theory (IRT). Previous studies assumed that examinees are randomly assigned to either test form. However, examinees can frequently select their own test forms and tests often differ according to examinees' abilities. In such…
Descriptors: Test Format, Item Response Theory, Test Items, Test Bias
Reshetar, Rosemary; Melican, Gerald J. – College Board, 2010
This paper discusses issues related to the design and psychometric work for mixed-format tests --tests containing both multiple-choice (MC) and constructed-response (CR) items. The issues of validity, fairness, reliability and score consistency can be addressed but for mixed-format tests there are many decisions to be made and no examination or…
Descriptors: Psychometrics, Test Construction, Multiple Choice Tests, Test Items
Wu, Pei-Chen; Huang, Tsai-Wei – Measurement and Evaluation in Counseling and Development, 2010
This study was to apply the mixed Rasch model to investigate person heterogeneity of Beck Depression Inventory-II-Chinese version (BDI-II-C) and its effects on dimensionality and construct validity. Person heterogeneity was reflected by two latent classes that differ qualitatively. Additionally, person heterogeneity adversely affected the…
Descriptors: Construct Validity, Validity, Depression (Psychology), Item Response Theory
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias
Conrad, Kendon J.; Riley, Barth B.; Conrad, Karen M.; Chan, Ya-Fen; Dennis, Michael L. – Evaluation Review, 2010
In assessing criminality, researchers have used counts of crimes, arrests, and so on, because interval measures were not available. Additionally, crime seriousness varies depending on demographic factors. This study examined the Crime and Violence Scale (CVS) regarding psychometric quality using item response theory (IRT) and invariance of the…
Descriptors: Racial Differences, Age Differences, Crime, Measures (Individuals)
Kim, Sooyeon; von Davier, Alina A.; Haberman, Shelby – Journal of Educational Measurement, 2008
This study addressed the sampling error and linking bias that occur with small samples in a nonequivalent groups anchor test design. We proposed a linking method called the synthetic function, which is a weighted average of the identity function and a traditional equating function (in this case, the chained linear equating function). Specifically,…
Descriptors: Equated Scores, Sample Size, Test Reliability, Comparative Analysis
Mouratidis, Athanasios; Lens, Willy; Sideridis, Georgios D. – Educational Psychology, 2010
In two cross-sectional studies, we investigated to what extent elementary (Study 1) and middle school (Study 2) students pursue similar, yet distinct, mastery-related and performance-related goals in physical education. We found that students were more likely to endorse outcome goals in conjunction with mastery-related goals and ability goals in…
Descriptors: Physical Education, Student Motivation, Program Effectiveness, Achievement Need
Ford, Donna Y. – Gifted Child Today, 2010
Each year, the U.S. Department of Education publishes its comprehensive report "The Condition of Education." As the author perused this year's edition, she was reminded of an important reality, one that all education professionals must be mindful of: America and its schools are more racially and culturally different than ever before. The term…
Descriptors: Academically Gifted, Student Diversity, Educational Indicators, Cultural Pluralism
ACT, Inc., 2013
This manual contains information about the American College Test (ACT) Plan® program. The principal focus of this manual is to document the Plan program's technical adequacy in light of its intended purposes. This manual supersedes the 2011 edition. The content of this manual responds to requirements of the testing industry as established in the…
Descriptors: College Entrance Examinations, Formative Evaluation, Evaluation Research, Test Bias
Burkes, LaShona L. – ProQuest LLC, 2009
The current study identified socioeconomic status (SES) group differences in student performance on an eighth grade mathematics assessment derived from the Third/Trends in International Mathematics and Science Study (TIMSS) 2003. Differential item functioning (DIF) methodology was applied to examine SES group differences on item performance for…
Descriptors: Test Bias, Socioeconomic Status, Student Characteristics, Grade 8

Direct link
Peer reviewed
