Publication Date
In 2025 | 16 |
Since 2024 | 97 |
Since 2021 (last 5 years) | 273 |
Since 2016 (last 10 years) | 617 |
Since 2006 (last 20 years) | 1413 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
Researchers | 110 |
Practitioners | 107 |
Teachers | 46 |
Administrators | 25 |
Policymakers | 24 |
Counselors | 12 |
Parents | 7 |
Students | 7 |
Support Staff | 4 |
Community | 2 |
Location
California | 60 |
Canada | 60 |
United States | 56 |
Turkey | 47 |
Australia | 43 |
Florida | 34 |
Germany | 26 |
Texas | 26 |
Netherlands | 25 |
China | 24 |
Iran | 21 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 1 |
Meets WWC Standards with or without Reservations | 1 |
Does not meet standards | 1 |
Jiban Khadka; Dirgha Raj Joshi; Krishna Prasad Adhikari; Bishnu Khanal – Journal of Educators Online, 2025
This study aims to explore the impact of the fairness of semester-end e-assessment in terms of policy provision, monitoring, and authenticity. The cross-sectional online survey design was employed among 346 students at Nepal Open University (NOU). The results were analyzed by using t-test, analysis of variance, and structural equation modeling.…
Descriptors: Foreign Countries, College Students, Open Universities, Computer Assisted Testing
Zeyuan Jing – ProQuest LLC, 2023
This dissertation presents a comprehensive review of the evolution of DIF analysis within educational measurement from the 1980s to the present. The review elucidates the concept of DIF, particularly emphasizing the crucial role of grouping for exhibiting DIF. Then, the dissertation introduces an innovative modification to the newly developed…
Descriptors: Item Response Theory, Algorithms, Measurement, Test Bias
Joseph A. Rios; Jiayi Deng – Educational and Psychological Measurement, 2024
Rapid guessing (RG) is a form of non-effortful responding that is characterized by short response latencies. This construct-irrelevant behavior has been shown in previous research to bias inferences concerning measurement properties and scores. To mitigate these deleterious effects, a number of response time threshold scoring procedures have been…
Descriptors: Reaction Time, Scores, Item Response Theory, Guessing (Tests)
Leala Holcomb; Wyatte C. Hall; Stephanie J. Gardiner-Walsh; Jessica Scott – Journal of Deaf Studies and Deaf Education, 2025
This study critically examines the biases and methodological shortcomings in studies comparing deaf and hearing populations, demonstrating their implications for both the reliability and ethics of research in deaf education. Upon reviewing the 20 most-cited deaf-hearing comparison studies, we identified recurring fallacies such as the presumption…
Descriptors: Literature Reviews, Deafness, Social Bias, Test Bias
Stefanie A. Wind; Yuan Ge – Measurement: Interdisciplinary Research and Perspectives, 2024
Mixed-format assessments made up of multiple-choice (MC) items and constructed response (CR) items that are scored using rater judgments include unique psychometric considerations. When these item types are combined to estimate examinee achievement, information about the psychometric quality of each component can depend on that of the other. For…
Descriptors: Interrater Reliability, Test Bias, Multiple Choice Tests, Responses
William C. M. Belzak; Daniel J. Bauer – Journal of Educational and Behavioral Statistics, 2024
Testing for differential item functioning (DIF) has undergone rapid statistical developments recently. Moderated nonlinear factor analysis (MNLFA) allows for simultaneous testing of DIF among multiple categorical and continuous covariates (e.g., sex, age, ethnicity, etc.), and regularization has shown promising results for identifying DIF among…
Descriptors: Test Bias, Algorithms, Factor Analysis, Error of Measurement
Sooyong Lee; Suhwa Han; Seung W. Choi – Journal of Educational Measurement, 2024
Research has shown that multiple-indicator multiple-cause (MIMIC) models can result in inflated Type I error rates in detecting differential item functioning (DIF) when the assumption of equal latent variance is violated. This study explains how the violation of the equal variance assumption adversely impacts the detection of nonuniform DIF and…
Descriptors: Factor Analysis, Bayesian Statistics, Test Bias, Item Response Theory
Farida Agus Setiawati; Tria Widyastuti; Kartika Nur Fathiyah; Tiara Shafa Nabila – European Journal of Psychology and Educational Research, 2024
Data obtained through questionnaires sometimes respond to the items presented by social norms, so sometimes they do not suit themselves. High social desirability (SD) in non-cognitive measurements will cause item bias. Several ways are used to reduce item bias, including freeing respondents from not writing their names or being anonymous,…
Descriptors: Social Desirability, Test Bias, Self Concept, Undergraduate Students
Belzak, William C. M. – Educational Measurement: Issues and Practice, 2023
Test developers and psychometricians have historically examined measurement bias and differential item functioning (DIF) across a single categorical variable (e.g., gender), independently of other variables (e.g., race, age, etc.). This is problematic when more complex forms of measurement bias may adversely affect test responses and, ultimately,…
Descriptors: Test Bias, High Stakes Tests, Artificial Intelligence, Test Items
Farshad Effatpanah; Purya Baghaei; Hamdollah Ravand; Olga Kunina-Habenicht – International Journal of Testing, 2025
This study applied the Mixed Rasch Model (MRM) to the listening comprehension section of the International English Language Testing System (IELTS) to detect latent class differential item functioning (DIF) by exploring multiple profiles of second/foreign language listeners. Item responses of 462 examinees to an IELTS listening test were subjected…
Descriptors: Item Response Theory, Second Language Learning, Listening Comprehension, English (Second Language)
Jung Yeon Park; Sean Joo; Zikun Li; Hyejin Yoon – Educational Measurement: Issues and Practice, 2025
This study examines potential assessment bias based on students' primary language status in PISA 2018. Specifically, multilingual (MLs) and nonmultilingual (non-MLs) students in the United States are compared with regard to their response time as well as scored responses across three cognitive domains (reading, mathematics, and science).…
Descriptors: Achievement Tests, Secondary School Students, International Assessment, Test Bias
Minghui Yao; Yunjie Xu – Sociological Methods & Research, 2024
As a crucial method in organizational and social behavior research, self-report surveys must manage method bias. Method biases are distorted scores in survey response, distorted variance in variables, and distorted relational estimates between variables caused by method designs. Studies on method bias have focused on "post hoc"…
Descriptors: Statistical Bias, Social Science Research, Questionnaires, Test Bias
Tatiana Artamonova; Maria Hasler-Barker; Edna Velásquez – Journal of Latinos and Education, 2024
This paper discusses Texas Examinations of Educator Standards Program Languages Other Than English -- Spanish exam (TExES LOTE - Spanish) and its potential bias, particularly against teacher candidates with Spanish as a heritage language (HL) background. In Texas, most teacher candidates, or college students of Spanish preparing for secondary…
Descriptors: Language Tests, Test Bias, Spanish, Native Language
Hwanggyu Lim; Danqi Zhu; Edison M. Choe; Kyung T. Han – Journal of Educational Measurement, 2024
This study presents a generalized version of the residual differential item functioning (RDIF) detection framework in item response theory, named GRDIF, to analyze differential item functioning (DIF) in multiple groups. The GRDIF framework retains the advantages of the original RDIF framework, such as computational efficiency and ease of…
Descriptors: Item Response Theory, Test Bias, Test Reliability, Test Construction
Weese, James D.; Turner, Ronna C.; Ames, Allison; Crawford, Brandon; Liang, Xinya – Educational and Psychological Measurement, 2022
A simulation study was conducted to investigate the heuristics of the SIBTEST procedure and how it compares with ETS classification guidelines used with the Mantel-Haenszel procedure. Prior heuristics have been used for nearly 25 years, but they are based on a simulation study that was restricted due to computer limitations and that modeled item…
Descriptors: Test Bias, Heuristics, Classification, Statistical Analysis