Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Turkan, Azmi; Cetin, Bayram – Journal of Education and Practice, 2017
Validity and reliability are among the most crucial characteristics of a test. One of the steps to make sure that a test is valid and reliable is to examine the bias in test items. The purpose of this study was to examine the bias in 2012 Placement Test items in terms of gender variable using Rasch Model in Turkey. The sample of this study was…
Descriptors: Item Response Theory, Gender Differences, Test Bias, Test Items
Martinková, Patricia; Drabinová, Adéla; Liaw, Yuan-Ling; Sanders, Elizabeth A.; McFarland, Jenny L.; Price, Rebecca M. – CBE - Life Sciences Education, 2017
We provide a tutorial on differential item functioning (DIF) analysis, an analytic method useful for identifying potentially biased items in assessments. After explaining a number of methodological approaches, we test for gender bias in two scenarios that demonstrate why DIF analysis is crucial for developing assessments, particularly because…
Descriptors: Test Bias, Test Items, Gender Bias, Science Tests
Kaspersen, Eivind; Pepin, Birgit; Sikko, Svein Arne – International Journal of Research & Method in Education, 2017
Several attempts have been made to measure and categorize beliefs and practices of mathematics teachers [Swan, M. 2006. "Designing and Using Research Instruments to Describe the Beliefs and Practices of Mathematics Teachers." "Research in Education" 75 (1): 58-70]. One of the reasons for measuring both beliefs and practices is…
Descriptors: Student Teacher Attitudes, Student Teachers, Beliefs, Mathematics Instruction
Li, Zhushan – Journal of Educational Measurement, 2014
Logistic regression is a popular method for detecting uniform and nonuniform differential item functioning (DIF) effects. Theoretical formulas for the power and sample size calculations are derived for likelihood ratio tests and Wald tests based on the asymptotic distribution of the maximum likelihood estimators for the logistic regression model.…
Descriptors: Test Bias, Sample Size, Statistical Analysis, Regression (Statistics)
Solano-Flores, Guillermo – Applied Measurement in Education, 2014
This article addresses validity and fairness in the testing of English language learners (ELLs)--students in the United States who are developing English as a second language. It discusses limitations of current approaches to examining the linguistic features of items and their effect on the performance of ELL students. The article submits that…
Descriptors: English Language Learners, Test Items, Probability, Test Bias
Maeda, Yukiko; Yoon, So Yoon – Journal of Psychoeducational Assessment, 2016
We investigated the extent to which the observed gender differences in mental rotation ability among the 2,468 freshmen studying engineering at a Midwest public university attributed to the gender bias of a test. The Revised Purdue Spatial Visualization Tests: Visualization of Rotations (Revised PSVT:R) is a spatial test frequently used to measure…
Descriptors: Gender Differences, Spatial Ability, College Freshmen, Engineering Education
Orosco, Michael J. – International Journal of Science and Mathematics Education, 2016
The psychometric properties of a 10-item math motivation scale were empirically validated with an independent sample consisting of 182 elementary-school students. Analysis of the model dimensionality supported a one-factor structure fit. Item parameter estimates from a Classical Test Theory framework revealed that most items were highly…
Descriptors: Psychometrics, Student Motivation, Mathematics Instruction, Elementary School Students
Ahmadi, Alireza; Bazvand, Ali Darabi – Iranian Journal of Language Teaching Research, 2016
Differential Item Functioning (DIF) exists when examinees of equal ability from different groups have different probabilities of successful performance in a certain item. This study examined gender differential item functioning across the PhD Entrance Exam of TEFL (PEET) in Iran, using both logistic regression (LR) and one-parameter item response…
Descriptors: Test Bias, Gender Bias, College Entrance Examinations, English (Second Language)
Kopf, Julia; Zeileis, Achim; Strobl, Carolin – Educational and Psychological Measurement, 2015
Differential item functioning (DIF) indicates the violation of the invariance assumption, for instance, in models based on item response theory (IRT). For item-wise DIF analysis using IRT, a common metric for the item parameters of the groups that are to be compared (e.g., for the reference and the focal group) is necessary. In the Rasch model,…
Descriptors: Test Items, Equated Scores, Test Bias, Item Response Theory
Avenia-Tapper, Brianna; Llosa, Lorena – Educational Assessment, 2015
This article addresses the issue of language-related construct-irrelevant variance on content area tests from the perspective of systemic functional linguistics. We propose that the construct relevance of language used in content area assessments, and consequent claims of construct-irrelevant variance and bias, should be determined according to…
Descriptors: English Language Learners, Science Tests, Difficulty Level, Language Usage
Akour, Mutasem; Sabah, Saed; Hammouri, Hind – Journal of Psychoeducational Assessment, 2015
The purpose of this study was to apply two types of Differential Item Functioning (DIF), net and global DIF, as well as the framework of Differential Step Functioning (DSF) to real testing data to investigate measurement invariance related to test language. Data from the Program for International Student Assessment (PISA)-2006 polytomously scored…
Descriptors: Test Bias, Science Tests, Test Items, Scoring
Davis, Julius; Martin, Danny Bernard – Journal of Urban Mathematics Education, 2018
Couched within a larger critique of assessment practices and how they are used to stigmatize African American children, the authors examine teachers' instructional practices in response to demands of increasing test scores. Many mathematics teachers might be unaware of how these test-driven instructional practices can simultaneously reflect…
Descriptors: Racial Bias, African American Children, Whites, Children
Achieve, Inc., 2018
This document describes the features of a statewide summative science assessment that has been designed to embody standards based on the "Framework for K-12 Science Education," such as the Next Generation Science Standards (NGSS)--to reflect its intent, grounded in the specific expectations of three-dimensional standards. Importantly,…
Descriptors: Summative Evaluation, Science Tests, Standardized Tests, National Standards
Alsadaawi, Abdullah Saleh – International Education Studies, 2017
The Saudi National Assessment Centre administers the Computer Science Teacher Test for teacher certification. The aim of this study is to explore gender differences in candidates' scores, and investigate dimensionality, reliability, and differential item functioning using confirmatory factor analysis and item response theory. The confirmatory…
Descriptors: Factor Structure, Test Items, Test Reliability, Teacher Certification
Demir, Ergul – Eurasian Journal of Educational Research, 2018
Purpose: The answer-copying tendency has the potential to detect suspicious answer patterns for prior distributions of statistical detection techniques. The aim of this study is to develop a valid and reliable measurement tool as a scale in order to observe the tendency of university students' copying of answers. Also, it is aimed to provide…
Descriptors: College Students, Cheating, Test Construction, Student Behavior

Peer reviewed
Direct link
