Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Magis, David; De Boeck, Paul – Educational and Psychological Measurement, 2012
The identification of differential item functioning (DIF) is often performed by means of statistical approaches that consider the raw scores as proxies for the ability trait level. One of the most popular approaches, the Mantel-Haenszel (MH) method, belongs to this category. However, replacing the ability level by the simple raw score is a source…
Descriptors: Test Bias, Data, Error of Measurement, Raw Scores
Wolfe, Edward W.; McVay, Aaron – Educational Measurement: Issues and Practice, 2012
Historically, research focusing on rater characteristics and rating contexts that enable the assignment of accurate ratings and research focusing on statistical indicators of accurate ratings has been conducted by separate communities of researchers. This study demonstrates how existing latent trait modeling procedures can identify groups of…
Descriptors: Researchers, Research, Correlation, Test Bias
Partnership for Assessment of Readiness for College and Careers, 2018
The purpose of this technical report is to describe the third operational administration of the Partnership for Assessment of Readiness for College and Careers (PARCC) assessments in the 2016-2017 academic year. PARCC is a state-led consortium creating next-generation assessments that, compared to traditional K-12 assessments, more accurately…
Descriptors: College Readiness, Career Readiness, Common Core State Standards, Language Arts
Markle, Ross; Olivera-Aguilar, Margarita; Jackson, Teresa; Noeth, Richard; Robbins, Steven – ETS Research Report Series, 2013
The "SuccessNavigator"™ assessment is an online, 30 minute self-assessment of psychosocial and study skills designed for students entering postsecondary education. In addition to providing feedback in areas such as classroom and study behaviors, commitment to educational goals, management of academic stress, and connection to social…
Descriptors: Self Evaluation (Individuals), Computer Assisted Testing, Test Reliability, Test Validity
Dizon, Francis Gary – ProQuest LLC, 2013
The 1979 landmark case of "Larry P. v. Riles" continues to be one of the most debated topics in school psychology. In this case, Judge Peckham ruled that standardized, norm-referenced intelligence tests were culturally biased towards African-Americans, resulting in overrepresentation of African-Americans in Educably Mentally Retarded…
Descriptors: School Psychologists, Court Litigation, Intelligence Tests, Standardized Tests
Carmichael, Colin – Mathematics Education Research Group of Australasia, 2013
With reports of declining enrolments in mathematics related degrees and low female participation rates in these degrees, the issue of gender differences in mathematics remains relevant. Results of recent studies suggest gender differences in mathematics are nuanced and that small differences in the early years can manifest as larger differences in…
Descriptors: Gender Differences, Mathematics Achievement, Longitudinal Studies, Foreign Countries
Hou, Likun – ProQuest LLC, 2013
Analyzing examinees' responses using cognitive diagnostic models (CDMs) have the advantages of providing richer diagnostic information. To ensure the validity of the results from these models, differential item functioning (DIF) in CDMs needs to be investigated. In this dissertation, the model-based DIF detection method, Wald-CDM procedure is…
Descriptors: Test Bias, Models, Cognitive Processes, Diagnostic Tests
Chen, Ssu-Kuang; Hwang, Fang-Ming; Lin, Sunny S. J. – Social Indicators Research, 2013
A scale measuring quality of life (QOL) is important in adolescent research. Using the graded response model (GRM), this study evaluates the psychometric properties of the satisfaction ratings of the Quality of Life Profile Adolescent Version (QOLPAV). Data for 1,392 adolescents were used to check IRT assumptions such as unidimensionality and…
Descriptors: Quality of Life, Measures (Individuals), Life Satisfaction, Adolescents
Kline, Rex B. – Educational Research and Evaluation, 2013
Test fairness and test bias are not synonymous concepts. Test bias refers to statistical evidence that the psychometrics or interpretation of test scores depend on group membership, such as gender or race, when such differences are not expected. A test that is grossly biased may be judged to be unfair, but test fairness concerns the broader, more…
Descriptors: Factor Analysis, Social Justice, Psychometrics, Test Bias
Baker-Doyle, Kira; Petchauer, Emery – Teacher Education Quarterly, 2015
In many countries, including the United States, England, Korea, Hong Kong, and Japan, individuals must pass some form of examination for entry into or completion of a teacher education program (Wang, Coleman, Coley, & Phelps, 2003). These exams are meant to act as gatekeeping mechanisms for teacher quality. In the majority of the countries…
Descriptors: Teacher Evaluation, Licensing Examinations (Professions), High Stakes Tests, Standardized Tests
Suh, Youngsuk; Bolt, Daniel M. – Journal of Educational Measurement, 2011
In multiple-choice items, differential item functioning (DIF) in the correct response may or may not be caused by differentially functioning distractors. Identifying distractors as causes of DIF can provide valuable information for potential item revision or the design of new test items. In this paper, we examine a two-step approach based on…
Descriptors: Test Items, Test Bias, Multiple Choice Tests, Simulation
Cormier, Damien C.; McGrew, Kevin S.; Evans, Jeffrey J. – Journal of Psychoeducational Assessment, 2011
The linguistic demand of spoken instructions on individually administered norm-referenced psychological and educational tests is of concern when examining individuals who have varying levels of language processing ability or varying cultural backgrounds. The authors present a new method for analyzing the level of verbosity, complexity, and total…
Descriptors: Intelligence Tests, Oral Language, Difficulty Level, Test Bias
Kim, Eun Sook; Yoon, Myeongsun – Structural Equation Modeling: A Multidisciplinary Journal, 2011
This study investigated two major approaches in testing measurement invariance for ordinal measures: multiple-group categorical confirmatory factor analysis (MCCFA) and item response theory (IRT). Unlike the ordinary linear factor analysis, MCCFA can appropriately model the ordered-categorical measures with a threshold structure. A simulation…
Descriptors: Measurement, Factor Analysis, Item Response Theory, Comparative Analysis
Ding, Lin; Caballero, Marcos D. – Physical Review Special Topics - Physics Education Research, 2014
In a recent study, Caballero and colleagues conducted a large-scale evaluation using the Force Concept Inventory (FCI) to compare student learning outcomes between two introductory physics curricula: the Matter and Interactions (M&I) mechanics course and a pedagogically-reformed-traditional-content (PRTC) mechanics course. Using a conventional…
Descriptors: Science Instruction, Physics, Scientific Concepts, Mechanics (Physics)
Lewis, Scott E. – Journal of Chemical Education, 2014
Validity of educational research instruments and student assessments has appropriately become a growing interest in the chemistry education research community. Of particular concern is an attention to the consequences to students that result from the interpretation of assessment scores and whether those consequences are swayed by invalidity within…
Descriptors: Science Instruction, Chemistry, Test Validity, College Science

Peer reviewed
Direct link
