Publication Date
In 2025 | 16 |
Since 2024 | 97 |
Since 2021 (last 5 years) | 273 |
Since 2016 (last 10 years) | 617 |
Since 2006 (last 20 years) | 1413 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
Researchers | 110 |
Practitioners | 107 |
Teachers | 46 |
Administrators | 25 |
Policymakers | 24 |
Counselors | 12 |
Parents | 7 |
Students | 7 |
Support Staff | 4 |
Community | 2 |
Location
California | 60 |
Canada | 60 |
United States | 56 |
Turkey | 47 |
Australia | 43 |
Florida | 34 |
Germany | 26 |
Texas | 26 |
Netherlands | 25 |
China | 24 |
Iran | 21 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Meets WWC Standards without Reservations | 1 |
Meets WWC Standards with or without Reservations | 1 |
Does not meet standards | 1 |
Christina M. Wilbar – ProQuest LLC, 2021
Historically, psychological tests were assumed to be objective, culture-free, and generalizable to any racial/ethnic minorities or socio-cultural group, even though most tests were standardized, validated, and found reliable with samples consisting only of White, middle-class, English- speaking individuals (The Council of National Psychology…
Descriptors: Individualized Education Programs, Test Bias, Special Education, Minority Group Children
Phuong Thi Tuyet Nguyen – ProQuest LLC, 2021
Source-based writing, in which writers read or listen to academic content before writing, has been considered to better assess academic writing skills than independent writing tasks (Read, 1990; Weigle, 2004). Because scores resulting from ratings of test takers' source-based writing task responses are treated as indicators of their academic…
Descriptors: Test Reliability, Test Validity, Writing Tests, Academic Language
Paris, Joseph H.; Torsney, Benjamin; Fiorot, Sara; Pressimone Beckowski, Catherine – Journal of College Access, 2022
An increasing number of postsecondary institutions in the United States have introduced test-optional admissions policies primarily due to criticism of standardized admissions tests as potentially biased predictors of student success. However, the impact of the test-optional movement is largely unknown and continues to evolve amid the COVID-19…
Descriptors: College Admission, Admission Criteria, College Entrance Examinations, Standardized Tests
Lions, Séverin; Monsalve, Carlos; Dartnell, Pablo; Blanco, María Paz; Ortega, Gabriel; Lemarié, Julie – Applied Measurement in Education, 2022
Multiple-choice tests are widely used in education, often for high-stakes assessment purposes. Consequently, these tests should be constructed following the highest standards. Many efforts have been undertaken to advance item-writing guidelines intended to improve tests. One important issue is the unwanted effects of the options' position on test…
Descriptors: Multiple Choice Tests, High Stakes Tests, Test Construction, Guidelines
Mumba, Brian – Journal on Educational Psychology, 2022
Researchers in educational measurement use Differential Item Functioning (DIF) to examine whether test items are functioning uniquely across subgroups of test participants while taking into account their ability level. DIF is essential for test validity arguments, thus making it a necessary part of validity studies. This study examines DIF across…
Descriptors: Test Bias, Test Items, Gender Differences, Grade 2
Daniel Katz; Anne Corinne Huggins-Manley; Walter Leite – Grantee Submission, 2022
According to the Standards for Educational and Psychological Testing (2014), one aspect of test fairness concerns examinees having comparable opportunities to learn prior to taking tests. Meanwhile, many researchers are developing platforms enhanced by artificial intelligence (AI) that can personalize curriculum to individual student needs. This…
Descriptors: High Stakes Tests, Test Bias, Testing Problems, Prior Learning
Daniel Katz; Anne Corinne Huggins-Manley; Walter Leite – Applied Measurement in Education, 2022
According to the "Standards for Educational and Psychological Testing" (2014), one aspect of test fairness concerns examinees having comparable opportunities to learn prior to taking tests. Meanwhile, many researchers are developing platforms enhanced by artificial intelligence (AI) that can personalize curriculum to individual student…
Descriptors: High Stakes Tests, Test Bias, Testing Problems, Prior Learning
Evan Rosenman; Rina Friedberg; Michael Baiocchi – Society for Research on Educational Effectiveness, 2022
Background and Context: In 2016, our team designed and implemented a cluster-randomized trial of a school-based empowerment training program, targeting adolescent girls in Nairobi, Kenya (Baiocchi et al., 2019; Rosenman et al., 2020). In that study, the primary outcome was the experience of sexual violence in the prior year. Participants disclosed…
Descriptors: Foreign Countries, Adolescents, Females, Sexual Abuse
Benjamin W. Y. Hornsby; Stephen Camarata; Sun-Joo Cho; Hilary Davis; Ronan McGarrigle; Fred H. Bess – Journal of Speech, Language, and Hearing Research, 2022
Purpose: Growing evidence suggests that fatigue associated with listening difficulties is particularly problematic for children with hearing loss (CHL). However, sensitive, reliable, and valid measures of listening-related fatigue do not exist. To address this gap, this article describes the development, psychometric evaluation, and preliminary…
Descriptors: Test Construction, Fatigue (Biology), Hearing Impairments, Listening
Rustam, Ahmad; Naga, Dali Santun; Supriyati, Yetti – International Journal of Education and Literacy Studies, 2019
Detection of differential item functioning (DIF) is needed in the development of tests to obtain useful items. The Mantel-Haenszel method and standardization are tools for DIF detection based on classical theory assumptions. The study was conducted to highlight the sensitivity and accuracy between the Mantel-Haenszel method and the standardization…
Descriptors: Statistical Analysis, Test Bias, Accuracy, Multiple Choice Tests
Wolkowitz, Amanda A.; Wright, Keith D. – Journal of Educational Measurement, 2019
This article explores the amount of equating error at a passing score when equating scores from exams with small samples sizes. This article focuses on equating using classical test theory methods of Tucker linear, Levine linear, frequency estimation, and chained equipercentile equating. Both simulation and real data studies were used in the…
Descriptors: Error Patterns, Sample Size, Test Theory, Test Bias
D'Urso, E. Damiano; Tijmstra, Jesper; Vermunt, Jeroen K.; De Roover, Kim – Educational and Psychological Measurement, 2023
Assessing the measurement model (MM) of self-report scales is crucial to obtain valid measurements of individuals' latent psychological constructs. This entails evaluating the number of measured constructs and determining which construct is measured by which item. Exploratory factor analysis (EFA) is the most-used method to evaluate these…
Descriptors: Factor Analysis, Measurement Techniques, Self Evaluation (Individuals), Psychological Patterns
Corradi, David – Assessment & Evaluation in Higher Education, 2023
Juries are a high-stake practice in higher education to assess complex competencies. However common, research remains behind in detailing the psychometric qualities of juries, especially when using rubrics or rating scales as an assessment tool. In this study, I analyze a case of a jury assessment (N = 191) of product development where both…
Descriptors: Court Litigation, Educational Practices, Higher Education, Rating Scales
Rodriguez, Rebekah M.; Silvia, Paul J.; Kaufman, James C.; Reiter-Palmon, Roni; Puryear, Jeb S. – Creativity Research Journal, 2023
The original 90-item Creative Behavior Inventory (CBI) was a landmark self-report scale in creativity research, and the 28-item brief form developed nearly 20 years ago continues to be a popular measure of everyday creativity. Relatively little is known, however, about the psychometric properties of this widely used scale. In the current research,…
Descriptors: Creativity Tests, Creativity, Creative Thinking, Psychometrics
Shear, Benjamin R. – Journal of Educational Measurement, 2018
When contextual features of test-taking environments differentially affect item responding for different test takers and these features vary across test administrations, they may cause differential item functioning (DIF) that varies across test administrations. Because many common DIF detection methods ignore potential DIF variance, this article…
Descriptors: Test Bias, Regression (Statistics), Hierarchical Linear Modeling