Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 9 |
Descriptor
Source
Author
Publication Type
Education Level
Elementary Education | 4 |
Elementary Secondary Education | 4 |
Grade 4 | 2 |
Grade 8 | 2 |
Intermediate Grades | 2 |
Junior High Schools | 2 |
Middle Schools | 2 |
Secondary Education | 2 |
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
United States | 2 |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 3 |
Assessments and Surveys
National Assessment of… | 32 |
SAT (College Admission Test) | 3 |
National Longitudinal Study… | 1 |
Program for International… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Muhammad, Gholnecsar E.; Ortiz, Nickolaus A.; Neville, Mary L. – Reading Teacher, 2021
According to data from the National Assessment of Educational Progress, or NAEP, the U.S. educational system has consistently failed Black and Brown children across both reading and mathematics. Educational research has further uncovered the ways that reading and mathematics assessment and curriculum are often biased and culturally and…
Descriptors: National Competency Tests, Reading Achievement, Mathematics Achievement, African American Students
Wenjing Guo – ProQuest LLC, 2021
Constructed response (CR) items are widely used in large-scale testing programs, including the National Assessment of Educational Progress (NAEP) and many district and state-level assessments in the United States. One unique feature of CR items is that they depend on human raters to assess the quality of examinees' work. The judgment of human…
Descriptors: National Competency Tests, Responses, Interrater Reliability, Error of Measurement
Fu, Jianbin – ETS Research Report Series, 2016
The multidimensional item response theory (MIRT) models with covariates proposed by Haberman and implemented in the "mirt" program provide a flexible way to analyze data based on item response theory. In this report, we discuss applications of the MIRT models with covariates to longitudinal test data to measure skill differences at the…
Descriptors: Item Response Theory, Longitudinal Studies, Test Bias, Goodness of Fit
Sachse, Karoline A.; Roppelt, Alexander; Haag, Nicole – Journal of Educational Measurement, 2016
Trend estimation in international comparative large-scale assessments relies on measurement invariance between countries. However, cross-national differential item functioning (DIF) has been repeatedly documented. We ran a simulation study using national item parameters, which required trends to be computed separately for each country, to compare…
Descriptors: Comparative Analysis, Measurement, Test Bias, Simulation
Mislevy, Robert J. – Educational Measurement: Issues and Practice, 2012
This article presents the author's observations on Neil Dorans's NCME Career Award Address: "The Contestant Perspective on Taking Tests: Emanations from the Statue within." He calls attention to some points that Dr. Dorans made in his address, and offers his thoughts in response.
Descriptors: Testing, Test Reliability, Psychometrics, Scores
Beretvas, S. Natasha; Cawthon, Stephanie W.; Lockhart, L. Leland; Kaye, Alyssa D. – Educational and Psychological Measurement, 2012
This pedagogical article is intended to explain the similarities and differences between the parameterizations of two multilevel measurement model (MMM) frameworks. The conventional two-level MMM that includes item indicators and models item scores (Level 1) clustered within examinees (Level 2) and the two-level cross-classified MMM (in which item…
Descriptors: Test Bias, Comparative Analysis, Test Items, Difficulty Level
Braun, Henry; Zhang, Jinming; Vezzu, Sailesh – ETS Research Report Series, 2008
At present, although the percentages of students with disabilities (SDs) and/or students who are English language learners (ELL) excluded from a NAEP administration are reported, no statistical adjustment is made for these excluded students in the calculation of NAEP results. However, the exclusion rates for both SD and ELL students vary…
Descriptors: Research Methodology, Computation, Disabilities, English Language Learners
Ockey, Gary J. – Language Assessment Quarterly, 2007
When testing English language learners (ELLs) in subject matter areas, construct irrelevant variance could result from English, the language in which the test is presented. Differential item functioning (DIF) techniques have been used to determine if items are operating differently for population subgroups and might therefore be appropriate for…
Descriptors: Test Bias, Validity, Word Problems (Mathematics), Mathematics Tests
Shermis, Mark D.; DiVesta, Francis J. – Rowman & Littlefield Publishers, Inc., 2011
"Classroom Assessment in Action" clarifies the multi-faceted roles of measurement and assessment and their applications in a classroom setting. Comprehensive in scope, Shermis and Di Vesta explain basic measurement concepts and show students how to interpret the results of standardized tests. From these basic concepts, the authors then…
Descriptors: Student Evaluation, Standardized Tests, Scores, Measurement
Newkirk, Thomas – 1977
The validity of current standardized competency tests for writing is in doubt as is the need for such testing at all. Some tests, especially those requiring little writing, may not be testing what they purport to test (content validity). Instructional validity (testing what has actually been taught) raises the issue that what is being tested is…
Descriptors: Basic Skills, Standardized Tests, Test Bias, Test Interpretation

Holmes, Barbara J. – Journal of Negro Education, 1986
Abolishing standards underlying teacher competency testing to reverse bias against minorities actually results in lowered expectations and excuses poor performance. Rather, we should advocate educational progress for Black children at all educational levels. Better preparation of current students will eventually result in an improved and larger…
Descriptors: Achievement Tests, Black Achievement, Black Students, Black Teachers
Holmes, Barbara J. – 1980
In recent years, the controversy surrounding testing has grown, and the charge of bias is the most often cited criticism of testing and assessment. A review of the literature indicates that psychometricians and other researchers speak of bias as a property of the test or of items in the test. Conversely, test critics speak of bias as a quality or…
Descriptors: Educational Assessment, Educational Objectives, Federal Programs, National Surveys

Olson, Paul A. – National Elementary Principal, 1975
The National Assessment of Educational Progress, in its power base, conception, creation, and execution, runs against the view that respects the local, unique culture as the prime determinant in education and supports the idea that education is an entity that a national program can clarify for delivery to children. (Author/IRT) Aspect of National…
Descriptors: Accountability, Cultural Context, Cultural Influences, Educational Assessment
Phelps, Richard P. – Fordham Report, 1999
The objections of testing experts to standardized testing are evaluated. The report begins with a foreword by Chester E. Finn, Jr., followed by an executive summary and an introduction. Four case studies include: (1) experts' opposition to high-stakes testing in Texas; (2) in North Carolina; (3) concerns raised in connection with the National…
Descriptors: Academic Achievement, Case Studies, Elementary Secondary Education, High Stakes Tests

Wainer, Howard – Journal of Educational Statistics, 1990
It is suggested that some of the technology applied to state Scholastic Aptitude Test scores to measure states' educational performance (particularly use of a truncated Gaussian model) may make it possible to adjust National Assessment of Educational Progress (NAEP) scores to make inferences about state educational progress possible. (SLD)
Descriptors: Academic Achievement, Educational Assessment, Elementary Secondary Education, Mathematical Models