Publication Date
| In 2026 | 0 |
| Since 2025 | 38 |
| Since 2022 (last 5 years) | 225 |
| Since 2017 (last 10 years) | 570 |
| Since 2007 (last 20 years) | 1377 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 110 |
| Practitioners | 107 |
| Teachers | 46 |
| Administrators | 25 |
| Policymakers | 24 |
| Counselors | 12 |
| Parents | 7 |
| Students | 7 |
| Support Staff | 4 |
| Community | 2 |
Location
| California | 61 |
| Canada | 60 |
| United States | 57 |
| Turkey | 47 |
| Australia | 43 |
| Florida | 34 |
| Germany | 26 |
| Texas | 26 |
| China | 25 |
| Netherlands | 25 |
| Iran | 22 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Westby, Carol – Communication Disorders Quarterly, 2007
This article reviews the concept of intelligence from different cultural perspectives and explains why the traditional approach to determining "who is smart" is inappropriate for students from culturally/linguistically diverse backgrounds and inadequate even for determining if mainstream students will be successful in daily living. The concept of…
Descriptors: Intelligence, Cultural Differences, Cultural Relevance, Student Diversity
Papanastasiou, Elena C.; Reckase, Mark D. – International Journal of Testing, 2007
Because of the increased popularity of computerized adaptive testing (CAT), many admissions tests, as well as certification and licensure examinations, have been transformed from their paper-and-pencil versions to computerized adaptive versions. A major difference between paper-and-pencil tests and CAT from an examinee's point of view is that in…
Descriptors: Simulation, Adaptive Testing, Computer Assisted Testing, Test Items
Meade, Adam W.; Lautenschlager, Gary J.; Johnson, Emily C. – Applied Psychological Measurement, 2007
This article highlights issues associated with the use of the differential functioning of items and tests (DFIT) methodology for assessing measurement invariance (or differential functioning) with Likert-type data. Monte Carlo analyses indicate relatively low sensitivity of the DFIT methodology for identifying differential item functioning (DIF)…
Descriptors: Measures (Individuals), Monte Carlo Methods, Likert Scales, Effect Size
Birenbaum, Menucha – Studies in Educational Evaluation, 2007
High quality assessment practice is expected to yield valid and useful score-based interpretations about what the examinees know and are able to do with respect to a defined target domain. Given this assertion, the article presents a framework based on the "unified view of validity," advanced by Cronbach and Messick over two decades ago, to assist…
Descriptors: Quality Control, Student Evaluation, Validity, Evaluation Methods
Braun, Henry; Zhang, Jinming; Vezzu, Sailesh – ETS Research Report Series, 2008
At present, although the percentages of students with disabilities (SDs) and/or students who are English language learners (ELL) excluded from a NAEP administration are reported, no statistical adjustment is made for these excluded students in the calculation of NAEP results. However, the exclusion rates for both SD and ELL students vary…
Descriptors: Research Methodology, Computation, Disabilities, English Language Learners
Sellbom, Martin; Bagby, R. Michael – Psychological Assessment, 2008
In the current investigation, the authors examined the validity of the L-r and K-r scales on the recently developed Minnesota Multiphasic Personality Inventory-2-Restructured Form (MMPI-2-RF; Y. S. Ben-Porath & A. Tellegen, in press) in measuring underreported response bias. Three archival samples previously collected for examining MMPI-2…
Descriptors: Schizophrenia, Response Style (Tests), Test Validity, Child Custody
Walker, Ruth; Barwell, Graham – International Journal for the Scholarship of Teaching and Learning, 2009
Peer bias is recognised as a primary factor in negative student perceptions of peer assessment strategies. This study trialled the use of classroom response systems, widely known as clickers, in small seminar classes in order to actively engage students in their subject's assessment process while providing the anonymity that would lessen the…
Descriptors: Educational Technology, Peer Evaluation, Audience Response Systems, Focus Groups
Zhang, Ying; Elder, Catherine – Language Assessment Quarterly, 2009
The College English Test-Spoken English Test is a nationwide spoken English test designed to assess the oral communicative ability of Chinese university and college students who have undertaken compulsory English study at a Chinese university. This article describes the test and evaluates it in terms of reliability, validity, authenticity,…
Descriptors: Test Results, Language Tests, Rating Scales, Foreign Countries
Vaughn, Brandon K. – Journal on School Educational Technology, 2008
This study considers the importance of contextual effects on the quality of assessments on item bias and differential item functioning (DIF) in measurement. Often, in educational studies, students are clustered in teachers or schools, and the clusters could impact psychometric issues yet are largely ignored by traditional item analyses. A…
Descriptors: Test Bias, Educational Assessment, Educational Quality, Context Effect
Thelk, Amy – Research & Practice in Assessment, 2008
Differential Item Functioning (DIF) occurs when there is a greater probability of solving an item based on group membership after controlling for ability. Following administration of a 50-item scientific and quantitative reasoning exam to 286 two-year and 1174 four-year students, items were evaluated for DIF. Two-year students performed…
Descriptors: Test Bias, Probability, Test Items, Student Evaluation
Coe, Robert – Oxford Review of Education, 2008
The comparability of examinations in different subjects has been a controversial topic for many years and a number of criticisms have been made of statistical approaches to estimating the "difficulties" of achieving particular grades in different subjects. This paper argues that if comparability is understood in terms of a linking…
Descriptors: Test Items, Grades (Scholastic), Foreign Countries, Test Bias
Ferrara, Steve – Educational Assessment, 2008
The No Child Left Behind Act of 2001 requires all states to assess the English proficiency of English language learners each school year. Under Title I and Title III of No Child Left Behind, states are required to measure the annual growth of students' English language development in reading, listening, writing, and speaking and in comprehension…
Descriptors: Speech Communication, Federal Legislation, Second Language Learning, Psychometrics
Sireci, Stephen G.; Han, Kyung T.; Wells, Craig S. – Educational Assessment, 2008
In the United States, when English language learners (ELLs) are tested, they are usually tested in English and their limited English proficiency is a potential cause of construct-irrelevant variance. When such irrelevancies affect test scores, inaccurate interpretations of ELLs' knowledge, skills, and abilities may occur. In this article, we…
Descriptors: Test Use, Educational Assessment, Psychological Testing, Validity
Peer reviewedDawson, George; And Others – Science Teacher, 1974
Discusses possible sources of bias in reading tests used to evaluate science materials. (PEB)
Descriptors: Bias, Readability, Reading Difficulty, Reading Skills
Southwest Regional Resource Center, Salt Lake City, UT. – 1977
Intended for personnel in State Educational Agencies, the document provides guidelines, procedures, and forms for implementation of nondiscriminatory assessment practices with handicapped children and adults. In an introductory section, a model is presented for establishing the relationships between various components of an unbiased assessment…
Descriptors: Conceptual Schemes, Disabilities, Evaluation Methods, Guidelines

Direct link
