Publication Date
| In 2026 | 0 |
| Since 2025 | 200 |
| Since 2022 (last 5 years) | 1070 |
| Since 2017 (last 10 years) | 2580 |
| Since 2007 (last 20 years) | 4941 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Ramirez, Sylvia Z.; Lukenbill, James F. – Research in Developmental Disabilities: A Multidisciplinary Journal, 2007
This paper describes the development of the fear survey for adults with mental retardation (FSAMR) and provides initial evidence of its psychometric properties. The FSAMR was designed to be sensitive to the assessment needs of individuals with mental retardation. The items were developed through open-ended interviews, a review of existing…
Descriptors: Psychometrics, Test Validity, Fear, Mental Retardation
Scharf, Eric M.; Baldwin, Lynne P. – Active Learning in Higher Education: The Journal of the Institute for Learning and Teaching, 2007
The reasoning behind popular methods for analysing the raw data generated by multiple choice question (MCQ) tests is not always appreciated, occasionally with disastrous results. This article discusses and analyses three options for processing the raw data produced by MCQ tests. The article shows that one extreme option is not to penalize a…
Descriptors: Guessing (Tests), Test Items, Multiple Choice Tests, Questioning Techniques
Herman, Joan L.; Osmundson, Ellen; Dietel, Ronald – Assessment and Accountability Comprehensive Center, 2010
This report describes the purposes of benchmark assessments and provides recommendations for selecting and using benchmark assessments--addressing validity, alignment, reliability, fairness and bias and accessibility, instructional sensitivity, utility, and reporting issues. We also present recommendations on building capacity to support schools'…
Descriptors: Multiple Choice Tests, Test Items, Benchmarking, Educational Assessment
O'Shea, Mary B. – ProQuest LLC, 2010
Although much is known about how students perform on standardized tests, little research exists concerning how students think and process while taking such tests. This mixed methods action research study was designed to investigate if a constructivist approach to test preparation could yield improved results for 37 English language arts freshmen…
Descriptors: Test Preparation, Test Items, Statistical Analysis, Grade 9
Nering, Michael L., Ed.; Ostini, Remo, Ed. – Routledge, Taylor & Francis Group, 2010
This comprehensive "Handbook" focuses on the most used polytomous item response theory (IRT) models. These models help us understand the interaction between examinees and test questions where the questions have various response categories. The book reviews all of the major models and includes discussions about how and where the models…
Descriptors: Guides, Item Response Theory, Test Items, Correlation
Weiss, Michael Kevin – ProQuest LLC, 2009
How can the secondary Geometry course serve as an opportunity for students to learn to "be like" a mathematician--that is, to acquire a mathematical sensibility? In the first part of this dissertation, I investigate what might be meant by "mathematical sensibility". By analyzing narratives of mathematicians and their work, I identify a collection…
Descriptors: Feedback (Response), Geometry, Mathematics Instruction, Secondary School Mathematics
Al-Shabatat, Ahmad Mohammad; Abbas, Merza; Ismail, Hairul Nizam – International Journal of Special Education, 2009
Many people believe that environmental factors promote giftedness and invest in many programs to adopt gifted students providing them with challenging activities. Intellectual giftedness is founded on fluid intelligence and extends to more specific abilities through the growth and inputs from the environment. Acknowledging the roles played by the…
Descriptors: Intelligence, Test Items, Academically Gifted, Foreign Countries
Cui, Ying; Leighton, Jacqueline P. – Journal of Educational Measurement, 2009
In this article, we introduce a person-fit statistic called the hierarchy consistency index (HCI) to help detect misfitting item response vectors for tests developed and analyzed based on a cognitive model. The HCI ranges from -1.0 to 1.0, with values close to -1.0 indicating that students respond unexpectedly or differently from the responses…
Descriptors: Test Length, Simulation, Correlation, Research Methodology
Sawaki, Yasuyo; Kim, Hae-Jin; Gentile, Claudia – Language Assessment Quarterly, 2009
In cognitive diagnosis a Q-matrix (Tatsuoka, 1983, 1990), which is an incidence matrix that defines the relationships between test items and constructs of interest, has great impact on the nature of performance feedback that can be provided to score users. The purpose of the present study was to identify meaningful skill coding categories that…
Descriptors: Feedback (Response), Test Items, Test Content, Identification
Fernandes, Anthony; Anhalt, Cynthia O.; Civil, Marta – Teaching Children Mathematics, 2009
To assess students' developing literacy, teachers regularly listen as individuals read. Similarly, listening to students talk about their mathematical reasoning is a powerful way for teachers to assess students' problem-solving abilities and mathematical understanding. Traditional paper-and-pencil tests have long been the main avenue to evaluate…
Descriptors: Mathematics Instruction, Test Items, Second Language Learning, Problem Solving
Tatsuoka, Curtis – Measurement: Interdisciplinary Research and Perspectives, 2009
In this commentary, the author addresses what is referred to as the deterministic input, noisy "and" gate (DINA) model. The author mentions concerns with how this model has been formulated and presented. In particular, the author points out that there is a lack of recognition of the confounding of profiles that generally arises and then discusses…
Descriptors: Test Items, Classification, Psychometrics, Item Response Theory
Xu, Xueli; von Davier, Matthias – ETS Research Report Series, 2008
Xu and von Davier (2006) demonstrated the feasibility of using the general diagnostic model (GDM) to analyze National Assessment of Educational Progress (NAEP) proficiency data. Their work showed that the GDM analysis not only led to conclusions for gender and race groups similar to those published in the NAEP Report Card, but also allowed…
Descriptors: National Competency Tests, Models, Data Analysis, Reading Tests
Elosua, Paula; Lopez-Jauregui, Alicia – Journal of Experimental Education, 2008
The comparison of scores from linguistically different tests is a twofold matter: the adaptation of tests and the comparison of scores. These 2 aspects of measurement invariance intersect at the need to guarantee the psychometric equivalence between the original and adapted versions. In this study, the authors examined comparability in 2 stages.…
Descriptors: Psychometrics, Item Response Theory, Equated Scores, Comparative Analysis
Leighton, Jacqueline P.; Gokiert, Rebecca J. – Educational Assessment, 2008
The purpose of the present investigation was to identify the relationship among different indicators of uncertainty that lead to potential item misalignment. The item-based indicators included ratings of ambiguity and cognitive complexity. The student-based indicators included (a) frequency of cognitive monitoring per item, (b) levels of…
Descriptors: Test Items, Cognitive Processes, Item Analysis, Self Concept
Kwak, Nohoon; And Others – 1997
This paper introduces a new method for detecting differential item functioning (DIF), the unsigned Mantel-Haenszel (UMH) statistic, and compares this method with two other chi-square methods, the Mantel-Haenszel (MH) and the absolute mean deviation (AMD) statistics, in terms of power and agreement between expected and actual false positive rates.…
Descriptors: Chi Square, Identification, Item Bias, Test Items

Peer reviewed
Direct link
