Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Gehlbach, Hunter – Journal of Early Adolescence, 2015
As pressure builds to assess students, teachers, and schools, educational practitioners and policy makers are increasingly looking toward student perception surveys as a promising means to collect high-quality, useful data. For instance, the widely cited Measures of Effective Teaching study lists student perception surveys as one of the three key…
Descriptors: Surveys, Evaluation Methods, Early Adolescents, Student Evaluation
Sha, Li; Schunn, Christian; Bathgate, Meghan – Journal of Research in Science Teaching, 2015
Cumulatively, participation in optional science learning experiences in school, after school, at home, and in the community may have a large impact on student interest in and knowledge of science. Therefore, interventions can have large long-term effects if they change student choice preferences for such optional science learning experiences. To…
Descriptors: Grade 5, Grade 6, Early Adolescents, Learning Experience
Wu, Yi-Fang – ProQuest LLC, 2015
Item response theory (IRT) uses a family of statistical models for estimating stable characteristics of items and examinees and defining how these characteristics interact in describing item and test performance. With a focus on the three-parameter logistic IRT (Birnbaum, 1968; Lord, 1980) model, the current study examines the accuracy and…
Descriptors: Item Response Theory, Test Items, Accuracy, Computation
Partnership for Assessment of Readiness for College and Careers, 2015
The Partnership for Assessment of Readiness for College and Careers (PARCC) is a group of states working together to develop a modern assessment that replaces previous state standardized tests. It provides better information for teachers and parents to identify where a student needs help, or is excelling, so they are able to enhance instruction to…
Descriptors: Literacy, Language Arts, Scoring Formulas, Scoring
Shin, Hyo Jeong – ProQuest LLC, 2015
This dissertation is comprised of three papers that propose and apply psychometric models to deal with complexities and challenges in large-scale assessments, focusing on modeling rater effects and complex learning progressions. In particular, three papers investigate extensions and applications of multilevel and multidimensional item response…
Descriptors: Item Response Theory, Psychometrics, Models, Measurement
Ferrara, Steve; Steedle, Jeffrey; Kinsman, Amy – Partnership for Assessment of Readiness for College and Careers, 2015
We report results from the following three analyses of PARCC [Partnership for Assessment of Readiness for College and Careers] cognitive complexity measures, based on 2014 field test item and task development and field test data. We conducted classification and regression tree analyses using 2014 PARCC field test data to do the following: (1)…
Descriptors: Cognitive Processes, Difficulty Level, Test Items, Mathematics Tests
Petr Kacovsky – Journal of Baltic Science Education, 2015
This study describes a research focused on grammar school students' misconceptions in the context of common, everyday thermal phenomena. Almost 500 Czech grammar school students from 24 classes were asked to fill in the reduced version of the Thermal Concept Evaluation in order to measure the asset of traditional instruction in eliminating…
Descriptors: Misconceptions, Science Instruction, Elementary School Students, Scientific Concepts
Raykov, Tenko; Marcoulides, George A.; Lee, Chun-Lung; Chang, Chi – Educational and Psychological Measurement, 2013
This note is concerned with a latent variable modeling approach for the study of differential item functioning in a multigroup setting. A multiple-testing procedure that can be used to evaluate group differences in response probabilities on individual items is discussed. The method is readily employed when the aim is also to locate possible…
Descriptors: Test Bias, Statistical Analysis, Models, Hypothesis Testing
Gains in Marking Reliability from Item-Level Marking: Is the Sum of the Parts Better than the Whole?
Wheadon, Christopher; Pinot de Moira, Anne – Educational Research and Evaluation, 2013
Marking of high-stakes examinations in England has traditionally been administered by schools and colleges sending their examination papers directly to examiners. As a consequence, the work of one candidate has, historically, been marked by one examiner, as has work of an entire centre. Previous studies have suggested that the marking of both…
Descriptors: Foreign Countries, Scoring, High Stakes Tests, Reliability
Simonsen, Hanne Gram; Lind, Marianne; Hansen, Pernille; Holm, Elisabeth; Mevik, Bjorn-Helge – Clinical Linguistics & Phonetics, 2013
In this article, we present a study of imageability ratings for a set of 1599 Norwegian words (896 nouns, 483 verbs and 220 adjectives) from a web-based survey. To a large extent, the results are in accordance with previous studies of other languages: high imageability scores in general, higher imageability scores for nouns than for verbs, and an…
Descriptors: Test Items, Children, Linguistics, Test Construction
Wolf, Raffaela – ProQuest LLC, 2013
Preservation of equity properties was examined using four equating methods--IRT True Score, IRT Observed Score, Frequency Estimation, and Chained Equipercentile--in a mixed-format test under a common-item nonequivalent groups (CINEG) design. Equating of mixed-format tests under a CINEG design can be influenced by factors such as attributes of the…
Descriptors: Testing, Item Response Theory, Equated Scores, Test Items
He, Wei; Diao, Qi; Hauser, Carl – Online Submission, 2013
This study compares the four existing procedures handling the item selection in severely constrained computerized adaptive tests (CAT). These procedures include weighted deviation model (WDM), weighted penalty model (WPM), maximum priority index (MPI), and shadow test approach (STA). Severely constrained CAT refer to those adaptive tests seeking…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Item Banks
Jin, Ying – ProQuest LLC, 2013
Previous research has demonstrated that DIF methods that do not account for multilevel data structure could result in too frequent rejection of the null hypothesis (i.e., no DIF) when the intraclass correlation coefficient (?) of the studied item was the same as ? of the total score. The current study extended previous research by comparing the…
Descriptors: Test Bias, Models, Correlation, Test Items
Adams, Raymond J.; Wu, Margaret L.; Wilson, Mark – Educational and Psychological Measurement, 2012
The Rasch rating (or partial credit) model is a widely applied item response model that is used to model ordinal observed variables that are assumed to collectively reflect a common latent variable. In the application of the model there is considerable controversy surrounding the assessment of fit. This controversy is most notable when the set of…
Descriptors: Item Response Theory, Models, Computation, Classification
Andrich, David; Hagquist, Curt – Journal of Educational and Behavioral Statistics, 2012
The literature in modern test theory on procedures for identifying items with differential item functioning (DIF) among two groups of persons includes the Mantel-Haenszel (MH) procedure. Generally, it is not recognized explicitly that if there is real DIF in some items which favor one group, then as an artifact of this procedure, artificial DIF…
Descriptors: Test Bias, Test Items, Item Response Theory, Statistical Analysis

Peer reviewed
Direct link
