Publication Date
| In 2026 | 0 |
| Since 2025 | 215 |
| Since 2022 (last 5 years) | 1084 |
| Since 2017 (last 10 years) | 2594 |
| Since 2007 (last 20 years) | 4955 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedChang, Hua-Hua; Qian, Jiahe; Yang, Zhiliang – Applied Psychological Measurement, 2001
Proposed a refinement, based on the stratification of items developed by D. Weiss (1973), of the computerized adaptive testing item selection procedure of H. Chang and Z. Ying (1999). Simulation studies using an item bank from the Graduate Record Examination show the benefits of the new procedure. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Simulation
Peer reviewedBandalos, Deborah L. – Structural Equation Modeling, 2002
Used simulation to study the effects of the practice of item parceling. Results indicate that certain types of item parceling can obfuscate a multidimensional factor structure in a way that acceptable values of fit indexes are found for a misspecified solution. Discusses why the use of parceling cannot be recommended when items are…
Descriptors: Estimation (Mathematics), Factor Structure, Goodness of Fit, Test Items
Peer reviewedMuraki, Eiji – Journal of Educational Measurement, 1999
Extended an Item Response Theory (IRT) method for detection of differential item functioning to the partial credit model and applied the method to simulated data using a stepwise procedure. Then applied the stepwise DIF analysis based on the multiple-group partial credit model to writing trend data from the National Assessment of Educational…
Descriptors: Groups, Item Bias, Item Response Theory, Simulation
Peer reviewedAnkenmann, Robert D. – Journal of Educational Measurement, 1996
This book is designed to be an instructional guide rather than a technical manual. For that reason, it provides a comprehensive and integrated overview of the procedures for detecting differential item functioning with citations to more technically detailed references. (SLD)
Descriptors: Evaluation Methods, Identification, Item Bias, Test Construction
Peer reviewedZeng, Lingjia – Applied Psychological Measurement, 1997
Proposes a marginal Bayesian estimation procedure to improve item parameter estimates for the three parameter logistic model. Computer simulation suggests that implementing the marginal Bayesian estimation algorithm with four-parameter beta prior distributions and then updating the priors with empirical means of updated intermediate estimates can…
Descriptors: Algorithms, Bayesian Statistics, Estimation (Mathematics), Statistical Distributions
Peer reviewedMaller, Susan J. – Educational and Psychological Measurement, 2001
Used the national standardization sample (n=2,200) of the Wechsler Intelligence Scale for Children Third Edition (WISC-III) to investigate differential item functioning (DIF) in 6 WISC-III subtests. Detected both uniform DIF and nonuniform DIF, finding DIF for about one third of the items studied. Discusses implications for use of the WISC-III.…
Descriptors: Children, Intelligence Tests, Item Bias, Test Items
Peer reviewedPlake, Barbara S.; Impara, James C. – Educational Assessment, 2001
Examined the reliability and accuracy of item performance estimates from an Angoff standard setting application with 29 panelists on 1 year and 30 in the next year. Results provide evidence that item performance estimates were both reasonable and reliable. Discusses factors that might have influenced the results. (SLD)
Descriptors: Estimation (Mathematics), Evaluators, Performance Factors, Reliability
Wang, Wen-Chung; Chen, Po-Hsi; Cheng, Ying-Yao – Psychological Methods, 2004
A conventional way to analyze item responses in multiple tests is to apply unidimensional item response models separately, one test at a time. This unidimensional approach, which ignores the correlations between latent traits, yields imprecise measures when tests are short. To resolve this problem, one can use multidimensional item response models…
Descriptors: Item Response Theory, Test Items, Testing, Test Validity
Sijtsma, Klaas; van der Ark, L. Andries – Multivariate Behavioral Research, 2003
This article first discusses a statistical test for investigating whether or not the pattern of missing scores in a respondent-by-item data matrix is random. Since this is an asymptotic test, we investigate whether it is useful in small but realistic sample sizes. Then, we discuss two known simple imputation methods, person mean (PM) and two-way…
Descriptors: Test Items, Questionnaires, Statistical Analysis, Models
DeMars, Christine E. – Applied Measurement in Education, 2004
Three methods of detecting item drift were compared: the procedure in BILOG-MG for estimating linear trends in item difficulty, the CUSUM procedure that Veerkamp and Glas (2000) used to detect trends in difficulty or discrimination, and a modification of Kim, Cohen, and Park's (1995) x 2 test for multiple-group differential item functioning (DIF),…
Descriptors: Comparative Analysis, Test Items, Testing, Item Analysis
Martiniello, Maria – Educational Assessment, 2009
This article examines nonmathematical linguistic complexity as a source of differential item functioning (DIF) in math word problems for English language learners (ELLs). Specifically, this study investigates the relationship between item measures of linguistic complexity, nonlinguistic forms of representation and DIF measures based on item…
Descriptors: Mathematics Tests, Grade 4, Test Bias, Word Problems (Mathematics)
Ministerial Council on Education, Employment, Training and Youth Affairs (NJ1), 2008
The second national sample assessment of Information and Communication Technology (ICT) literacy at Year 6 and Year 10 will occur in 2008. The assessment happens every three years and the first occurred in 2005. This document provides information about the ICT literacy assessment including: (1) Education Ministers' decisions regarding ICT; (2) the…
Descriptors: Foreign Countries, National Programs, Testing Programs, National Competency Tests
D'Agostino, Jerome V.; Welsh, Megan E.; Cimetta, Adriana D.; Falco, Lia D.; Smith, Shannon; VanWinkle, Waverely Hester; Powers, Sonya J. – Applied Measurement in Education, 2008
Central to the standards-based assessment validation process is an examination of the alignment between state standards and test items. Several alignment analysis systems have emerged recently, but most rely on either traditional rating or matching techniques. Little, if any, analyses have been reported on the degree of consistency between the two…
Descriptors: Test Items, Student Evaluation, State Standards, Evaluation Methods
Brooks, Michelle; Davies, Stephanie – British Journal of Learning Disabilities, 2008
People with a learning disability, the experts of their own experience are increasingly involved in research. We will be discussing in this paper their centrality in the development of a psychological therapy outcome measure for people with learning disabilities. Their involvement needs to go beyond giving their views to being included in the…
Descriptors: Participatory Research, Learning Disabilities, Psychotherapy, Measures (Individuals)
Koong, Chorng-Shiuh; Wu, Chi-Ying – Computers & Education, 2010
Multiple intelligences, with its hypothesis and implementation, have ascended to a prominent status among the many instructional methodologies. Meanwhile, pedagogical theories and concepts are in need of more alternative and interactive assessments to prove their prevalence (Kinugasa, Yamashita, Hayashi, Tominaga, & Yamasaki, 2005). In general,…
Descriptors: Multiple Intelligences, Test Items, Grading, Programming

Direct link
