Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedChang, Hua-Hua; Qian, Jiahe; Yang, Zhiliang – Applied Psychological Measurement, 2001
Proposed a refinement, based on the stratification of items developed by D. Weiss (1973), of the computerized adaptive testing item selection procedure of H. Chang and Z. Ying (1999). Simulation studies using an item bank from the Graduate Record Examination show the benefits of the new procedure. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Simulation
Peer reviewedBandalos, Deborah L. – Structural Equation Modeling, 2002
Used simulation to study the effects of the practice of item parceling. Results indicate that certain types of item parceling can obfuscate a multidimensional factor structure in a way that acceptable values of fit indexes are found for a misspecified solution. Discusses why the use of parceling cannot be recommended when items are…
Descriptors: Estimation (Mathematics), Factor Structure, Goodness of Fit, Test Items
Peer reviewedMuraki, Eiji – Journal of Educational Measurement, 1999
Extended an Item Response Theory (IRT) method for detection of differential item functioning to the partial credit model and applied the method to simulated data using a stepwise procedure. Then applied the stepwise DIF analysis based on the multiple-group partial credit model to writing trend data from the National Assessment of Educational…
Descriptors: Groups, Item Bias, Item Response Theory, Simulation
Peer reviewedAnkenmann, Robert D. – Journal of Educational Measurement, 1996
This book is designed to be an instructional guide rather than a technical manual. For that reason, it provides a comprehensive and integrated overview of the procedures for detecting differential item functioning with citations to more technically detailed references. (SLD)
Descriptors: Evaluation Methods, Identification, Item Bias, Test Construction
Peer reviewedZeng, Lingjia – Applied Psychological Measurement, 1997
Proposes a marginal Bayesian estimation procedure to improve item parameter estimates for the three parameter logistic model. Computer simulation suggests that implementing the marginal Bayesian estimation algorithm with four-parameter beta prior distributions and then updating the priors with empirical means of updated intermediate estimates can…
Descriptors: Algorithms, Bayesian Statistics, Estimation (Mathematics), Statistical Distributions
Peer reviewedMaller, Susan J. – Educational and Psychological Measurement, 2001
Used the national standardization sample (n=2,200) of the Wechsler Intelligence Scale for Children Third Edition (WISC-III) to investigate differential item functioning (DIF) in 6 WISC-III subtests. Detected both uniform DIF and nonuniform DIF, finding DIF for about one third of the items studied. Discusses implications for use of the WISC-III.…
Descriptors: Children, Intelligence Tests, Item Bias, Test Items
Peer reviewedPlake, Barbara S.; Impara, James C. – Educational Assessment, 2001
Examined the reliability and accuracy of item performance estimates from an Angoff standard setting application with 29 panelists on 1 year and 30 in the next year. Results provide evidence that item performance estimates were both reasonable and reliable. Discusses factors that might have influenced the results. (SLD)
Descriptors: Estimation (Mathematics), Evaluators, Performance Factors, Reliability
Wang, Wen-Chung; Chen, Po-Hsi; Cheng, Ying-Yao – Psychological Methods, 2004
A conventional way to analyze item responses in multiple tests is to apply unidimensional item response models separately, one test at a time. This unidimensional approach, which ignores the correlations between latent traits, yields imprecise measures when tests are short. To resolve this problem, one can use multidimensional item response models…
Descriptors: Item Response Theory, Test Items, Testing, Test Validity
Sijtsma, Klaas; van der Ark, L. Andries – Multivariate Behavioral Research, 2003
This article first discusses a statistical test for investigating whether or not the pattern of missing scores in a respondent-by-item data matrix is random. Since this is an asymptotic test, we investigate whether it is useful in small but realistic sample sizes. Then, we discuss two known simple imputation methods, person mean (PM) and two-way…
Descriptors: Test Items, Questionnaires, Statistical Analysis, Models
DeMars, Christine E. – Applied Measurement in Education, 2004
Three methods of detecting item drift were compared: the procedure in BILOG-MG for estimating linear trends in item difficulty, the CUSUM procedure that Veerkamp and Glas (2000) used to detect trends in difficulty or discrimination, and a modification of Kim, Cohen, and Park's (1995) x 2 test for multiple-group differential item functioning (DIF),…
Descriptors: Comparative Analysis, Test Items, Testing, Item Analysis
Thelk, Amy – Research & Practice in Assessment, 2008
Differential Item Functioning (DIF) occurs when there is a greater probability of solving an item based on group membership after controlling for ability. Following administration of a 50-item scientific and quantitative reasoning exam to 286 two-year and 1174 four-year students, items were evaluated for DIF. Two-year students performed…
Descriptors: Test Bias, Probability, Test Items, Student Evaluation
Sinharay, Sandip; Holland, Paul W. – Educational Testing Service, 2008
The nonequivalent groups with anchor test (NEAT) design involves missing data that are missing by design. Three popular equating methods that can be used with a NEAT design are the poststratification equating method, the chain equipercentile equating method, and the item-response-theory observed-score-equating method. These three methods each…
Descriptors: Equated Scores, Test Items, Item Response Theory, Data
Lu, Irene R. R.; Thomas, D. Roland – Structural Equation Modeling: A Multidisciplinary Journal, 2008
This article considers models involving a single structural equation with latent explanatory and/or latent dependent variables where discrete items are used to measure the latent variables. Our primary focus is the use of scores as proxies for the latent variables and carrying out ordinary least squares (OLS) regression on such scores to estimate…
Descriptors: Least Squares Statistics, Computation, Item Response Theory, Structural Equation Models
Martiniello, Maria – Harvard Educational Review, 2008
In this article, Maria Martiniello reports the findings of a study of the linguistic complexity of math word problems that were found to exhibit differential item functioning for English-language learners (ELLs) and non-ELLs taking the Massachusetts Comprehensive Assessment System (MCAS) fourth-grade math test. It builds on prior research showing…
Descriptors: Reading Comprehension, Test Items, Protocol Analysis, Mathematics Tests
Solano-Flores, Guillermo – Educational Researcher, 2008
The testing of English language learners (ELLs) is, to a large extent, a random process because of poor implementation and factors that are uncertain or beyond control. Yet current testing practices and policies appear to be based on deterministic views of language and linguistic groups and erroneous assumptions about the capacity of assessment…
Descriptors: Generalizability Theory, Testing, Second Language Learning, Error of Measurement

Direct link
