Publication Date
| In 2026 | 0 |
| Since 2025 | 215 |
| Since 2022 (last 5 years) | 1084 |
| Since 2017 (last 10 years) | 2594 |
| Since 2007 (last 20 years) | 4955 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Wang, Wen-Chung – 1998
The conventional two-group differential item functioning (DIF) analysis is extended to an analysis of variance-like (ANOVA-like) DIF analysis where multiple factors with multiple groups are compared simultaneously. Moreover, DIF is treated as a parameter to be estimated rather than simply a sign to be detected. This proposed approach allows the…
Descriptors: Analysis of Variance, Foreign Countries, Item Bias, Item Response Theory
Longford, Nicholas T. – 1994
This study is a critical evaluation of the roles for coding and scoring of missing responses to multiple-choice items in educational tests. The focus is on tests in which the test-takers have little or no motivation; in such tests omitting and not reaching (as classified by the currently adopted operational rules) is quite frequent. Data from the…
Descriptors: Algorithms, Classification, Coding, Models
Bowles, Ryan; Pommerich, Mary – 2001
Many arguments have been made against allowing examinees to review and change their answers after completing a computer adaptive test (CAT). These arguments include: (1) increased bias; (2) decreased precision; and (3) susceptibility of test-taking strategies. Results of simulations suggest that the strength of these arguments is reduced or…
Descriptors: Adaptive Testing, Algorithms, Computer Assisted Testing, Review (Reexamination)
De Ayala, R. J.; Plake, Barbara S.; Impara, James C.; Kozmicky, Michelle – 2000
This study investigated the effect on examinees' ability estimate under item response theory (IRT) when they are presented an item, have ample time to answer the item, but decide not to respond to the item. Simulation data were modeled on an empirical data set of 25,546 examinees that was calibrated using the 3-parameter logistic model. The study…
Descriptors: Ability, Estimation (Mathematics), Item Response Theory, Maximum Likelihood Statistics
Zhu, Renbang; Yu, Feng – 2003
To ensure fairness, it is of critical importance that testing programs make sure that essay items given to examinees are equivalent in difficulty. The purpose of this study was to evaluate the stability and accuracy of a logistic regression based polytomous essay difficulty index. Preliminary results from a simulation study (9 conditions with a…
Descriptors: Difficulty Level, Essay Tests, Indexes, Measurement Techniques
Thompson, Tony D.; Davey, Tim – 2000
This paper applies specific information item selection using a method developed by T. Davey and M. Fan (2000) to a multiple-choice passage-based reading test that is being developed for computer administration. Data used to calibrate the multidimensional item parameters for the simulation study consisted of item responses from randomly equivalent…
Descriptors: Adaptive Testing, Computer Assisted Testing, Reading Tests, Selection
Price, Larry R. – 1999
Data from a 50-item translated test used for certification were used to assess the percentage and type of agreement between the Mantel-Haenszel (MH) and Differential Functioning of Items and Tests (DFIT) techniques for the detection of differential item functioning (DIF). The DFIT procedure flagged 10 of 30 items as exhibiting significant DIF…
Descriptors: Certification, Item Bias, Licensing Examinations (Professions), Tables (Data)
Habick, Timothy – 1999
With the advent of computer-based testing (CBT) and the need to increase the number of items available in computer adaptive test pools, the idea of item variants was conceived. An item variant can be defined as an item with content based on an existing item to a greater or lesser degree. Item variants were first proposed as a way to enhance test…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Test Construction
Pommerich, Mary; Nicewander, W. Alan – 1998
A simulation study was performed to determine whether a group's average percent correct in a content domain could be accurately estimated for groups taking a single test form and not the entire domain of items. Six Item Response Theory (IRT)-based domain score estimation methods were evaluated, under conditions of few items per content area per…
Descriptors: Ability, Estimation (Mathematics), Groups, Item Response Theory
Optimal Stratification of Item Pools in a-Stratified Computerized Adaptive Testing. Research Report.
van der Linden, Wim J. – 2000
A method based on 0-1 linear programming (LP) is presented to stratify an item pool optimally for use in "alpha"-stratified adaptive testing. Because the 0-1 LP model belongs to the subclass of models with a network-flow structure, efficient solutions are possible. The method is applied to a previous item pool from the computerized…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Linear Programming
Louisiana State Department of Education, 2004
Louisiana Educational Assessment Program for the 21st Century (LEAP 21) is an integral part of the Louisiana school and district accountability system passed by the state legislature and signed into law in 1997. In March 2004, students in grade 8 took LEAP 21 English Language Arts, Mathematics, Science, and Social Studies tests. This document…
Descriptors: Academic Achievement, Grade 8, Test Items, Educational Quality
Peer reviewedLu, K. H. – Educational and Psychological Measurement, 1971
Descriptors: Difficulty Level, Statistical Analysis, Statistical Significance, Test Items
Peer reviewedAiken, Lewis R. – Educational and Psychological Measurement, 1982
Five types of multiple-choice items that can be used to assess higher-order educational objectives are examined. The item types do not exhaust the possibilities, but they are standard forms found helpful in writing items to measure more than recognitive memory. (Author/CM)
Descriptors: Cognitive Measurement, Educational Objectives, Evaluation Methods, Multiple Choice Tests
Peer reviewedLam, Tony C. M.; Klockars, Alan J. – Journal of Educational Measurement, 1982
Ratings given to questionnaire items on four types of rating scales were compared. This study shows that the differences between scales are contingent upon the particular anchors used for the intermediate options. The results suggest that the mean score is predictably influenced by changes in the intermediate anchors. (Author/PN)
Descriptors: Higher Education, Measurement Techniques, Measures (Individuals), Psychometrics
Peer reviewedSilverstein, A. B. – Journal of Consulting and Clinical Psychology, 1982
Assessed the validity of short forms that reduce the number of items within subtests rather than the number of subtests. Used data from the standardization samples for the Wechsler Intelligence Scale for Children, Wechsler Adult Intelligence Scale, Wechsler Preschool and Primary Scale of Intelligence, WISC-Revised, and WAIS-Revised. (Author)
Descriptors: Correlation, Intelligence Tests, Mathematical Formulas, Test Format


