Publication Date
| In 2026 | 0 |
| Since 2025 | 220 |
| Since 2022 (last 5 years) | 1089 |
| Since 2017 (last 10 years) | 2599 |
| Since 2007 (last 20 years) | 4960 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 226 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 66 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedLinacre, John M.; Wright, Benjamin D. – Journal of Applied Measurement, 2002
Describes an extension to the Rasch model for fundamental measurement in which there is parameterization not only for examinee ability and item difficulty but also for judge severity. Discusses variants of this model and judging plans, and explains its use in an empirical testing situation. (SLD)
Descriptors: Ability, Difficulty Level, Evaluators, Item Response Theory
Peer reviewedMiller, G. Edward; Beretvas, S. Natasha – Journal of Applied Measurement, 2002
Presents empirically based item selection guidelines for moving the cut score on equated tests consisting of "n" dichotomous items calibrated assuming the Rasch model. Derivations of lemmas that underlie the guidelines are provided as well as a simulated example. (SLD)
Descriptors: Cutting Scores, Equated Scores, Item Response Theory, Selection
Peer reviewedRudas, Tamas; Zwick, Rebecca – Journal of Educational and Behavioral Statistics, 1997
The mixture index of fit (T. Rudas et al, 1994) is used to estimate the fraction of a population for which differential item functioning (DIF) occurs, and this approach is compared to the Mantel Haenszel test of DIF. The proposed noniterative procedure provides information about data portions contributing to DIF. (SLD)
Descriptors: Comparative Analysis, Estimation (Mathematics), Item Bias, Maximum Likelihood Statistics
Peer reviewedVeldkamp, Bernard P. – Applied Psychological Measurement, 2002
Presents two mathematical programming approaches for the assembly of ability tests from item pools calibrated under a multidimensional item response theory model. Item selection is based on the Fisher information matrix. Illustrates the method through empirical examples for a two-dimensional mathematics item pool. (SLD)
Descriptors: Ability, Item Banks, Item Response Theory, Selection
Peer reviewedDavis, Laurie Laughlin; Pastor, Dena A.; Dodd, Barbara G.; Chiang, Claire; Fitzpatrick, Steven J. – Journal of Applied Measurement, 2003
Examined the effectiveness of the Sympson-Hetter technique and rotated content balancing relative to no exposure control and no content rotation conditions in a computerized adaptive testing system based on the partial credit model. Simulation results show the Sympson-Hetter technique can be used with minimal impact on measurement precision,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Simulation
Peer reviewedEnright, Mary K.; Morley, Mary; Sheehan, Kathleen M. – Applied Measurement in Education, 2002
Studied the impact of systematic item feature variation on item statistical characteristics and the degree to which such information could be used as collateral information to supplement examinee performance data and reduce pretest sample size by generating 2 families of 48 word problem variants for the Graduate Record Examinations. Results with…
Descriptors: College Entrance Examinations, Sample Size, Statistical Analysis, Test Construction
Peer reviewedZwick, Rebecca; Thayer, Dorothy T. – Applied Psychological Measurement, 2002
Used a simulation to investigate the applicability to computerized adaptive test data of a differential item functioning (DIF) analysis method. Results show the performance of this empirical Bayes enhancement of the Mantel Haenszel DIF analysis method to be quite promising. (SLD)
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Item Bias
Peer reviewedde Gruijter, Dato N. M. – Applied Psychological Measurement, 1990
Following a brief discussion of test construction by linear programing, the results of a study by F. B. Baker and others (1988) with respect to a uniform target is replicated. It is demonstrated that the result depends on characteristics of the item pool. (SLD)
Descriptors: Item Response Theory, Linear Programing, Mathematical Models, Test Construction
Peer reviewedNicewander, W. Alan – Psychometrika, 1990
An estimate and upper-bound estimate for the reliability of a test composed of binary items is derived from the multidimensional latent trait theory of R. D. Bock and M. Aitken (1981). The practical uses of such estimates are discussed. (SLD)
Descriptors: Estimation (Mathematics), Factor Analysis, Item Response Theory, Test Items
Peer reviewedBaker, Frank B.; And Others – Applied Psychological Measurement, 1988
Linear programing was used to select items from item pools (N=1,500) based on one-, two-, and three-parameter models so that a target test information function (TTIF) was reached. Focus was on the distributional characteristics of selected items. The linear-programing approach focuses on the worst feature of the TTIF. (TJH)
Descriptors: Item Banks, Latent Trait Theory, Linear Programing, Test Construction
Peer reviewedD'Amato, Rik Carl; And Others – Journal of School Psychology, 1988
Investigated the overlap between the Wechsler Intelligence Scale for Children - Revised (WISC-R) and the Halstead-Reitan Neuropsychological Battery (HRNB) in light of their use in diagnosing children's learning problems using scores for children (N=1,181) on the WISC-R and the HRNB. Results showed primary overlap between measures was attributed to…
Descriptors: Adolescents, Children, Intelligence Tests, Test Items
Peer reviewedParshall, Cynthia G.; Miller, Timothy R. – Journal of Educational Measurement, 1995
Exact testing was evaluated as a method for conducting Mantel-Haenszel differential item functioning (DIF) analyses with relatively small samples. A series of computer simulations found that the asymptotic Mantel-Haenszel and the exact method yielded very similar results across sample size, levels of DIF, and data sets. (SLD)
Descriptors: Comparative Analysis, Computer Simulation, Identification, Item Bias
Peer reviewedBacon, Donald R.; And Others – Educational and Psychological Measurement, 1995
The potential for bias in reliability estimation and for errors in item selection when alpha or unit-weighted omega coefficients are used is explored under simulated conditions. Results suggest that composite reliability may be an assessment tool but should not be an item selection tool in structural equations. (SLD)
Descriptors: Bias, Estimation (Mathematics), Reliability, Selection
Peer reviewedAckerman, Terry A.; Evans, John A. – Applied Psychological Measurement, 1994
The effect of the conditioning score on the results of differential item functioning (DIF) analysis was examined with simulated data. The study demonstrates that results of DIF that rely on a conditioning score can be quite different depending on the conditioning variable that is selected. (SLD)
Descriptors: Construct Validity, Identification, Item Bias, Selection
Peer reviewedEngelhard, George, Jr. – Educational and Psychological Measurement, 1992
A historical perspective is provided of the concept of invariance in measurement theory, describing sample-invariant item calibration and item-invariant measurement of individuals. Invariance as a key measurement concept is illustrated through the measurement theories of E. L. Thorndike, L. L. Thurstone, and G. Rasch. (SLD)
Descriptors: Behavioral Sciences, Educational History, Measurement Techniques, Psychometrics


