NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 4,801 to 4,815 of 9,533 results Save | Export
Peer reviewed Peer reviewed
Cole, Jack; And Others – Adult Learning, 1993
Highlights test biases that can hamper the performance of adult learners. The biases can originate from the content of the questions or from special situations related to motivation, competition, membership in specific groups, or instructional methods. (JOW)
Descriptors: Adult Education, Literacy Education, Test Bias, Test Items
Peer reviewed Peer reviewed
Osterlind, Steven J. – Educational Research Quarterly, 1990
Criteria for planning, designing, and writing test items are suggested. The criteria were developed via a discussion by subject matter specialists, psychometricians, and test construction experts. Seven criteria proposed for test items of merit address the congruence of an item with its intended purpose, technical assumptions, and editorial…
Descriptors: Criteria, Guidelines, Test Construction, Test Items
Peer reviewed Peer reviewed
Sijtsma, Klaas; Hemker, Bas T. – Psychometrika, 1998
The absence of the invariant item ordering (IIO) property in two nonparametric polytomous item response theory (IRT) models is discussed, and two nonparametric models are discussed that imply an IIO. Only two parametric polytomous IRT models are found to imply an IIO. A method is proposed to investigate whether an IIO is implied with empirical…
Descriptors: Item Response Theory, Models, Nonparametric Statistics, Test Items
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T.; Lewis, Charles – Journal of Educational Measurement, 1999
Developed an empirical Bayes enhancement to Mantel-Haenszel (MH) analysis of differential item functioning (DIF) in which it is assumed that the MH statistics are normally distributed and that the prior distribution of underlying DIF parameters is also normal. (Author/SLD)
Descriptors: Bayesian Statistics, Item Bias, Statistical Distributions, Test Items
Peer reviewed Peer reviewed
Wollack, James A.; Cohen, Allan S. – Applied Psychological Measurement, 1998
Investigated empirical Type I error rates and the power of omega (index of answer copying developed by J. Wollack, 1997) when item and trait (theta) parameters were unknown and estimated from datasets of 100 and 500 examinees. Type I error was unaffected by estimating item parameters, with power slightly lower for the smaller sample. (SLD)
Descriptors: Cheating, Estimation (Mathematics), Plagiarism, Sample Size
Peer reviewed Peer reviewed
Mazor, Kathleen M.; Hambleton, Ronald K.; Clauser, Brian E. – Applied Psychological Measurement, 1998
Studied whether matching on multiple test scores would reduce false-positive error rates compared to matching on a single number-correct score using simulation. False-positive error rates were reduced for most datasets. Findings suggest that assessing the dimensional structure of a test can be important in analysis of differential item functioning…
Descriptors: Error of Measurement, Item Bias, Scores, Test Items
Peer reviewed Peer reviewed
Roussos, Louis A.; Schnipke, Deborah L.; Pashley, Peter J. – Journal of Educational and Behavioral Statistics, 1999
Derives a general formula for the population parameter being estimated by the Mantel-Haenszel differential item functioning (DIF) statistics. The formula is appropriate for uniform or nonuniform DIF and can be used regardless of the form of the item response function. Shows the relationship between this parameter and item difficulty. (SLD)
Descriptors: Difficulty Level, Estimation (Mathematics), Item Bias, Test Items
Peer reviewed Peer reviewed
Becker, Kirk – Popular Measurement, 2000
Describes the steps for estimating a population mean from three test items from a test that has been calibrated. Because the Rasch model allows for a direct transformation between raw scores and logit measures, if a population mean in logits is known relative to a set of item calibrations, the population mean in raw score units can be determined.…
Descriptors: Estimation (Mathematics), Raw Scores, Test Construction, Test Items
Peer reviewed Peer reviewed
van der Linden, Wim J. – Psychometrika, 1998
Dichotomous item response theory (IRT) models can be viewed as families of stochastically ordered distributions of responses to test items. This paper explores several properties of such distributions, especially those related to transfer to other distributions. Results are formulated as a series of theorems and corollaries that apply to…
Descriptors: Item Response Theory, Responses, Statistical Distributions, Test Items
Peer reviewed Peer reviewed
Plake, Barbara S.; Impara, James C.; Irwin, Patrick M. – Journal of Educational Measurement, 2000
Examined intra- and inter-rater consistency of item performance estimated from an Angoff standard setting over 2 years, with 29 panelists one year, and 30 the next. Results provide evidence that item performance estimates were consistent within and across panels within and across years. Factors that might have influenced this high degree of…
Descriptors: Evaluators, Prediction, Reliability, Standard Setting
Peer reviewed Peer reviewed
Magazine, Sherry L.; And Others – Educational and Psychological Measurement, 1996
Examination of the Affective and Continuance Commitment Scales of J. P. Meyer and N. J. Allen using confirmatory factor analysis for 333 subjects with the LISREL 7 computer program provided strong support across multiple diagnostics for existence of a reverse coding factor defined by the 6 negatively worded scale items. (Author/SLD)
Descriptors: Affective Behavior, Behavior Rating Scales, Coding, Test Items
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy; Lewis, Charles – Journal of Educational and Behavioral Statistics, 2000
Studied a method for flagging differential item functioning (DIF) based on loss functions. Builds on earlier research that led to the development of an empirical Bayes enhancement to the Mantel-Haenszel DIF analysis. Tested the method through simulation and found its performance better than some commonly used DIF classification systems. (SLD)
Descriptors: Bayesian Statistics, Identification, Item Bias, Simulation
Peer reviewed Peer reviewed
Muniz, Jose; Hambleton, Ronald K.; Xing, Dehui – International Journal of Testing, 2001
Studied two procedures for detecting potentially flawed items in translated tests with small samples: (1) conditional item "p" value comparisons; and (2) delta plots. Varied several factors in this simulation study. Findings show that the two procedures can be valuable in identifying flawed test items, especially when the size of the…
Descriptors: Identification, Sample Size, Simulation, Test Items
Peer reviewed Peer reviewed
Flowers, Claudia P.; Oshima, T. C.; Raju, Nambury S. – Applied Psychological Measurement, 1999
Examined the polytomous differential functioning of items and tests (DFIT) framework proposed by N. Raju and others through simulation. Findings show that the DFIT framework is effective in identifying differential item functioning and differential test functioning. (SLD)
Descriptors: Identification, Item Bias, Models, Test Bias
Peer reviewed Peer reviewed
Katz, Stuart; Lautenschlager, Gary J. – Educational Assessment, 2001
Conducted a regression analysis to assess the contributions of passage and no-passage factors to item variance on the Scholastic Aptitude Test reading comprehension task. Results show that no-passage factors play a larger role than do passage factors, accounting for as much as three-fourths of systematic variance in item difficulty and more than…
Descriptors: Reading Comprehension, Reading Tests, Regression (Statistics), Test Items
Pages: 1  |  ...  |  317  |  318  |  319  |  320  |  321  |  322  |  323  |  324  |  325  |  ...  |  636