NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 4,876 to 4,890 of 9,547 results Save | Export
Peer reviewed Peer reviewed
French, Ann W.; Miller, Timothy R. – Journal of Educational Measurement, 1996
A computer simulation study was conducted to determine the feasibility of using logistic regression procedures to detect differential item functioning (DIF) in polytomous items. Results indicate that logistic regression is powerful in detecting most forms of DIF, although it requires large amounts of data manipulation and careful interpretation.…
Descriptors: Computer Simulation, Identification, Item Bias, Test Interpretation
Peer reviewed Peer reviewed
Pelton, Timothy W.; Bunderson, C. Victor – Journal of Applied Measurement, 2003
Attempted to illuminate practical limitations on the Rasch model by focusing on the recovery of the density scale through five simulation trials. Results show that when error distributions are insufficient, the results may be ordinal at best, and when error distributions are nonsymmetrical, the positions of items may be biased with respect to the…
Descriptors: Error of Measurement, Item Response Theory, Simulation, Test Items
Peer reviewed Peer reviewed
Scheiblechner, Hartmann – Psychometrika, 2003
Presented nonparametric tests for testing the validity of polytomous unidimensional ordinal probabilistic polytomous item response theory models along with procedures for testing the comonotonicity of two item sets and for item selection. Describes advantages of the new approach. (SLD)
Descriptors: Item Response Theory, Nonparametric Statistics, Selection, Test Items
Peer reviewed Peer reviewed
Bolt, Daniel – Psychometrika, 2003
Any item response theory (IRT) researcher or practitioner will find something of interest in this book, which covers a broad range of topics in essays by well-known researchers. Chapters are organized into sections devoted to parametric and nonparametric IRT topics. (SLD)
Descriptors: Item Response Theory, Measurement Techniques, Test Construction, Test Items
Peer reviewed Peer reviewed
Stanton, Jeffrey M.; Bachiochi, Peter D.; Robie, Chet; Perez, Lisa M.; Smith, Patricia C. – Educational and Psychological Measurement, 2002
Studied the Work Satisfaction subscale of the Job Descriptive Index (JDI) to determine the difference between measuring work stress and measuring work satisfaction. Results from samples of 1,623 and 314 adults provide evidence supporting the removal of some contaminating items from the JDI. (SLD)
Descriptors: Adults, Measures (Individuals), Stress Variables, Test Construction
Peer reviewed Peer reviewed
Davey, Tim; And Others – Applied Psychological Measurement, 1996
Scales defined by most item response theory (IRT) models are truly invariant with respect to certain linear transformations of parameters. The problem is to find the proper transformation to place calibrations on a common scale. This paper explores issues of extending and adapting unidimensional linking procedures to multidimensional IRT models.…
Descriptors: Equated Scores, Item Response Theory, Models, Scaling
Peer reviewed Peer reviewed
Zenisky, April L.; Hambleton, Ronald K.; Robin, Frederic – Educational and Psychological Measurement, 2003
Studied a two-stage methodology for evaluating differential item functioning (DIF) in large-scale assessment data using a sample of 60,000 students taking a large-scale assessment. Findings illustrate the merit of iterative approached for DIF detection, since items identified at one stage were not necessarily the same as those identified at the…
Descriptors: Item Bias, Large Scale Assessment, Research Methodology, Test Items
Peer reviewed Peer reviewed
Gelin, Michaela N.; Zumbo, Bruno D. – Educational and Psychological Measurement, 2003
Investigated potentially biased scale items on the Center for Epidemiological Studies Depression scale (CES-D; Radloff, 1977) in a sample of 600 adults. Overall, results indicate that the scoring method has an effect on differential item functioning (DIF), and that DIF is a property of the item, scoring method, and purpose of the assessment. (SLD)
Descriptors: Depression (Psychology), Item Bias, Scoring, Test Items
Peer reviewed Peer reviewed
Gierl, Mark J.; Bolt, Daniel M. – International Journal of Testing, 2001
Presents an overview of nonparametric regression as it allies to differential item functioning analysis and then provides three examples to illustrate how nonparametric regression can be applied to multilingual, multicultural data to study group differences. (SLD)
Descriptors: Groups, Item Bias, Nonparametric Statistics, Regression (Statistics)
Peer reviewed Peer reviewed
Harmon, Lenore W.; Borgen, Fred H. – Journal of Career Assessment, 1995
Data from over 50,000 people in 50 occupational groups were used to revise the Strong Interest Inventory. New General Reference Samples containing over 18,000 people were used to construct scales, and nearly every scale was revised. (SK)
Descriptors: Evaluation Criteria, Interest Inventories, Measures (Individuals), Occupations
Peer reviewed Peer reviewed
Engelhard, George, Jr.; Davis, Melodee; Hansche, Linda – Applied Measurement in Education, 1999
Examined whether reviewers on item-review committees can identify accurately test items that exhibit a variety of flaws. Results with 39 reviewers of a 75-item test show that reviewers exhibit fairly high accuracy rates overall, with statistically significant differences in judgmental accuracy among reviewers. (SLD)
Descriptors: Decision Making, Judges, Review (Reexamination), Test Construction
Peer reviewed Peer reviewed
Lee, Guemin; Frisbie, David A. – Applied Measurement in Education, 1999
Studied the appropriateness and implications of using a generalizability theory approach to estimating the reliability of scores from tests composed of testlets. Analyses of data from two national standardization samples suggest that manipulating the number of passages is a more productive way to obtain efficient measurement than manipulating the…
Descriptors: Generalizability Theory, Models, National Surveys, Reliability
Peer reviewed Peer reviewed
Nering, Michael L.; Meijer, Rob R. – Applied Psychological Measurement, 1998
Compared the person-response function (PRF) method for identifying examinees who respond to test items in a manner divergent from the underlying test model to the "l(z)" index of Drasgow and others (1985). Although performance of the "l(z)" index was superior in most cases, the PRF was useful in some conditions. (SLD)
Descriptors: Comparative Analysis, Item Response Theory, Models, Responses
Peer reviewed Peer reviewed
van der Linden, Wim J.; Adema, Jos J. – Journal of Educational Measurement, 1998
Proposes an algorithm for the assembly of multiple test forms in which the multiple-form problem is reduced to a series of computationally less intensive two-form problems. Illustrates how the method can be implemented using 0-1 linear programming and gives two examples. (SLD)
Descriptors: Algorithms, Linear Programming, Test Construction, Test Format
Peer reviewed Peer reviewed
Camilli, Gregory; Congdon, Peter – Journal of Educational and Behavioral Statistics, 1999
Demonstrates a method for studying differential item functioning (DIF) that can be used with dichotomous or polytomous items and that is valid for data that follow a partial credit Item Response Theory model. A simulation study shows that positively biased Type I error rates are in accord with results from previous studies. (SLD)
Descriptors: Estimation (Mathematics), Item Bias, Item Response Theory, Test Items
Pages: 1  |  ...  |  322  |  323  |  324  |  325  |  326  |  327  |  328  |  329  |  330  |  ...  |  637