NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 4,126 to 4,140 of 9,552 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hartig, Johannes; Hohler, Jana – Studies in Educational Evaluation, 2009
Multidimensional item response theory (MIRT) provides an ideal foundation for modeling performance in complex domains, taking into account multiple basic abilities simultaneously, and representing different mixtures of the abilities required for different test items. This article provides a brief overview of different MIRT models, and the…
Descriptors: Psychometrics, Test Items, Item Response Theory, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Pelucchi, Bruna; Hay, Jessica F.; Saffran, Jenny R. – Cognition, 2009
Numerous recent studies suggest that human learners, including both infants and adults, readily track sequential statistics computed between adjacent elements. One such statistic, transitional probability, is typically calculated as the likelihood that one element predicts another. However, little is known about whether listeners are sensitive to…
Descriptors: Infants, Test Items, Prediction, Probability
Peer reviewed Peer reviewed
Direct linkDirect link
Guler, Nese; Penfield, Randall D. – Journal of Educational Measurement, 2009
In this study, we investigate the logistic regression (LR), Mantel-Haenszel (MH), and Breslow-Day (BD) procedures for the simultaneous detection of both uniform and nonuniform differential item functioning (DIF). A simulation study was used to assess and compare the Type I error rate and power of a combined decision rule (CDR), which assesses DIF…
Descriptors: Test Bias, Simulation, Test Items, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Sturman, Edward D.; Cribbie, Robert A.; Flett, Gordon L. – Journal of Psychoeducational Assessment, 2009
This article presents a method for assessing the internal consistency of scales that works equally well with short and long scales, namely, the average proportional distance. The method provides information on the average distance between item scores for a particular scale. In this article, we sought to demonstrate how this relatively simple…
Descriptors: Computation, Reliability, Measures (Individuals), Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Wood, Timothy J. – Advances in Health Sciences Education, 2009
Reusing questions on an examination is a concern because test administrators do not want to unfairly aid examinees by exposing them to questions they have seen on previous examinations. The purpose of this study was to investigate the effect that prior exposure of questions has on the performance of repeat examinees. Two recent administrations of…
Descriptors: Item Response Theory, Multiple Choice Tests, Memory, Test Results
Peer reviewed Peer reviewed
Direct linkDirect link
Lopez Rivas, Gabriel E.; Stark, Stephen; Chernyshenko, Oleksandr S. – Applied Psychological Measurement, 2009
The purpose of this simulation study is to investigate the effects of anchor subtest composition on the accuracy of item response theory (IRT) likelihood ratio (LR) differential item functioning (DIF) detection (Thissen, Steinberg, & Wainer, 1988). Here, the IRT LR test was implemented with a free baseline approach wherein a baseline model was…
Descriptors: Simulation, Item Response Theory, Test Bias, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Sampson, James P., Jr.; Shy, Jonathan D.; Hartley, Sarah Lucas; Reardon, Robert C.; Peterson, Gary W. – Journal of Career Development, 2009
Students (N = 247) responded to Self-Directed Search (SDS) per the standard response format and were also instructed to record a question mark (?) for items about which they were uncertain (item response indecision [IRI]). The initial responses of the 114 participants with a (?) were then reversed and a second SDS summary code was obtained and…
Descriptors: Vocational Interests, Interest Inventories, Career Counseling, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Tasdemir, Mehmet – Journal of Instructional Psychology, 2010
This study aims at comparing the difficulty levels, discrimination powers and powers of testing achievement of multiple choice tests and true-false tests, and thus revealing the rightness or wrongness of the commonly believed hypothesis that multiple choice tests don't bear the same properties as true-false tests. The research was performed with…
Descriptors: Achievement Tests, Multiple Choice Tests, Objective Tests, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Vaughn, Brandon K.; Wang, Qiu – Educational and Psychological Measurement, 2010
A nonparametric tree classification procedure is used to detect differential item functioning for items that are dichotomously scored. Classification trees are shown to be an alternative procedure to detect differential item functioning other than the use of traditional Mantel-Haenszel and logistic regression analysis. A nonparametric…
Descriptors: Test Bias, Classification, Nonparametric Statistics, Regression (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Andrich, David; Kreiner, Svend – Applied Psychological Measurement, 2010
Models of modern test theory imply statistical independence among responses, generally referred to as "local independence." One violation of local independence occurs when the response to one item governs the response to a subsequent item. Expanding on a formulation of this kind of violation as a process in the dichotomous Rasch model,…
Descriptors: Test Theory, Item Response Theory, Test Items, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Jiyoung; Chi, Youngshin; Huensch, Amanda; Jun, Heesung; Li, Hongli; Roullion, Vanessa – Language Assessment Quarterly, 2010
This article discusses a case study on an item writing process that reflects on our practical experience in an item development project. The purpose of the article is to share our lessons from the experience aiming to demystify item writing process. The study investigated three issues that naturally emerged during the project: how item writers use…
Descriptors: Test Items, Writing Processes, Group Dynamics, Case Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Battisti, Bryce Thomas; Hanegan, Nikki; Sudweeks, Richard; Cates, Rex – International Journal of Science and Mathematics Education, 2010
Concept inventories are often used to assess current student understanding although conceptual change models are problematic. Due to controversies with conceptual change models and the realities of student assessment, it is important that concept inventories are evaluated using a variety of theoretical models to improve quality. This study used a…
Descriptors: Item Response Theory, Testing, Test Items, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Do-Hong; Huynh, Huynh – Educational Assessment, 2010
This study investigated whether scores obtained from the online and paper-and-pencil administrations of the statewide end-of-course English test were equivalent for students with and without disabilities. Score comparability was evaluated by examining equivalence of factor structure (measurement invariance) and differential item and bundle…
Descriptors: Computer Assisted Testing, Language Tests, English, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Acar, Tulin; Kelecioglu, Hulya – Educational Sciences: Theory and Practice, 2010
The aim of this research is to determine differential item functioning (DIF) by hierarchical linear modeling (HGLM) on test items and comparing these results by the DIF results determined by logistic regression (LR) and IRT-likelihood ratio (IRT-LR) techniques. Investigating the concordance between the techniques in determining the items with DIF,…
Descriptors: Test Bias, Comparative Analysis, Item Response Theory, Regression (Statistics)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Guler, Nese; Gelbal, Selahattin – Educational Sciences: Theory and Practice, 2010
In this study, the Classical test theory and generalizability theory were used for determination to reliability of scores obtained from measurement tool of mathematics success. 24 open-ended mathematics question of the TIMSS-1999 was applied to 203 students in 2007-spring semester. Internal consistency of scores was found as 0.92. For…
Descriptors: Generalizability Theory, Test Theory, Test Reliability, Interrater Reliability
Pages: 1  |  ...  |  272  |  273  |  274  |  275  |  276  |  277  |  278  |  279  |  280  |  ...  |  637