NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 3 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Yi-Hsuan; Zhang, Jinming – International Journal of Testing, 2017
Simulations were conducted to examine the effect of differential item functioning (DIF) on measurement consequences such as total scores, item response theory (IRT) ability estimates, and test reliability in terms of the ratio of true-score variance to observed-score variance and the standard error of estimation for the IRT ability parameter. The…
Descriptors: Test Bias, Test Reliability, Performance, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Stone, Gregory Ethan; Beltyukova, Svetlana; Fox, Christine M. – International Journal of Testing, 2008
Judge-mediated examinations are defined as those for which expert evaluation (using rubrics) is required to determine correctness, completeness, and reasonability of test-taker responses. The use of multifaceted Rasch modeling has led to improvements in the reliability of scoring such examinations. The establishment of criterion-referenced…
Descriptors: Interrater Reliability, High Stakes Tests, Standard Setting, Minimum Competencies
Peer reviewed Peer reviewed
Hsu, Tse-chi; Wu, Kuo-liang; Yu, Jya-yi Wu; Lee, Ming-yen – International Journal of Testing, 2002
Explored the feasibility of applying a method that incorporates collateral information to equate tests constructed for a college entrance examination by comparing its results with those of item response theory (IRT) true score equating. Simulation results suggest that, overall, equating results based on collateral information are relatively…
Descriptors: College Entrance Examinations, Equated Scores, Item Response Theory, Simulation