Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 12 |
Descriptor
Simulation | 22 |
Test Items | 18 |
Adaptive Testing | 14 |
Computer Assisted Testing | 14 |
Selection | 9 |
Item Response Theory | 8 |
Item Banks | 7 |
Comparative Analysis | 5 |
Ability | 4 |
Accuracy | 4 |
Test Construction | 3 |
More ▼ |
Source
Applied Psychological… | 5 |
Educational and Psychological… | 3 |
Journal of Educational… | 3 |
ETS Research Report Series | 2 |
Journal of Educational and… | 2 |
Applied Measurement in… | 1 |
Psychometrika | 1 |
Author
Chang, Hua-Hua | 22 |
Hau, Kit-Tai | 4 |
Leung, Chi-Keung | 2 |
Wang, Chun | 2 |
Wen, Jian-Bing | 2 |
Yi, Qing | 2 |
Ying, Zhiliang | 2 |
Zhang, Jinming | 2 |
Ali, Usama S. | 1 |
Boughton, Keith A. | 1 |
Chen, Pei-Hua | 1 |
More ▼ |
Publication Type
Journal Articles | 17 |
Reports - Research | 15 |
Reports - Evaluative | 6 |
Speeches/Meeting Papers | 5 |
Reports - Descriptive | 1 |
Education Level
Audience
Location
Canada | 1 |
United States | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 1 |
What Works Clearinghouse Rating
Choe, Edison M.; Kern, Justin L.; Chang, Hua-Hua – Journal of Educational and Behavioral Statistics, 2018
Despite common operationalization, measurement efficiency of computerized adaptive testing should not only be assessed in terms of the number of items administered but also the time it takes to complete the test. To this end, a recent study introduced a novel item selection criterion that maximizes Fisher information per unit of expected response…
Descriptors: Computer Assisted Testing, Reaction Time, Item Response Theory, Test Items
Kang, Hyeon-Ah; Lu, Ying; Chang, Hua-Hua – Applied Measurement in Education, 2017
Increasing use of item pools in large-scale educational assessments calls for an appropriate scaling procedure to achieve a common metric among field-tested items. The present study examines scaling procedures for developing a new item pool under a spiraled block linking design. The three scaling procedures are considered: (a) concurrent…
Descriptors: Item Response Theory, Accuracy, Educational Assessment, Test Items
Guo, Rui; Zheng, Yi; Chang, Hua-Hua – Journal of Educational Measurement, 2015
An important assumption of item response theory is item parameter invariance. Sometimes, however, item parameters are not invariant across different test administrations due to factors other than sampling error; this phenomenon is termed item parameter drift. Several methods have been developed to detect drifted items. However, most of the…
Descriptors: Item Response Theory, Test Items, Evaluation Methods, Equated Scores
Wang, Chun; Chang, Hua-Hua; Boughton, Keith A. – Applied Psychological Measurement, 2013
Multidimensional computerized adaptive testing (MCAT) is able to provide a vector of ability estimates for each examinee, which could be used to provide a more informative profile of an examinee's performance. The current literature on MCAT focuses on the fixed-length tests, which can generate less accurate results for those examinees whose…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Length, Item Banks
Tao, Jian; Shi, Ning-Zhong; Chang, Hua-Hua – Journal of Educational and Behavioral Statistics, 2012
For mixed-type tests composed of both dichotomous and polytomous items, polytomous items often yield more information than dichotomous ones. To reflect the difference between the two types of items, polytomous items are usually pre-assigned with larger weights. We propose an item-weighted likelihood method to better assess examinees' ability…
Descriptors: Test Items, Weighted Scores, Maximum Likelihood Statistics, Statistical Bias
Cui, Ying; Gierl, Mark J.; Chang, Hua-Hua – Journal of Educational Measurement, 2012
This article introduces procedures for the computation and asymptotic statistical inference for classification consistency and accuracy indices specifically designed for cognitive diagnostic assessments. The new classification indices can be used as important indicators of the reliability and validity of classification results produced by…
Descriptors: Classification, Accuracy, Cognitive Tests, Diagnostic Tests
Chen, Pei-Hua; Chang, Hua-Hua; Wu, Haiyan – Educational and Psychological Measurement, 2012
Two sampling-and-classification-based procedures were developed for automated test assembly: the Cell Only and the Cell and Cube methods. A simulation study based on a 540-item bank was conducted to compare the performance of the procedures with the performance of a mixed-integer programming (MIP) method for assembling multiple parallel test…
Descriptors: Test Items, Selection, Test Construction, Item Response Theory
Ali, Usama S.; Chang, Hua-Hua – ETS Research Report Series, 2014
Adaptive testing is advantageous in that it provides more efficient ability estimates with fewer items than linear testing does. Item-driven adaptive pretesting may also offer similar advantages, and verification of such a hypothesis about item calibration was the main objective of this study. A suitability index (SI) was introduced to adaptively…
Descriptors: Adaptive Testing, Simulation, Pretests Posttests, Test Items
Wang, Chun; Chang, Hua-Hua – Psychometrika, 2011
Over the past thirty years, obtaining diagnostic information from examinees' item responses has become an increasingly important feature of educational and psychological testing. The objective can be achieved by sequentially selecting multidimensional items to fit the class of latent traits being assessed, and therefore Multidimensional…
Descriptors: Psychological Testing, Adaptive Testing, Scientific Concepts, Item Analysis
Cheng, Ying; Chang, Hua-Hua; Douglas, Jeffrey; Guo, Fanmin – Educational and Psychological Measurement, 2009
a-stratification is a method that utilizes items with small discrimination (a) parameters early in an exam and those with higher a values when more is learned about the ability parameter. It can achieve much better item usage than the maximum information criterion (MIC). To make a-stratification more practical and more widely applicable, a method…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Selection
Yi, Qing; Zhang, Jinming; Chang, Hua-Hua – Applied Psychological Measurement, 2008
Criteria had been proposed for assessing the severity of possible test security violations for computerized tests with high-stakes outcomes. However, these criteria resulted from theoretical derivations that assumed uniformly randomized item selection. This study investigated potential damage caused by organized item theft in computerized adaptive…
Descriptors: Test Items, Simulation, Item Analysis, Safety

Chang, Hua-Hua; And Others – Journal of Educational Measurement, 1996
An extension to the SIBTEST procedure of R. Shealy and W. Stout (1993) to detect differential item functioning (DIF) is proposed to handle polytomous items. Results of two simulations suggest that the modified SIBTEST performs reasonably well and sometimes can provide better control of impact-induced Type I error inflation. (SLD)
Descriptors: Comparative Analysis, Identification, Item Bias, Simulation

Chang, Hua-Hua; Qian, Jiahe; Yang, Zhiliang – Applied Psychological Measurement, 2001
Proposed a refinement, based on the stratification of items developed by D. Weiss (1973), of the computerized adaptive testing item selection procedure of H. Chang and Z. Ying (1999). Simulation studies using an item bank from the Graduate Record Examination show the benefits of the new procedure. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Simulation
Chang, Hua-Hua; And Others – 1995
Recently, R. Shealy and W. Stout (1993) proposed a procedure for detecting differential item functioning (DIF) called SIBTEST. Current versions of SIBTEST can only be used for dichotomously scored items, but this paper presents an extension to handle polytomous items. The paper presents: (1) a discussion of an appropriate definition of DIF for…
Descriptors: Evaluation Methods, Identification, Item Bias, Robustness (Statistics)
Yi, Qing; Zhang, Jinming; Chang, Hua-Hua – ETS Research Report Series, 2006
Chang and Zhang (2002, 2003) proposed several baseline criteria for assessing the severity of possible test security violations for computerized tests with high-stakes outcomes. However, these criteria were obtained from theoretical derivations that assumed uniformly randomized item selection. The current study investigated potential damage caused…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Computer Security
Previous Page | Next Page ยป
Pages: 1 | 2