NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 19 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Daria Gerasimova – Journal of Educational Measurement, 2024
I propose two practical advances to the argument-based approach to validity: developing a living document and incorporating preregistration. First, I present a potential structure for the living document that includes an up-to-date summary of the validity argument. As the validation process may span across multiple studies, the living document…
Descriptors: Validity, Documentation, Methods, Research Reports
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Zhongmin; Liu, Chunyan; He, Yong; Chen, Hanwei – Journal of Educational Measurement, 2018
Allowing item review in computerized adaptive testing (CAT) is getting more attention in the educational measurement field as more and more testing programs adopt CAT. The research literature has shown that allowing item review in an educational test could result in more accurate estimates of examinees' abilities. The practice of item review in…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Test Wiseness
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Hyung Jin; Brennan, Robert L.; Lee, Won-Chan – Journal of Educational Measurement, 2017
In equating, when common items are internal and scoring is conducted in terms of the number of correct items, some pairs of total scores ("X") and common-item scores ("V") can never be observed in a bivariate distribution of "X" and "V"; these pairs are called "structural zeros." This simulation…
Descriptors: Test Items, Equated Scores, Comparative Analysis, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Debeer, Dries; Ali, Usama S.; van Rijn, Peter W. – Journal of Educational Measurement, 2017
Test assembly is the process of selecting items from an item pool to form one or more new test forms. Often new test forms are constructed to be parallel with an existing (or an ideal) test. Within the context of item response theory, the test information function (TIF) or the test characteristic curve (TCC) are commonly used as statistical…
Descriptors: Test Format, Test Construction, Statistical Analysis, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
French, Brian F.; Finch, W. Holmes – Journal of Educational Measurement, 2015
SIBTEST is a differential item functioning (DIF) detection method that is accurate and effective with small samples, in the presence of group mean differences, and for assessment of both uniform and nonuniform DIF. The presence of multilevel data with DIF detection has received increased attention. Ignoring such structure can inflate Type I error.…
Descriptors: Test Bias, Data, Simulation, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Haberman, Shelby; Yao, Lili – Journal of Educational Measurement, 2015
Admission decisions frequently rely on multiple assessments. As a consequence, it is important to explore rational approaches to combine the information from different educational tests. For example, U.S. graduate schools usually receive both TOEFL iBT® scores and GRE® General scores of foreign applicants for admission; however, little guidance…
Descriptors: College Entrance Examinations, Repetition, Methods, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Häggström, Jenny; Wiberg, Marie – Journal of Educational Measurement, 2014
The selection of bandwidth in kernel equating is important because it has a direct impact on the equated test scores. The aim of this article is to examine the use of double smoothing when selecting bandwidths in kernel equating and to compare double smoothing with the commonly used penalty method. This comparison was made using both an equivalent…
Descriptors: Equated Scores, Data Analysis, Comparative Analysis, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Powers, Sonya; Kolen, Michael J. – Journal of Educational Measurement, 2014
Accurate equating results are essential when comparing examinee scores across exam forms. Previous research indicates that equating results may not be accurate when group differences are large. This study compared the equating results of frequency estimation, chained equipercentile, item response theory (IRT) true-score, and IRT observed-score…
Descriptors: Accuracy, Equated Scores, Differences, Groups
Peer reviewed Peer reviewed
Direct linkDirect link
Yao, Lihua – Journal of Educational Measurement, 2014
The intent of this research was to find an item selection procedure in the multidimensional computer adaptive testing (CAT) framework that yielded higher precision for both the domain and composite abilities, had a higher usage of the item pool, and controlled the exposure rate. Five multidimensional CAT item selection procedures (minimum angle;…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Items, Selection
Peer reviewed Peer reviewed
Direct linkDirect link
Rutkowski, Leslie; Zhou, Yan – Journal of Educational Measurement, 2015
Given the importance of large-scale assessments to educational policy conversations, it is critical that subpopulation achievement is estimated reliably and with sufficient precision. Despite this importance, biased subpopulation estimates have been found to occur when variables in the conditioning model side of a latent regression model contain…
Descriptors: Error of Measurement, Error Correction, Regression (Statistics), Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Deng, Hui; Ansley, Timothy; Chang, Hua-Hua – Journal of Educational Measurement, 2010
In this study we evaluated and compared three item selection procedures: the maximum Fisher information procedure (F), the a-stratified multistage computer adaptive testing (CAT) (STR), and a refined stratification procedure that allows more items to be selected from the high a strata and fewer items from the low a strata (USTR), along with…
Descriptors: Computer Assisted Testing, Adaptive Testing, Selection, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Omar, M. Hafidz – Journal of Educational Measurement, 2010
Methods of statistical process control were briefly investigated in the field of educational measurement as early as 1999. However, only the use of a cumulative sum chart was explored. In this article other methods of statistical quality control are introduced and explored. In particular, methods in the form of Shewhart mean and standard deviation…
Descriptors: Charts, Quality Control, Measurement, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Zhongmin; Kolen, Michael J. – Journal of Educational Measurement, 2009
This article considers two new smoothing methods in equipercentile equating, the cubic B-spline presmoothing method and the direct presmoothing method. Using a simulation study, these two methods are compared with established methods, the beta-4 method, the polynomial loglinear method, and the cubic spline postsmoothing method, under three sample…
Descriptors: Equated Scores, Methods, Sample Size, Test Content
Peer reviewed Peer reviewed
Direct linkDirect link
de La Torre, Jimmy; Deng, Weiling – Journal of Educational Measurement, 2008
The standardized log-likelihood of a response vector (l[subscript z]) is a popular IRT-based person-fit test statistic for identifying model-misfitting response patterns. Traditional use of l[subscript z] is overly conservative in detecting aberrance due to its incorrect assumption regarding its theoretical null distribution. This study proposes a…
Descriptors: Goodness of Fit, Measures (Individuals), Test Reliability, Responses
Peer reviewed Peer reviewed
Direct linkDirect link
Davis, Susan L.; Buckendahl, Chad W.; Plake, Barbara S. – Journal of Educational Measurement, 2008
As an alternative to adaptation, tests may also be developed simultaneously in multiple languages. Although the items on such tests could vary substantially, scores from these tests may be used to make the same types of decisions about different groups of examinees. The ability to make such decisions is contingent upon setting performance…
Descriptors: Test Results, Testing Programs, Multilingualism, Standard Setting
Previous Page | Next Page »
Pages: 1  |  2