NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Type
Reports - Evaluative27
Journal Articles26
Information Analyses1
Speeches/Meeting Papers1
Education Level
Higher Education1
Audience
Researchers1
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Showing 1 to 15 of 27 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Sijia Huang; Dubravka Svetina Valdivia – Educational and Psychological Measurement, 2024
Identifying items with differential item functioning (DIF) in an assessment is a crucial step for achieving equitable measurement. One critical issue that has not been fully addressed with existing studies is how DIF items can be detected when data are multilevel. In the present study, we introduced a Lord's Wald X[superscript 2] test-based…
Descriptors: Item Analysis, Item Response Theory, Algorithms, Accuracy
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Eray Selçuk; Ergül Demir – International Journal of Assessment Tools in Education, 2024
This research aims to compare the ability and item parameter estimations of Item Response Theory according to Maximum likelihood and Bayesian approaches in different Monte Carlo simulation conditions. For this purpose, depending on the changes in the priori distribution type, sample size, test length, and logistics model, the ability and item…
Descriptors: Item Response Theory, Item Analysis, Test Items, Simulation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Williams, Matt N.; Gomez Grajales, Carlos Alberto; Kurkiewicz, Dason – Practical Assessment, Research & Evaluation, 2013
In 2002, an article entitled "Four assumptions of multiple regression that researchers should always test" by Osborne and Waters was published in "PARE." This article has gone on to be viewed more than 275,000 times (as of August 2013), and it is one of the first results displayed in a Google search for "regression…
Descriptors: Multiple Regression Analysis, Misconceptions, Reader Response, Predictor Variables
Peer reviewed Peer reviewed
Direct linkDirect link
Murphy, Daniel L.; Dodd, Barbara G.; Vaughn, Brandon K. – Applied Psychological Measurement, 2010
This study examined the performance of the maximum Fisher's information, the maximum posterior weighted information, and the minimum expected posterior variance methods for selecting items in a computerized adaptive testing system when the items were grouped in testlets. A simulation study compared the efficiency of ability estimation among the…
Descriptors: Simulation, Adaptive Testing, Item Analysis, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Barrada, Juan Ramon; Olea, Julio; Ponsoda, Vicente; Abad, Francisco Jose – Applied Psychological Measurement, 2010
In a typical study comparing the relative efficiency of two item selection rules in computerized adaptive testing, the common result is that they simultaneously differ in accuracy and security, making it difficult to reach a conclusion on which is the more appropriate rule. This study proposes a strategy to conduct a global comparison of two or…
Descriptors: Test Items, Simulation, Adaptive Testing, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Finkelman, Matthew; Nering, Michael L.; Roussos, Louis A. – Journal of Educational Measurement, 2009
In computerized adaptive testing (CAT), ensuring the security of test items is a crucial practical consideration. A common approach to reducing item theft is to define maximum item exposure rates, i.e., to limit the proportion of examinees to whom a given item can be administered. Numerous methods for controlling exposure rates have been proposed…
Descriptors: Test Items, Adaptive Testing, Item Analysis, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Cools, Wilfried; De Fraine, Bieke; Van den Noortgate, Wim; Onghena, Patrick – School Effectiveness and School Improvement, 2009
In educational effectiveness research, multilevel data analyses are often used because research units (most frequently, pupils or teachers) are studied that are nested in groups (schools and classes). This hierarchical data structure complicates designing the study because the structure has to be taken into account when approximating the accuracy…
Descriptors: Effective Schools Research, Program Effectiveness, School Effectiveness, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Yi, Qing; Zhang, Jinming; Chang, Hua-Hua – Applied Psychological Measurement, 2008
Criteria had been proposed for assessing the severity of possible test security violations for computerized tests with high-stakes outcomes. However, these criteria resulted from theoretical derivations that assumed uniformly randomized item selection. This study investigated potential damage caused by organized item theft in computerized adaptive…
Descriptors: Test Items, Simulation, Item Analysis, Safety
Peer reviewed Peer reviewed
Direct linkDirect link
Boulet, John R.; van Zanten, Marta; de Champlain, Andre; Hawkins, Richard E.; Peitzman, Steven J. – Advances in Health Sciences Education, 2008
While checklists are often used to score standardized patient based clinical assessments, little research has focused on issues related to their development or the level of agreement with respect to the importance of specific items. Five physicians independently reviewed checklists from 11 simulation scenarios that were part of the former…
Descriptors: Check Lists, Foreign Medical Graduates, Patients, Clinical Experience
Peer reviewed Peer reviewed
Direct linkDirect link
French, Brian F.; Maller, Susan J. – Educational and Psychological Measurement, 2007
Two unresolved implementation issues with logistic regression (LR) for differential item functioning (DIF) detection include ability purification and effect size use. Purification is suggested to control inaccuracies in DIF detection as a result of DIF items in the ability estimate. Additionally, effect size use may be beneficial in controlling…
Descriptors: Effect Size, Test Bias, Guidelines, Error of Measurement
Peer reviewed Peer reviewed
Ramsay, J. O. – Psychometrika, 1995
A similarity-based smoothing approach to nondimensional item analysis was studied. Simulated and actual data are presented to show that when responses are determined by a latent ability variable, this similarity-based smoothing procedure can reveal the dimensionality of ability satisfactorily. (SLD)
Descriptors: Ability, Item Analysis, Item Response Theory, Nonparametric Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Van Rijn, P. W.; Eggen, T. J. H. M.; Hemker, B. T.; Sanders, P. F. – Applied Psychological Measurement, 2002
In the present study, a procedure that has been used to select dichotomous items in computerized adaptive testing was applied to polytomous items. This procedure was designed to select the item with maximum weighted information. In a simulation study, the item information function was integrated over a fixed interval of ability values and the item…
Descriptors: Intervals, Simulation, Adaptive Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Hartig, Johannes; Holzel, Britta; Moosbrugger, Helfried – Multivariate Behavioral Research, 2007
Numerous studies have shown increasing item reliabilities as an effect of the item position in personality scales. Traditionally, these context effects are analyzed based on item-total correlations. This approach neglects that trends in item reliabilities can be caused either by an increase in true score variance or by a decrease in error…
Descriptors: True Scores, Error of Measurement, Structural Equation Models, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
van de Velden, Michel; Bijmolt, Tammo H. A. – Psychometrika, 2006
A method is presented for generalized canonical correlation analysis of two or more matrices with missing rows. The method is a combination of Carroll's (1968) method and the missing data approach of the OVERALS technique (Van der Burg, 1988). In a simulation study we assess the performance of the method and compare it to an existing procedure…
Descriptors: Multivariate Analysis, Matrices, Simulation, Comparative Testing
Peer reviewed Peer reviewed
Reckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models
Previous Page | Next Page »
Pages: 1  |  2