NotesFAQContact Us
Collection
Advanced
Search Tips
Source
International Journal of…14
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 14 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Xiaowen Liu – International Journal of Testing, 2024
Differential item functioning (DIF) often arises from multiple sources. Within the context of multidimensional item response theory, this study examined DIF items with varying secondary dimensions using the three DIF methods: SIBTEST, Mantel-Haenszel, and logistic regression. The effect of the number of secondary dimensions on DIF detection rates…
Descriptors: Item Analysis, Test Items, Item Response Theory, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Cao, Mengyang; Song, Q. Chelsea; Tay, Louis – International Journal of Testing, 2018
There is a growing use of noncognitive assessments around the world, and recent research has posited an ideal point response process underlying such measures. A critical issue is whether the typical use of dominance approaches (e.g., average scores, factor analysis, and the Samejima's graded response model) in scoring such measures is adequate.…
Descriptors: Comparative Analysis, Item Response Theory, Factor Analysis, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Magis, David; Raiche, Gilles; Beland, Sebastien; Gerard, Paul – International Journal of Testing, 2011
We present an extension of the logistic regression procedure to identify dichotomous differential item functioning (DIF) in the presence of more than two groups of respondents. Starting from the usual framework of a single focal group, we propose a general approach to estimate the item response functions in each group and to test for the presence…
Descriptors: Language Skills, Identification, Foreign Countries, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Young-Sun; Park, Yoon Soo; Taylan, Didem – International Journal of Testing, 2011
Studies of international mathematics achievement such as the Trends in Mathematics and Science Study (TIMSS) have employed classical test theory and item response theory to rank individuals within a latent ability continuum. Although these approaches have provided insights into comparisons between countries, they have yet to examine how specific…
Descriptors: Mathematics Achievement, Achievement Tests, Models, Cognitive Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Evers, Arne – International Journal of Testing, 2012
In this article, the characteristics of five test review models are described. The five models are the US review system at the Buros Center for Testing, the German Test Review System of the Committee on Tests, the Brazilian System for the Evaluation of Psychological Tests, the European EFPA Review Model, and the Dutch COTAN Evaluation System for…
Descriptors: Program Evaluation, Test Reviews, Trend Analysis, International Education
Peer reviewed Peer reviewed
Direct linkDirect link
Byrne, Barbara M.; van de Vijver, Fons J. R. – International Journal of Testing, 2010
A critical assumption in cross-cultural comparative research is that the instrument measures the same construct(s) in exactly the same way across all groups (i.e., the instrument is measurement and structurally equivalent). Structural equation modeling (SEM) procedures are commonly used in testing these assumptions of multigroup equivalence.…
Descriptors: Measures (Individuals), Cross Cultural Studies, Measurement, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Svetina, Dubravka; Gorin, Joanna S.; Tatsuoka, Kikumi K. – International Journal of Testing, 2011
As a construct definition, the current study develops a cognitive model describing the knowledge, skills, and abilities measured by critical reading test items on a high-stakes assessment used for selection decisions in the United States. Additionally, in order to establish generalizability of the construct meaning to other similarly structured…
Descriptors: Reading Tests, Reading Comprehension, Critical Reading, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Furlan, Luis Alberto; Cassady, Jerrell C.; Perez, Edgardo Raul – International Journal of Testing, 2009
A new Spanish version of the Cognitive Test Anxiety Scale (CTAS) was created to be used explicitly with Argentinean university students. The scale was translated and verified through blind back translation and given to a large sample of students majoring in psychology or chemistry (N = 752). Exploratory Factor Analysis (N = 376) showed an internal…
Descriptors: Factor Structure, Cognitive Tests, Measures (Individuals), Factor Analysis
Peer reviewed Peer reviewed
Byrne, Barbara M. – International Journal of Testing, 2001
Uses a confirmatory factor analytic (CFA) model as a paradigmatic basis for the comparison of three widely used structural equation modeling computer programs: (1) AMOS 4.0; (2) EQS 6; and (3) LISREL 8. Comparisons focus on aspects of programs that bear on the specification and testing of CFA models and the treatment of incomplete, nonnormally…
Descriptors: Comparative Analysis, Computer Software, Data Analysis, Statistical Distributions
Peer reviewed Peer reviewed
Direct linkDirect link
Sese, Albert; Palmer, Alfonso L.; Montano, Juan J. – International Journal of Testing, 2004
The study of measurement models in psychometrics by means of dimensionality reduction techniques such as Principal Components Analysis (PCA) is a very common practice. In recent times, an upsurge of interest in the study of artificial neural networks apt to computing a principal component extraction has been observed. Despite this interest, the…
Descriptors: Psychometrics, Computer Simulation, Models, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Ramesh, Anuradha; Hazucha, Joy F.; Bank, Jurgen – International Journal of Testing, 2008
A major challenge that decisions makers face in multi-national organizations is how to compare managers from different parts of the globe. This challenge is both psychometric and practical. We draw on the cross-cultural psychology literature to propose a three-step framework to compare personality data from different countries. The first step…
Descriptors: Personality, Norms, Psychometrics, International Organizations
Peer reviewed Peer reviewed
Direct linkDirect link
Meyer, Kevin D.; Foster, Jeff L. – International Journal of Testing, 2008
With the increasing globalization of human resources practices, a commensurate increase in demand has occurred for multi-language ("global") personality norms for use in selection and development efforts. The combination of data from multiple translations of a personality assessment into a single norm engenders error from multiple sources. This…
Descriptors: Global Approach, Cultural Differences, Norms, Human Resources