NotesFAQContact Us
Collection
Advanced
Search Tips
Source
International Journal of…72
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 72 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Xiaowen Liu – International Journal of Testing, 2024
Differential item functioning (DIF) often arises from multiple sources. Within the context of multidimensional item response theory, this study examined DIF items with varying secondary dimensions using the three DIF methods: SIBTEST, Mantel-Haenszel, and logistic regression. The effect of the number of secondary dimensions on DIF detection rates…
Descriptors: Item Analysis, Test Items, Item Response Theory, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Magarotto Machado, Gisele; Hauck-Filho, Nelson; Lima-Costa, Ariela Raissa; Carvalho, Lucas de Francisco – International Journal of Testing, 2023
In the current study, we used latent profile analysis to investigate the Dimensional Clinical Personality Inventory 2 capacity to discriminate psychopathy traits in a sample of adults. Participants were 628 adults from the general population recruited by convenience. Our latent profile analysis recovered two groups: Psychopathic Tendencies and…
Descriptors: Adults, Psychopathology, Empathy, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Badham, Louise; Furlong, Antony – International Journal of Testing, 2023
Multilingual summative assessments face significant challenges due to tensions that exist between multiple language provision and comparability. Yet, conventional approaches for investigating comparability in multilingual assessments fail to accommodate assessments that comprise extended responses that target complex constructs. This article…
Descriptors: Summative Evaluation, Multilingualism, Comparative Analysis, Literature
Peer reviewed Peer reviewed
Direct linkDirect link
Maritza Casas; Stephen G. Sireci – International Journal of Testing, 2025
In this study, we take a critical look at the degree to which the measurement of bullying and sense of belonging at school is invariant across groups of students defined by immigrant status. Our study focuses on the invariance of these constructs as measured on a recent PISA administration and includes a discussion of two statistical methods for…
Descriptors: Error of Measurement, Immigrants, Peer Groups, Bullying
Peer reviewed Peer reviewed
Direct linkDirect link
Chet Robie; Sabah Rasheed; Stephen D. Risavy; Piers Steel – International Journal of Testing, 2024
This meta-analysis examined the validity of an alternative to traditional assessments called the Wonderlic which is a brief measure of general mental ability. Our results showed significant, positive correlations between Wonderlic scores and academic performance in general ([r-bar] = 0.26), between Wonderlic scores and undergraduate GPA in…
Descriptors: Meta Analysis, Test Validity, Alternative Assessment, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Rujun Xu; James Soland – International Journal of Testing, 2024
International surveys are increasingly being used to understand nonacademic outcomes like math and science motivation, and to inform education policy changes within countries. Such instruments assume that the measure works consistently across countries, ethnicities, and languages--that is, they assume measurement invariance. While studies have…
Descriptors: Surveys, Statistical Bias, Achievement Tests, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Rios, Joseph A.; Guo, Hongwen; Mao, Liyang; Liu, Ou Lydia – International Journal of Testing, 2017
When examinees' test-taking motivation is questionable, practitioners must determine whether careless responding is of practical concern and if so, decide on the best approach to filter such responses. As there has been insufficient research on these topics, the objectives of this study were to: a) evaluate the degree of underestimation in the…
Descriptors: Response Style (Tests), Scores, Motivation, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Cao, Mengyang; Song, Q. Chelsea; Tay, Louis – International Journal of Testing, 2018
There is a growing use of noncognitive assessments around the world, and recent research has posited an ideal point response process underlying such measures. A critical issue is whether the typical use of dominance approaches (e.g., average scores, factor analysis, and the Samejima's graded response model) in scoring such measures is adequate.…
Descriptors: Comparative Analysis, Item Response Theory, Factor Analysis, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Arce-Ferrer, Alvaro J.; Bulut, Okan – International Journal of Testing, 2017
This study examines separate and concurrent approaches to combine the detection of item parameter drift (IPD) and the estimation of scale transformation coefficients in the context of the common item nonequivalent groups design with the three-parameter item response theory equating. The study uses real and synthetic data sets to compare the two…
Descriptors: Item Response Theory, Equated Scores, Identification, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Roschmann, Sarina; Witmer, Sara E.; Volker, Martin A. – International Journal of Testing, 2021
Accommodations are commonly provided to address language-related barriers students may experience during testing. Research on the validity of scores from accommodated test administrations remains somewhat inconclusive. The current study investigated item response patterns to understand whether accommodations, as used in practice among English…
Descriptors: Testing Accommodations, English Language Learners, Scores, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Martin-Raugh, Michelle P.; Anguiano-Carrsaco, Cristina; Jackson, Teresa; Brenneman, Meghan W.; Carney, Lauren; Barnwell, Patrick; Kochert, Jonathan – International Journal of Testing, 2018
Single-response situational judgment tests (SRSJTs) differ from multiple-response SJTs (MRSJTS) in that they present test takers with edited critical incidents and simply ask test takers to read over the action described and evaluate it according to its effectiveness. Research comparing the reliability and validity of SRSJTs and MRSJTs is thus far…
Descriptors: Test Format, Test Reliability, Test Validity, Predictive Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Guo, Xiuyan; Lei, Pui-Wa – International Journal of Testing, 2020
Little research has been done on the effects of peer raters' quality characteristics on peer rating qualities. This study aims to address this gap and investigate the effects of key variables related to peer raters' qualities, including content knowledge, previous rating experience, training on rating tasks, and rating motivation. In an experiment…
Descriptors: Peer Evaluation, Error Patterns, Correlation, Knowledge Level
Peer reviewed Peer reviewed
Direct linkDirect link
Evers, Arne; McCormick, Carina M.; Hawley, Leslie R.; Muñiz, José; Balboni, Giulia; Bartram, Dave; Boben, Dusica; Egeland, Jens; El-Hassan, Karma; Fernández-Hermida, José R.; Fine, Saul; Frans, Örjan; Gintiliené, Grazina; Hagemeister, Carmen; Halama, Peter; Iliescu, Dragos; Jaworowska, Aleksandra; Jiménez, Paul; Manthouli, Marina; Matesic, Krunoslav; Michaelsen, Lars; Mogaji, Andrew; Morley-Kirk, James; Rózsa, Sándor; Rowlands, Lorraine; Schittekatte, Mark; Sümer, H. Canan; Suwartono, Tono; Urbánek, Tomáš; Wechsler, Solange; Zelenevska, Tamara; Zanev, Svetoslav; Zhang, Jianxin – International Journal of Testing, 2017
On behalf of the International Test Commission and the European Federation of Psychologists' Associations a world-wide survey on the opinions of professional psychologists on testing practices was carried out. The main objective of this study was to collect data for a better understanding of the state of psychological testing worldwide. These data…
Descriptors: Testing, Attitudes, Surveys, Psychologists
Peer reviewed Peer reviewed
Direct linkDirect link
Maeda, Hotaka; Zhang, Bo – International Journal of Testing, 2017
The omega (?) statistic is reputed to be one of the best indices for detecting answer copying on multiple choice tests, but its performance relies on the accurate estimation of copier ability, which is challenging because responses from the copiers may have been contaminated. We propose an algorithm that aims to identify and delete the suspected…
Descriptors: Cheating, Test Items, Mathematics, Statistics
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5