Publication Date
In 2025 | 8 |
Since 2024 | 22 |
Since 2021 (last 5 years) | 54 |
Since 2016 (last 10 years) | 92 |
Since 2006 (last 20 years) | 301 |
Descriptor
Evaluation Methods | 468 |
Item Analysis | 468 |
Test Items | 104 |
Test Construction | 96 |
Foreign Countries | 83 |
Psychometrics | 78 |
Item Response Theory | 75 |
Test Validity | 75 |
Measurement Techniques | 71 |
Evaluation Research | 70 |
Student Evaluation | 59 |
More ▼ |
Source
Author
Alonzo, Julie | 8 |
Tindal, Gerald | 8 |
Lai, Cheng Fei | 7 |
Hambleton, Ronald K. | 4 |
Raykov, Tenko | 4 |
Chun Wang | 3 |
Gongjun Xu | 3 |
Brennan, Robert L. | 2 |
Dancer, L. Suzanne | 2 |
De Maeyer, Sven | 2 |
Gierl, Mark J. | 2 |
More ▼ |
Publication Type
Education Level
Location
Australia | 11 |
United Kingdom | 10 |
Oregon | 8 |
United States | 8 |
China | 7 |
United Kingdom (England) | 6 |
Netherlands | 5 |
Canada | 4 |
California | 3 |
Germany | 3 |
Greece | 3 |
More ▼ |
Laws, Policies, & Programs
Individuals with Disabilities… | 8 |
Elementary and Secondary… | 1 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Martin Bäckström; Fredrik Björklund – Educational and Psychological Measurement, 2024
The forced-choice response format is often considered superior to the standard Likert-type format for controlling social desirability in personality inventories. We performed simulations and found that the trait information based on the two formats converges when the number of items is high and forced-choice items are mixed with regard to…
Descriptors: Likert Scales, Item Analysis, Personality Traits, Personality Measures
Kazuhiro Yamaguchi – Journal of Educational and Behavioral Statistics, 2025
This study proposes a Bayesian method for diagnostic classification models (DCMs) for a partially known Q-matrix setting between exploratory and confirmatory DCMs. This Q-matrix setting is practical and useful because test experts have pre-knowledge of the Q-matrix but cannot readily specify it completely. The proposed method employs priors for…
Descriptors: Models, Classification, Bayesian Statistics, Evaluation Methods
Jiaying Xiao; Chun Wang; Gongjun Xu – Grantee Submission, 2024
Accurate item parameters and standard errors (SEs) are crucial for many multidimensional item response theory (MIRT) applications. A recent study proposed the Gaussian Variational Expectation Maximization (GVEM) algorithm to improve computational efficiency and estimation accuracy (Cho et al., 2021). However, the SE estimation procedure has yet to…
Descriptors: Error of Measurement, Models, Evaluation Methods, Item Analysis
Zachary K. Collier; Minji Kong; Olushola Soyoye; Kamal Chawla; Ann M. Aviles; Yasser Payne – Journal of Educational and Behavioral Statistics, 2024
Asymmetric Likert-type items in research studies can present several challenges in data analysis, particularly concerning missing data. These items are often characterized by a skewed scaling, where either there is no neutral response option or an unequal number of possible positive and negative responses. The use of conventional techniques, such…
Descriptors: Likert Scales, Test Items, Item Analysis, Evaluation Methods
Wind, Stefanie A.; Ge, Yuan – Measurement: Interdisciplinary Research and Perspectives, 2023
In selected-response assessments such as attitude surveys with Likert-type rating scales, examinees often select from rating scale categories to reflect their locations on a construct. Researchers have observed that some examinees exhibit "response styles," which are systematic patterns of responses in which examinees are more likely to…
Descriptors: Goodness of Fit, Responses, Likert Scales, Models
Klauth, Bo – ProQuest LLC, 2023
In conducting confirmatory factor analysis with ordered response items, the literature suggests that when the number of responses is five and item skewness (IS) is approximately normal, researchers can employ maximum likelihood with robust standard errors (MLR). However, MLR can yield biased factor loadings (FL) and FL standard errors (FLSE) when…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Error of Measurement
Erik Forsberg; Anders Sjöberg – Measurement: Interdisciplinary Research and Perspectives, 2025
This paper reports a validation study based on descriptive multidimensional item response theory (DMIRT), implemented in the R package "D3mirt" by using the ERS-C, an extended version of the Relevance subscale from the Moral Foundations Questionnaire including two new items for collectivism (17 items in total). Two latent models are…
Descriptors: Evaluation Methods, Programming Languages, Altruism, Collectivism
Hannah Gadd Ardrey – ProQuest LLC, 2024
The purpose of the study was to investigate secondary choral music educators' and administrators' perceptions of the use of the Mississippi Professional Growth System (PGS) as an applicable tool for evaluating secondary choral music educators. While there is limited research regarding the evaluation of choral music educators, this study aimed to…
Descriptors: Secondary School Teachers, Music Teachers, Singing, Teacher Evaluation
Liang, Xinya; Cao, Chunhua – Journal of Experimental Education, 2023
To evaluate multidimensional factor structure, a popular method that combines features of confirmatory and exploratory factor analysis is Bayesian structural equation modeling with small-variance normal priors (BSEM-N). This simulation study evaluated BSEM-N as a variable selection and parameter estimation tool in factor analysis with sparse…
Descriptors: Factor Analysis, Bayesian Statistics, Structural Equation Models, Simulation
Stephen Humphry; Paul Montuoro; Carolyn Maxwell – Journal of Psychoeducational Assessment, 2024
This article builds upon a proiminent definition of construct validity that focuses on variation in attributes causing variation in measurement outcomes. This article synthesizes the defintion and uses Rasch measurement modeling to explicate a modified conceptualization of construct validity for assessments of developmental attributes. If…
Descriptors: Construct Validity, Measurement Techniques, Developmental Stages, Item Analysis
Yang Du; Susu Zhang – Journal of Educational and Behavioral Statistics, 2025
Item compromise has long posed challenges in educational measurement, jeopardizing both test validity and test security of continuous tests. Detecting compromised items is therefore crucial to address this concern. The present literature on compromised item detection reveals two notable gaps: First, the majority of existing methods are based upon…
Descriptors: Item Response Theory, Item Analysis, Bayesian Statistics, Educational Assessment
Markus T. Jansen; Ralf Schulze – Educational and Psychological Measurement, 2024
Thurstonian forced-choice modeling is considered to be a powerful new tool to estimate item and person parameters while simultaneously testing the model fit. This assessment approach is associated with the aim of reducing faking and other response tendencies that plague traditional self-report trait assessments. As a result of major recent…
Descriptors: Factor Analysis, Models, Item Analysis, Evaluation Methods
Guo, Wenjing; Choi, Youn-Jeng – Educational and Psychological Measurement, 2023
Determining the number of dimensions is extremely important in applying item response theory (IRT) models to data. Traditional and revised parallel analyses have been proposed within the factor analysis framework, and both have shown some promise in assessing dimensionality. However, their performance in the IRT framework has not been…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Guidelines
Novak, Josip; Rebernjak, Blaž – Measurement: Interdisciplinary Research and Perspectives, 2023
A Monte Carlo simulation study was conducted to examine the performance of [alpha], [lambda]2, [lambda][subscript 4], [lambda][subscript 2], [omega][subscript T], GLB[subscript MRFA], and GLB[subscript Algebraic] coefficients. Population reliability, distribution shape, sample size, test length, and number of response categories were varied…
Descriptors: Monte Carlo Methods, Evaluation Methods, Reliability, Simulation