NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers1
Location
China2
Turkey1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 23 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Kang, Hyeon-Ah; Han, Suhwa; Kim, Doyoung; Kao, Shu-Chuan – Educational and Psychological Measurement, 2022
The development of technology-enhanced innovative items calls for practical models that can describe polytomous testlet items. In this study, we evaluate four measurement models that can characterize polytomous items administered in testlets: (a) generalized partial credit model (GPCM), (b) testlet-as-a-polytomous-item model (TPIM), (c)…
Descriptors: Goodness of Fit, Item Response Theory, Test Items, Scoring
Haimiao Yuan – ProQuest LLC, 2022
The application of diagnostic classification models (DCMs) in the field of educational measurement is getting more attention in recent years. To make a valid inference from the model, it is important to ensure that the model fits the data. The purpose of the present study was to investigate the performance of the limited information…
Descriptors: Goodness of Fit, Educational Assessment, Educational Diagnosis, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Sedat Sen; Allan S. Cohen – Educational and Psychological Measurement, 2024
A Monte Carlo simulation study was conducted to compare fit indices used for detecting the correct latent class in three dichotomous mixture item response theory (IRT) models. Ten indices were considered: Akaike's information criterion (AIC), the corrected AIC (AICc), Bayesian information criterion (BIC), consistent AIC (CAIC), Draper's…
Descriptors: Goodness of Fit, Item Response Theory, Sample Size, Classification
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ketabi, Somaye; Alavi, Seyyed Mohammed; Ravand, Hamdollah – International Journal of Language Testing, 2021
Although Diagnostic Classification Models (DCMs) were introduced to education system decades ago, it seems that these models were not employed for the original aims upon which they had been designed. Using DCMs has been mostly common in analyzing large-scale non-diagnostic tests and these models have been rarely used in developing Cognitive…
Descriptors: Diagnostic Tests, Test Construction, Goodness of Fit, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Shafipoor, Mahdieh; Ravand, Hamdollah; Maftoon, Parviz – Language Testing in Asia, 2021
The current study compared the model fit indices, skill mastery probabilities, and classification accuracy of six Diagnostic Classification Models (DCMs): a general model (G-DINA) against five specific models (LLM, RRUM, ACDM, DINA, and DINO). To do so, the response data to the grammar and vocabulary sections of a General English Achievement Test,…
Descriptors: Goodness of Fit, Models, Classification, Grammar
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sahin, Murat Dogan – International Electronic Journal of Elementary Education, 2020
Advanced Item Response Theory (IRT) practices serve well in understanding the nature of latent variables which have been subject to research in various disciplines. In the current study, 7-12 aged 2536 children's responses to 20- item Visual Sequential Processing Memory (VSPM) sub-test of Anadolu-Sak Intelligence Scale (ASIS) were analyzed with…
Descriptors: Item Response Theory, Memory, Intelligence Tests, Children
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mousavi, Amin; Cui, Ying – Education Sciences, 2020
Often, important decisions regarding accountability and placement of students in performance categories are made on the basis of test scores generated from tests, therefore, it is important to evaluate the validity of the inferences derived from test results. One of the threats to the validity of such inferences is aberrant responding. Several…
Descriptors: Student Evaluation, Educational Testing, Psychological Testing, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Ren – Educational and Psychological Measurement, 2018
Attribute structure is an explicit way of presenting the relationship between attributes in diagnostic measurement. The specification of attribute structures directly affects the classification accuracy resulted from psychometric modeling. This study provides a conceptual framework for understanding misspecifications of attribute structures. Under…
Descriptors: Diagnostic Tests, Classification, Test Construction, Relationship
Peer reviewed Peer reviewed
Direct linkDirect link
Ravand, Hamdollah; Baghaei, Purya – International Journal of Testing, 2020
More than three decades after their introduction, diagnostic classification models (DCM) do not seem to have been implemented in educational systems for the purposes they were devised. Most DCM research is either methodological for model development and refinement or retrofitting to existing nondiagnostic tests and, in the latter case, basically…
Descriptors: Classification, Models, Diagnostic Tests, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Delafontaine, Jolien; Chen, Changsheng; Park, Jung Yeon; Van den Noortgate, Wim – Large-scale Assessments in Education, 2022
In cognitive diagnosis assessment (CDA), the impact of misspecified item-attribute relations (or "Q-matrix") designed by subject-matter experts has been a great challenge to real-world applications. This study examined parameter estimation of the CDA with the expert-designed Q-matrix and two refined Q-matrices for international…
Descriptors: Q Methodology, Matrices, Cognitive Measurement, Diagnostic Tests
Kilgus, Stephen P.; Bonifay, Wes E.; Eklund, Katie; von der Embse, Nathaniel P.; Peet, Casie; Izumi, Jared; Shim, Hyejin; Meyer, Lauren N. – Grantee Submission, 2020
The purpose of this study was to support the development and initial validation of the Intervention Selection Profile (ISP)-Skills, a brief 14-item teacher rating scale intended to inform the selection and delivery of instructional interventions at Tier 2. Teacher participants (n = 196) rated five students from their classroom across four measures…
Descriptors: Test Construction, Test Validity, Intervention, Rating Scales
Peer reviewed Peer reviewed
Direct linkDirect link
Min, Shangchao; Cai, Hongwen; He, Lianzhen – Language Assessment Quarterly, 2022
The present study examined the performance of the bi-factor multidimensional item response theory (MIRT) model and higher-order (HO) cognitive diagnostic models (CDM) in providing diagnostic information and general ability estimation simultaneously in a listening test. The data used were 1,611 examinees' item-level responses to an in-house EFL…
Descriptors: Listening Comprehension Tests, English (Second Language), Second Language Learning, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Huang, Hung-Yu – Journal of Educational Measurement, 2017
Cognitive diagnosis models (CDMs) have been developed to evaluate the mastery status of individuals with respect to a set of defined attributes or skills that are measured through testing. When individuals are repeatedly administered a cognitive diagnosis test, a new class of multilevel CDMs is required to assess the changes in their attributes…
Descriptors: Testing, Cognitive Measurement, Test Items, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
DiStefano, Christine; McDaniel, Heather L.; Zhang, Liyun; Shi, Dexin; Jiang, Zhehan – Educational and Psychological Measurement, 2019
A simulation study was conducted to investigate the model size effect when confirmatory factor analysis (CFA) models include many ordinal items. CFA models including between 15 and 120 ordinal items were analyzed with mean- and variance-adjusted weighted least squares to determine how varying sample size, number of ordered categories, and…
Descriptors: Factor Analysis, Effect Size, Data, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Tay-lim, Brenda Siok-Hoon; Zhang, Jinming – Applied Measurement in Education, 2015
To ensure the statistical result validity, model-data fit must be evaluated for each item. In practice, certain actions or treatments are needed for misfit items. If all misfit items are treated, much item information would be lost during calibration. On the other hand, if only severely misfit items are treated, the inclusion of misfit items may…
Descriptors: Test Items, Goodness of Fit, Classification, Item Response Theory
Previous Page | Next Page ยป
Pages: 1  |  2