NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
No Child Left Behind Act 20011
What Works Clearinghouse Rating
Showing 1 to 15 of 40 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Guo, Wenjing; Choi, Youn-Jeng – Educational and Psychological Measurement, 2023
Determining the number of dimensions is extremely important in applying item response theory (IRT) models to data. Traditional and revised parallel analyses have been proposed within the factor analysis framework, and both have shown some promise in assessing dimensionality. However, their performance in the IRT framework has not been…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Guidelines
Peer reviewed Peer reviewed
Direct linkDirect link
Novak, Josip; Rebernjak, Blaž – Measurement: Interdisciplinary Research and Perspectives, 2023
A Monte Carlo simulation study was conducted to examine the performance of [alpha], [lambda]2, [lambda][subscript 4], [lambda][subscript 2], [omega][subscript T], GLB[subscript MRFA], and GLB[subscript Algebraic] coefficients. Population reliability, distribution shape, sample size, test length, and number of response categories were varied…
Descriptors: Monte Carlo Methods, Evaluation Methods, Reliability, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Yuanfang Liu; Mark H. C. Lai; Ben Kelcey – Structural Equation Modeling: A Multidisciplinary Journal, 2024
Measurement invariance holds when a latent construct is measured in the same way across different levels of background variables (continuous or categorical) while controlling for the true value of that construct. Using Monte Carlo simulation, this paper compares the multiple indicators, multiple causes (MIMIC) model and MIMIC-interaction to a…
Descriptors: Classification, Accuracy, Error of Measurement, Correlation
Peer reviewed Peer reviewed
Direct linkDirect link
Pere J. Ferrando; David Navarro-González; Fabia Morales-Vives – Educational and Psychological Measurement, 2025
The problem of local item dependencies (LIDs) is very common in personality and attitude measures, particularly in those that measure narrow-bandwidth dimensions. At the structural level, these dependencies can be modeled by using extended factor analytic (FA) solutions that include correlated residuals. However, the effects that LIDs have on the…
Descriptors: Scores, Accuracy, Evaluation Methods, Factor Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Metsämuuronen, Jari – International Journal of Educational Methodology, 2020
A new index of item discrimination power (IDP), dimension-corrected Somers' D (D2) is proposed. Somers' D is one of the superior alternatives for item-total- (Rit) and item-rest correlation (Rir) in reflecting the real IDP with items with scales 0/1 and 0/1/2, that is, up to three categories. D also reaches the extreme value +1 and -1 correctly…
Descriptors: Item Analysis, Correlation, Test Items, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Hung, Su-Pin; Huang, Hung-Yu – Journal of Educational and Behavioral Statistics, 2022
To address response style or bias in rating scales, forced-choice items are often used to request that respondents rank their attitudes or preferences among a limited set of options. The rating scales used by raters to render judgments on ratees' performance also contribute to rater bias or errors; consequently, forced-choice items have recently…
Descriptors: Evaluation Methods, Rating Scales, Item Analysis, Preferences
Hosseinzadeh, Mostafa – ProQuest LLC, 2021
In real-world situations, multidimensional data may appear on large-scale tests or attitudinal surveys. A simple structure, multidimensional model may be used to evaluate the items, ignoring the cross-loading of some items on the secondary dimension. The purpose of this study was to investigate the influence of structure complexity magnitude of…
Descriptors: Item Response Theory, Models, Simulation, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Smith, Trevor I.; Bendjilali, Nasrine – Physical Review Physics Education Research, 2022
Several recent studies have employed item response theory (IRT) to rank incorrect responses to commonly used research-based multiple-choice assessments. These studies use Bock's nominal response model (NRM) for applying IRT to categorical (nondichotomous) data, but the response rankings only utilize half of the parameters estimated by the model.…
Descriptors: Item Response Theory, Test Items, Multiple Choice Tests, Science Tests
Bernhardt, Amery E. – ProQuest LLC, 2022
This quantitative correlational study dives into the heart of understanding the significance of model fidelity for implementing school threat assessment teams. The target population was instructional staff and threat assessment team members from schools in Dutchess, Putnam, and Westchester Counties in New York State that use the Comprehensive…
Descriptors: Evaluation Methods, Educational Environment, Correlation, Fidelity
Peer reviewed Peer reviewed
Direct linkDirect link
Park, Sung Eun; Ahn, Soyeon; Zopluoglu, Cengiz – Educational and Psychological Measurement, 2021
This study presents a new approach to synthesizing differential item functioning (DIF) effect size: First, using correlation matrices from each study, we perform a multigroup confirmatory factor analysis (MGCFA) that examines measurement invariance of a test item between two subgroups (i.e., focal and reference groups). Then we synthesize, across…
Descriptors: Item Analysis, Effect Size, Difficulty Level, Monte Carlo Methods
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Malec, Wojciech; Krzeminska-Adamek, Malgorzata – Practical Assessment, Research & Evaluation, 2020
The main objective of the article is to compare several methods of evaluating multiple-choice options through classical item analysis. The methods subjected to examination include the tabulation of choice distribution, the interpretation of trace lines, the point-biserial correlation, the categorical analysis of trace lines, and the investigation…
Descriptors: Comparative Analysis, Evaluation Methods, Multiple Choice Tests, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Schat, Esther; van der Knaap, Ewout; de Graaff, Rick – Intercultural Communication Education, 2021
Intercultural competence is a crucial element of foreign language education, yet the multifaceted nature of this construct makes it inherently difficult to assess. Although several tools for evaluating intercultural competence currently exist, research on their use in secondary school settings is scarce. This study reports on the development and…
Descriptors: Intercultural Communication, Communicative Competence (Languages), Second Language Learning, Second Language Instruction
Lopata, Christopher; Donnelly, James P.; Rodgers, Jonathan D.; Thomeer, Marcus L.; Booth, Adam J. – Autism: The International Journal of Research and Practice, 2020
This study assessed the reliability and criterion-related validity of teacher ratings on the Adapted Skillstreaming Checklist for a sample of 133 children, aged 6-11 years, with autism spectrum disorder (without intellectual disability). Internal consistency for the total sample was 0.93. For a subsample, test-retest reliability was very good (r =…
Descriptors: Check Lists, Validity, Reliability, Teacher Attitudes
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rihák, Jirí; Pelánek, Radek – International Educational Data Mining Society, 2017
Educational systems typically contain a large pool of items (questions, problems). Using data mining techniques we can group these items into knowledge components, detect duplicated items and outliers, and identify missing items. To these ends, it is useful to analyze item similarities, which can be used as input to clustering or visualization…
Descriptors: Item Analysis, Data Analysis, Visualization, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Rigney, Alexander M. – Journal of Psychoeducational Assessment, 2019
This report reviews the "Social Skills Improvement System Social-Emotional Learning Edition" (SSIS SEL; Gresham & Elliott, 2017), a multicomponent rating scale that includes a criterion and norm-referenced measure of social-emotional and academic functioning--based on a reformulation of the "Social Skills Improvement…
Descriptors: Rating Scales, Interpersonal Competence, Social Development, Emotional Development
Previous Page | Next Page »
Pages: 1  |  2  |  3