Publication Date
In 2025 | 0 |
Since 2024 | 12 |
Since 2021 (last 5 years) | 34 |
Since 2016 (last 10 years) | 51 |
Since 2006 (last 20 years) | 95 |
Descriptor
Error of Measurement | 95 |
Item Response Theory | 30 |
Scores | 25 |
Comparative Analysis | 21 |
Simulation | 20 |
Models | 19 |
Test Items | 19 |
Goodness of Fit | 18 |
Evaluation Methods | 17 |
Statistical Analysis | 16 |
Correlation | 15 |
More ▼ |
Source
ProQuest LLC | 95 |
Author
Publication Type
Dissertations/Theses -… | 95 |
Education Level
Elementary Education | 9 |
Higher Education | 9 |
Elementary Secondary Education | 7 |
Postsecondary Education | 6 |
Secondary Education | 6 |
Grade 8 | 3 |
Junior High Schools | 3 |
Middle Schools | 3 |
Grade 4 | 1 |
Grade 7 | 1 |
High Schools | 1 |
More ▼ |
Audience
Researchers | 1 |
Location
United States | 3 |
North Carolina | 2 |
Chile | 1 |
China | 1 |
Colombia | 1 |
Florida | 1 |
New York | 1 |
North Dakota | 1 |
Pakistan | 1 |
South Korea | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Ayse Busra Ceviren – ProQuest LLC, 2024
Latent change score (LCS) models are a powerful class of structural equation modeling that allows researchers to work with latent difference scores that minimize measurement error. LCS models define change as a function of prior status, which makes it well-suited for modeling developmental theories or processes. In LCS models, like other latent…
Descriptors: Structural Equation Models, Error of Measurement, Statistical Bias, Monte Carlo Methods
Jiangqiong Li – ProQuest LLC, 2024
When measuring latent constructs, for example, language ability, we use statistical models to specify appropriate relationships between the latent construct and observe responses to test items. These models rely on theoretical assumptions to ensure accurate parameter estimates for valid inferences based on the test results. This dissertation…
Descriptors: Goodness of Fit, Item Response Theory, Models, Measurement Techniques
Christopher E. Shank – ProQuest LLC, 2024
This dissertation compares the performance of equivalence test (EQT) and null hypothesis test (NHT) procedures for identifying invariant and noninvariant factor loadings under a range of experimental manipulations. EQT is the statistically appropriate approach when the research goal is to find evidence of group similarity rather than group…
Descriptors: Factor Analysis, Goodness of Fit, Intervals, Comparative Analysis
Jiayi Deng – ProQuest LLC, 2024
Test score comparability in international large-scale assessments (LSA) is of utmost importance in measuring the effectiveness of education systems and understanding the impact of education on economic growth. To effectively compare test scores on an international scale, score linking is widely used to convert raw scores from different linguistic…
Descriptors: Item Response Theory, Scoring Rubrics, Scoring, Error of Measurement
Klauth, Bo – ProQuest LLC, 2023
In conducting confirmatory factor analysis with ordered response items, the literature suggests that when the number of responses is five and item skewness (IS) is approximately normal, researchers can employ maximum likelihood with robust standard errors (MLR). However, MLR can yield biased factor loadings (FL) and FL standard errors (FLSE) when…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Error of Measurement
Jackson, Kayla – ProQuest LLC, 2023
Prior research highlights the benefits of multimode surveys and best practices for item-by-item (IBI) and matrix-type survey items. Some researchers have explored whether mode differences for online and paper surveys persist for these survey item types. However, no studies discuss measurement invariance when both item types and online modes are…
Descriptors: Test Items, Surveys, Error of Measurement, Item Response Theory
Ryan Derickson – ProQuest LLC, 2022
Item Response Theory (IRT) models are a popular analytic method for self report data. We show how traditional IRT models can be vulnerable to specific kinds of asymmetric measurement error (AME) in self-report data, because the models spread the error to all estimates -- even those of items that do not contribute error. We quantify the impact of…
Descriptors: Item Response Theory, Measurement Techniques, Error of Measurement, Models
Olasunkanmi James Kehinde – ProQuest LLC, 2024
The Q-matrix played a key role in implementations of diagnostic classification models (DCMs) or cognitive diagnostic models (CDMs) -- a family of psychometric models that are gaining attention in providing diagnostic information on students' mastery of cognitive attributes or skills. Using two Monte Carlo simulation studies, this dissertation…
Descriptors: Diagnostic Tests, Q Methodology, Learning Trajectories, Sample Size
Montserrat Beatriz Valdivia Medinaceli – ProQuest LLC, 2023
My dissertation examines three current challenges of international large-scale assessments (ILSAs) associated with the transition from linear testing to an adaptive testing design. ILSAs are important for making comparisons among populations and informing countries about the quality of their educational systems. ILSA's results inform policymakers…
Descriptors: International Assessment, Achievement Tests, Adaptive Testing, Test Items
Atehortua, Laura – ProQuest LLC, 2022
Intelligence tests are used in a variety of settings such as schools, clinics, and courts to assess the intellectual capacity of individuals of all ages. Intelligence tests are used to make high-stakes decisions such as special education placement, employment, eligibility for social security services, and determination of the death penalty.…
Descriptors: Adults, Intelligence Tests, Children, Error of Measurement
Ning Jiang – ProQuest LLC, 2022
The purpose of this study is to evaluate the performance of three commonly used model fit indices when measurement invariance is tested in the context of multiple-group CFA analysis with categorical-ordered data. As applied researchers are increasingly aware of the importance of testing measurement invariance, as well as Likert-type scales are…
Descriptors: Goodness of Fit, Factor Analysis, Data, Monte Carlo Methods
Wu, Tong – ProQuest LLC, 2023
This three-article dissertation aims to address three methodological challenges to ensure comparability in educational research, including scale linking, test equating, and propensity score (PS) weighting. The first study intends to improve test scale comparability by evaluating the effect of six missing data handling approaches, including…
Descriptors: Educational Research, Comparative Analysis, Equated Scores, Weighted Scores
Sophie Lilit Litschwartz – ProQuest LLC, 2021
In education research test scores are a common object of analysis. Across studies test scores can be an important outcome, a highly predictive covariate, or a means of assigning treatment. However, test scores are a measure of an underlying proficiency we can't observe directly and so contain error. This measurement error has implications for how…
Descriptors: Scores, Inferences, Educational Research, Evaluation Methods
Emily A. Brown – ProQuest LLC, 2024
Previous research has been limited regarding the measurement of computational thinking, particularly as a learning progression in K-12. This study proposes to apply a multidimensional item response theory (IRT) model to a newly developed measure of computational thinking utilizing both selected response and open-ended polytomous items to establish…
Descriptors: Models, Computation, Thinking Skills, Item Response Theory
Yu Wang – ProQuest LLC, 2024
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Cognitive Tests, Cognitive Measurement, Educational Diagnosis