Publication Date
In 2025 | 1 |
Since 2024 | 5 |
Since 2021 (last 5 years) | 28 |
Since 2016 (last 10 years) | 66 |
Since 2006 (last 20 years) | 95 |
Descriptor
Accuracy | 95 |
Comparative Analysis | 95 |
Models | 95 |
Prediction | 26 |
Classification | 22 |
Foreign Countries | 20 |
Computer Software | 19 |
Statistical Analysis | 19 |
Item Response Theory | 18 |
Correlation | 16 |
Simulation | 16 |
More ▼ |
Source
Author
Ratcliff, Roger | 3 |
Bosch, Nigel | 2 |
Candace Walkington | 2 |
Chun Wang | 2 |
Gongjun Xu | 2 |
Kelsey Schenck | 2 |
Koon, Sharon | 2 |
Min Wang | 2 |
Mitchell J. Nathan | 2 |
Monroe, Scott | 2 |
Petscher, Yaacov | 2 |
More ▼ |
Publication Type
Education Level
Audience
Teachers | 2 |
Administrators | 1 |
Parents | 1 |
Policymakers | 1 |
Location
Spain | 4 |
China | 3 |
Australia | 2 |
Florida | 2 |
Germany | 2 |
Iran | 2 |
Israel | 2 |
Netherlands | 2 |
North Carolina | 2 |
Argentina | 1 |
California (San Diego) | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Jiaying Xiao; Chun Wang; Gongjun Xu – Grantee Submission, 2024
Accurate item parameters and standard errors (SEs) are crucial for many multidimensional item response theory (MIRT) applications. A recent study proposed the Gaussian Variational Expectation Maximization (GVEM) algorithm to improve computational efficiency and estimation accuracy (Cho et al., 2021). However, the SE estimation procedure has yet to…
Descriptors: Error of Measurement, Models, Evaluation Methods, Item Analysis
Christopher E. Shank – ProQuest LLC, 2024
This dissertation compares the performance of equivalence test (EQT) and null hypothesis test (NHT) procedures for identifying invariant and noninvariant factor loadings under a range of experimental manipulations. EQT is the statistically appropriate approach when the research goal is to find evidence of group similarity rather than group…
Descriptors: Factor Analysis, Goodness of Fit, Intervals, Comparative Analysis
Xieling Chen; Haoran Xie; Di Zou; Lingling Xu; Fu Lee Wang – Educational Technology & Society, 2025
In massive open online course (MOOC) environments, computer-based analysis of course reviews enables instructors and course designers to develop intervention strategies and improve instruction to support learners' learning. This study aimed to automatically and effectively identify learners' concerned topics within their written reviews. First, we…
Descriptors: Classification, MOOCs, Teaching Skills, Artificial Intelligence
Song, Yoon Ah; Lee, Won-Chan – Applied Measurement in Education, 2022
This article presents the performance of item response theory (IRT) models when double ratings are used as item scores over single ratings when rater effects are present. Study 1 examined the influence of the number of ratings on the accuracy of proficiency estimation in the generalized partial credit model (GPCM). Study 2 compared the accuracy of…
Descriptors: Item Response Theory, Item Analysis, Scores, Accuracy
Rüttenauer, Tobias – Sociological Methods & Research, 2022
Spatial regression models provide the opportunity to analyze spatial data and spatial processes. Yet, several model specifications can be used, all assuming different types of spatial dependence. This study summarizes the most commonly used spatial regression models and offers a comparison of their performance by using Monte Carlo experiments. In…
Descriptors: Models, Monte Carlo Methods, Social Science Research, Data Analysis
Demir, Seda; Doguyurt, Mehmet Fatih – African Educational Research Journal, 2022
The purpose of this research was to compare the performances of the Fixed Effect Model (FEM) and the Random Effects Model (REM) in the meta-analysis studies conducted through 5, 10, 20 and 40 studies with an outlier and 4, 9, 19 and 39 studies without an outlier in terms of estimated common effect size, confidence interval coverage rate and…
Descriptors: Meta Analysis, Comparative Analysis, Research Reports, Effect Size
Kalkan, Ömür Kaya – Measurement: Interdisciplinary Research and Perspectives, 2022
The four-parameter logistic (4PL) Item Response Theory (IRT) model has recently been reconsidered in the literature due to the advances in the statistical modeling software and the recent developments in the estimation of the 4PL IRT model parameters. The current simulation study evaluated the performance of expectation-maximization (EM),…
Descriptors: Comparative Analysis, Sample Size, Test Length, Algorithms
Zhang, Mengxue; Heffernan, Neil; Lan, Andrew – International Educational Data Mining Society, 2023
Automated scoring of student responses to open-ended questions, including short-answer questions, has great potential to scale to a large number of responses. Recent approaches for automated scoring rely on supervised learning, i.e., training classifiers or fine-tuning language models on a small number of responses with human-provided score…
Descriptors: Scoring, Computer Assisted Testing, Mathematics Instruction, Mathematics Tests
Huang, Hung-Yu – Educational and Psychological Measurement, 2023
The forced-choice (FC) item formats used for noncognitive tests typically develop a set of response options that measure different traits and instruct respondents to make judgments among these options in terms of their preference to control the response biases that are commonly observed in normative tests. Diagnostic classification models (DCMs)…
Descriptors: Test Items, Classification, Bayesian Statistics, Decision Making
Salem, Alexandra C.; Gale, Robert; Casilio, Marianne; Fleegle, Mikala; Fergadiotis, Gerasimos; Bedrick, Steven – Journal of Speech, Language, and Hearing Research, 2023
Purpose: ParAlg (Paraphasia Algorithms) is a software that automatically categorizes a person with aphasia's naming error (paraphasia) in relation to its intended target on a picture-naming test. These classifications (based on lexicality as well as semantic, phonological, and morphological similarity to the target) are important for…
Descriptors: Semantics, Computer Software, Aphasia, Classification
Chun Wang; Ruoyi Zhu; Gongjun Xu – Grantee Submission, 2022
Differential item functioning (DIF) analysis refers to procedures that evaluate whether an item's characteristic differs for different groups of persons after controlling for overall differences in performance. DIF is routinely evaluated as a screening step to ensure items behavior the same across groups. Currently, the majority DIF studies focus…
Descriptors: Models, Item Response Theory, Item Analysis, Comparative Analysis
Logacev, Pavel – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2023
A number of studies have found evidence for the so-called "ambiguity advantage," that is, faster processing of ambiguous sentences compared with unambiguous counterparts. While a number of proposals regarding the mechanism underlying this phenomenon have been made, the empirical evidence so far is far from unequivocal. It is compatible…
Descriptors: Phrase Structure, Accuracy, Ambiguity (Semantics), Sentences
Rungsilp, Chutimon; Piromsopa, Krerk; Viriyopase, Atthaphon; U-Yen, Kongpop – International Association for Development of the Information Society, 2021
The study of mind-wandering is popular since it is linked to the emotional problems and working/learning performance. In terms of education, it impacts comprehension during learning which affects academic success. Therefore, we sought to develop a machine learning model for an embedded portable device that can categorize mind-wandering state to…
Descriptors: Brain Hemisphere Functions, Diagnostic Tests, Artificial Intelligence, Cognitive Processes
Yi Gui – ProQuest LLC, 2024
This study explores using transfer learning in machine learning for natural language processing (NLP) to create generic automated essay scoring (AES) models, providing instant online scoring for statewide writing assessments in K-12 education. The goal is to develop an instant online scorer that is generalizable to any prompt, addressing the…
Descriptors: Writing Tests, Natural Language Processing, Writing Evaluation, Scoring
Nahatame, Shingo – Language Learning, 2021
Although text readability has traditionally been measured based on simple linguistic features, recent studies have employed natural language processing techniques to develop new readability formulas that better represent theoretical accounts of reading processes. This study evaluated the construct validity of different readability formulas,…
Descriptors: Readability, Natural Language Processing, Readability Formulas, Reading Processes