Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 6 |
Descriptor
Computer Assisted Testing | 9 |
Markov Processes | 9 |
Monte Carlo Methods | 9 |
Bayesian Statistics | 6 |
Goodness of Fit | 4 |
Item Response Theory | 4 |
Models | 4 |
Test Items | 4 |
Language Tests | 3 |
Statistical Analysis | 3 |
Cognitive Processes | 2 |
More ▼ |
Source
ETS Research Report Series | 3 |
Journal of Educational and… | 3 |
Grantee Submission | 1 |
International Journal of… | 1 |
Psychological Methods | 1 |
Author
Douglas, Jeffrey A. | 2 |
Akihito Kamata | 1 |
Bradlow, Eric T. | 1 |
Chang, Hua-Hua | 1 |
Cornelis Potgieter | 1 |
Culpepper, Steven Andrew | 1 |
Eckes, Thomas | 1 |
Esther Ulitzsch | 1 |
Fan, Zhewen | 1 |
Fox, Jean-Paul | 1 |
Hornke, Lutz F. | 1 |
More ▼ |
Publication Type
Journal Articles | 8 |
Reports - Research | 8 |
Reports - Evaluative | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Secondary Education | 1 |
Audience
Location
Germany | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 1 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Xin Qiao; Akihito Kamata; Yusuf Kara; Cornelis Potgieter; Joseph Nese – Grantee Submission, 2023
In this article, the beta-binomial model for count data is proposed and demonstrated in terms of its application in the context of oral reading fluency (ORF) assessment, where the number of words read correctly (WRC) is of interest. Existing studies adopted the binomial model for count data in similar assessment scenarios. The beta-binomial model,…
Descriptors: Oral Reading, Reading Fluency, Bayesian Statistics, Markov Processes
Esther Ulitzsch; Steffi Pohl; Lale Khorramdel; Ulf Kroehne; Matthias von Davier – Journal of Educational and Behavioral Statistics, 2024
Questionnaires are by far the most common tool for measuring noncognitive constructs in psychology and educational sciences. Response bias may pose an additional source of variation between respondents that threatens validity of conclusions drawn from questionnaire data. We present a mixture modeling approach that leverages response time data from…
Descriptors: Item Response Theory, Response Style (Tests), Questionnaires, Secondary School Students
Eckes, Thomas; Jin, Kuan-Yu – International Journal of Testing, 2021
Severity and centrality are two main kinds of rater effects posing threats to the validity and fairness of performance assessments. Adopting Jin and Wang's (2018) extended facets modeling approach, we separately estimated the magnitude of rater severity and centrality effects in the web-based TestDaF (Test of German as a Foreign Language) writing…
Descriptors: Language Tests, German, Second Languages, Writing Tests
Wang, Shiyu; Yang, Yan; Culpepper, Steven Andrew; Douglas, Jeffrey A. – Journal of Educational and Behavioral Statistics, 2018
A family of learning models that integrates a cognitive diagnostic model and a higher-order, hidden Markov model in one framework is proposed. This new framework includes covariates to model skill transition in the learning environment. A Bayesian formulation is adopted to estimate parameters from a learning model. The developed methods are…
Descriptors: Skill Development, Cognitive Measurement, Cognitive Processes, Markov Processes
Wang, Chun; Fan, Zhewen; Chang, Hua-Hua; Douglas, Jeffrey A. – Journal of Educational and Behavioral Statistics, 2013
The item response times (RTs) collected from computerized testing represent an underutilized type of information about items and examinees. In addition to knowing the examinees' responses to each item, we can investigate the amount of time examinees spend on each item. Current models for RTs mainly focus on parametric models, which have the…
Descriptors: Reaction Time, Computer Assisted Testing, Test Items, Accuracy
Klein Entink, Rinke H.; Kuhn, Jorg-Tobias; Hornke, Lutz F.; Fox, Jean-Paul – Psychological Methods, 2009
In current psychological research, the analysis of data from computer-based assessments or experiments is often confined to accuracy scores. Response times, although being an important source of additional information, are either neglected or analyzed separately. In this article, a new model is developed that allows the simultaneous analysis of…
Descriptors: Psychological Studies, Monte Carlo Methods, Markov Processes, Educational Assessment
Wang, Xiaohui; Bradlow, Eric T.; Wainer, Howard – ETS Research Report Series, 2005
SCORIGHT is a very general computer program for scoring tests. It models tests that are made up of dichotomously or polytomously rated items or any kind of combination of the two through the use of a generalized item response theory (IRT) formulation. The items can be presented independently or grouped into clumps of allied items (testlets) or in…
Descriptors: Computer Assisted Testing, Statistical Analysis, Test Items, Bayesian Statistics
von Davier, Matthias – ETS Research Report Series, 2005
Probabilistic models with more than one latent variable are designed to report profiles of skills or cognitive attributes. Testing programs want to offer additional information beyond what a single test score can provide using these skill profiles. Many recent approaches to skill profile models are limited to dichotomous data and have made use of…
Descriptors: Models, Diagnostic Tests, Language Tests, Language Proficiency
Stricker, Lawrence J.; Rock, Donald A.; Lee, Yong-Won – ETS Research Report Series, 2005
This study assessed the factor structure of the LanguEdge™ test and the invariance of its factors across language groups. Confirmatory factor analyses of individual tasks and subsets of items in the four sections of the test, Listening, Reading, Speaking, and Writing, was carried out for Arabic-, Chinese-, and Spanish-speaking test takers. Two…
Descriptors: Factor Structure, Language Tests, Factor Analysis, Semitic Languages