NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 46 to 60 of 4,158 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Davis-Berg, Elizabeth C.; Minbiole, Julie – School Science Review, 2020
The completion rates were compared for long-form questions where a large blank answer space is provided and for long-form questions where the answer space has bullet-points prompts corresponding to the parts of the question. It was found that students were more likely to complete a question when bullet points were provided in the answer space.…
Descriptors: Test Format, Test Construction, Academic Achievement, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Xiao, Jiaying; Bulut, Okan – Educational and Psychological Measurement, 2020
Large amounts of missing data could distort item parameter estimation and lead to biased ability estimates in educational assessments. Therefore, missing responses should be handled properly before estimating any parameters. In this study, two Monte Carlo simulation studies were conducted to compare the performance of four methods in handling…
Descriptors: Data, Computation, Ability, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
van Groen, Maaike M.; Eggen, Theo J. H. M. – Journal of Applied Testing Technology, 2020
When developing a digital test, one of the first decisions that need to be made is which type of Computer-Based Test (CBT) to develop. Six different CBT types are considered here: linear tests, automatically generated tests, computerized adaptive tests, adaptive learning environments, educational simulations, and educational games. The selection…
Descriptors: Computer Assisted Testing, Formative Evaluation, Summative Evaluation, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Huber, Chuck – Measurement: Interdisciplinary Research and Perspectives, 2020
It is demonstrated that the popular three-parameter logistic model can lead to markedly inaccurate individual ability level estimates for mixture populations. A theoretically and empirically important setting is initially considered where (a) in one of two subpopulations (latent classes) the two-parameter logistic model holds for each item in a…
Descriptors: Item Response Theory, Models, Measurement Techniques, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2019
The Mantel-Haenszel delta difference (MH D-DIF) and the standardized proportion difference (STD P-DIF) are two observed-score methods that have been used to assess differential item functioning (DIF) at Educational Testing Service since the early 1990s. Latentvariable approaches to assessing measurement invariance at the item level have been…
Descriptors: Test Bias, Educational Testing, Statistical Analysis, Item Response Theory
Sinharay, Sandip; van Rijn, Peter W. – Journal of Educational and Behavioral Statistics, 2020
Response time models (RTMs) are of increasing interest in educational and psychological testing. This article focuses on the lognormal model for response times, which is one of the most popular RTMs. Several existing statistics for testing normality and the fit of factor analysis models are repurposed for testing the fit of the lognormal model. A…
Descriptors: Educational Testing, Psychological Testing, Goodness of Fit, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Yang; Wang, Xiaojing – Journal of Educational and Behavioral Statistics, 2020
Parametric methods, such as autoregressive models or latent growth modeling, are usually inflexible to model the dependence and nonlinear effects among the changes of latent traits whenever the time gap is irregular and the recorded time points are individually varying. Often in practice, the growth trend of latent traits is subject to certain…
Descriptors: Bayesian Statistics, Nonparametric Statistics, Regression (Statistics), Item Response Theory
Sinharay, Sandip; van Rijn, Peter – Grantee Submission, 2020
Response-time models are of increasing interest in educational and psychological testing. This paper focuses on the lognormal model for response times (van der Linden, 2006), which is one of the most popular response-time models. Several existing statistics for testing normality and the fit of factor-analysis models are repurposed for testing the…
Descriptors: Educational Testing, Psychological Testing, Goodness of Fit, Factor Analysis
Care, Esther; Kim, Helyn – Center for Universal Education at The Brookings Institution, 2020
This framework marks the first in a series of five reports detailing the work of the Optimizing Assessment for All (OAA) project at Brookings to strengthen education systems' capacity to integrate 21st century skills (21CS) into teaching and learning, using assessment as a lever for changing classroom practices. In a world of rapid advancement and…
Descriptors: 21st Century Skills, Foreign Countries, Educational Testing, Assessment Literacy
Berman, Amy I.; Haertel, Edward H.; Pellegrino, James W. – National Academy of Education, 2020
This National Academy of Education (NAEd) volume provides guidance to key stakeholders on how to accurately report and interpret comparability assertions concerning large-scale educational assessments as well as how to ensure greater comparability by paying close attention to key aspects of assessment design, content, and procedures. The goal of…
Descriptors: Educational Assessment, Educational Testing, Scores, Comparative Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Doneva, Rositsa; Gaftandzhieva, Siliva; Totkov, George – Turkish Online Journal of Distance Education, 2018
This paper presents a study on known approaches for quality assurance of educational test and test items. On its basis a comprehensive approach to the quality assurance of online educational testing is proposed to address the needs of all stakeholders (authors of online tests, teachers, students, experts, quality managers, etc.). According to the…
Descriptors: Educational Testing, Automation, Quality Assurance, Computer Assisted Testing
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Beck, Klaus – Frontline Learning Research, 2020
Many test developers try to ensure the content validity of their tests by having external experts review the items, e.g. in terms of relevance, difficulty, or clarity. Although this approach is widely accepted, a closer look reveals several pitfalls need to be avoided if experts' advice is to be truly helpful. The purpose of this paper is to…
Descriptors: Content Validity, Psychological Testing, Educational Testing, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Margarita Kilili-Lesta; Konstantinos Giannakou; Louiza Voniati – Autism & Developmental Language Impairments, 2024
Background and Aims: Children with autism spectrum disorder (ASD) present with impairments in social interaction and stereotypic behaviors. About a third may exhibit delays in verbal expression beyond preschool age, potentially categorizing them as nonverbal/minimally verbal (NV/MV), a condition that can persist into adulthood and affect their…
Descriptors: Foreign Countries, Autism Spectrum Disorders, Prognostic Tests, Educational Testing
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mousavi, Amin; Cui, Ying – Education Sciences, 2020
Often, important decisions regarding accountability and placement of students in performance categories are made on the basis of test scores generated from tests, therefore, it is important to evaluate the validity of the inferences derived from test results. One of the threats to the validity of such inferences is aberrant responding. Several…
Descriptors: Student Evaluation, Educational Testing, Psychological Testing, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Zwick, Rebecca – Educational Measurement: Issues and Practice, 2019
Selection decisions have a major impact on our education, occupation, and quality of life, and the role of standardized tests in selection has always been a source of controversy. Here, I consider various definitions of fairness in measurement and selection--those emerging from within educational measurement and statistics, those from philosophy,…
Descriptors: Culture Fair Tests, Decision Making, Standardized Tests, Selection Criteria
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  278