NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 23 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
He, Yinhong – Journal of Educational Measurement, 2023
Back random responding (BRR) behavior is one of the commonly observed careless response behaviors. Accurately detecting BRR behavior can improve test validities. Yu and Cheng (2019) showed that the change point analysis (CPA) procedure based on weighted residual (CPA-WR) performed well in detecting BRR. Compared with the CPA procedure, the…
Descriptors: Test Validity, Item Response Theory, Measurement, Monte Carlo Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Qi, Hongchao; Rizopoulos, Dimitris; Rosmalen, Joost – Research Synthesis Methods, 2023
The meta-analytic-predictive (MAP) approach is a Bayesian method to incorporate historical controls in new trials that aims to increase the statistical power and reduce the required sample size. Here we investigate how to calculate the sample size of the new trial when historical data is available, and the MAP approach is used in the analysis. In…
Descriptors: Sample Size, Computation, Meta Analysis, Bayesian Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Yuting Han; Zhehan Jiang; Lingling Xu; Fen Cai – AERA Online Paper Repository, 2024
To address the computational constraints of parameter estimation in the polytomous Cognitive Diagnosis Model (pCDM) in large-scale high data volume situations, this study proposes two two-stage polytomous attribute estimation methods: P_max and P_linear. The effects of the two-stage methods were studied via a Monte Carlo simulation study, and the…
Descriptors: Medical Education, Licensing Examinations (Professions), Measurement Techniques, Statistical Data
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Basman, Munevver – International Journal of Assessment Tools in Education, 2023
To ensure the validity of the tests is to check that all items have similar results across different groups of individuals. However, differential item functioning (DIF) occurs when the results of individuals with equal ability levels from different groups differ from each other on the same test item. Based on Item Response Theory and Classic Test…
Descriptors: Test Bias, Test Items, Test Validity, Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Musa Adekunle Ayanwale; Mdutshekelwa Ndlovu – Journal of Pedagogical Research, 2024
The COVID-19 pandemic has had a significant impact on high-stakes testing, including the national benchmark tests in South Africa. Current linear testing formats have been criticized for their limitations, leading to a shift towards Computerized Adaptive Testing [CAT]. Assessments with CAT are more precise and take less time. Evaluation of CAT…
Descriptors: Adaptive Testing, Benchmarking, National Competency Tests, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Elizabeth Talbott; Andres De Los Reyes; Junhui Yang; Mo Wang – Society for Research on Educational Effectiveness, 2024
Youth in the U.S. face a mental health crisis--one that has steadily worsened over the past 10 years (Centers for Disease Control, 2023). Youth with disabilities may experience increased mental health risk compared to their peers without disabilities. Recent national surveys and systematic reviews reveal that mental health risk may be elevated for…
Descriptors: Youth, Youth Problems, Mental Health, Crisis Intervention
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Boquan; Polce, Evan; Sprott, Julien C.; Jiang, Jack J. – Journal of Speech, Language, and Hearing Research, 2018
Purpose: The purpose of this study is to introduce a chaos level test to evaluate linear and nonlinear voice type classification method performances under varying signal chaos conditions without subjective impression. Study Design: Voice signals were constructed with differing degrees of noise to model signal chaos. Within each noise power, 100…
Descriptors: Acoustics, Monte Carlo Methods, Correlation, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Ubuz, Behiye; Aydinyer, Yurdagül – International Journal of Mathematical Education in Science and Technology, 2017
The current study aimed to construct a questionnaire that measures students' personality traits related to "striving for understanding" and "learning value of geometry" and then examine its psychometric properties. Through the use of multiple methods on two independent samples of 402 and 521 middle school students, two studies…
Descriptors: Mathematics Instruction, Test Construction, Questionnaires, Personality Traits
Peer reviewed Peer reviewed
Direct linkDirect link
Gordon-Hollingsworth, Arlene T.; Thompson, Julia E.; Geary, Meghan A.; Schexnaildre, Mark A.; Lai, Betty S.; Kelley, Mary Lou – Measurement and Evaluation in Counseling and Development, 2016
The Social Support Questionnaire for Children (SSQC) is a 50-item scale that assesses children's social support from parents, relatives, nonrelative adults, siblings, and peers. The SSQC demonstrates good psychometric properties (e.g., internal consistency, factorial validity). Furthermore, the SSQC appears to be an ethnically sensitive measure of…
Descriptors: Social Support Groups, Questionnaires, Children, Test Construction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Aminifar, Elahe; Alipour, Mohammad – European Journal of Educational Sciences, 2014
Item bank is one of the main components of adaptive tests. In this research, a test was made in order to design and calibrate items for Homogeneous Second Order Differential Equations. The items were designed according to the goal-content's table of the subject and the Bloom's taxonomy learning domain. Validity and reliability of these items was…
Descriptors: Test Items, Calculus, Mathematics Tests, Mathematics Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Aslanides, J. S.; Savage, C. M. – Physical Review Special Topics - Physics Education Research, 2013
We report on a concept inventory for special relativity: the development process, data analysis methods, and results from an introductory relativity class. The Relativity Concept Inventory tests understanding of relativistic concepts. An unusual feature is confidence testing for each question. This can provide additional information; for example,…
Descriptors: Physics, Science Tests, Scientific Concepts, Confidence Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Reeve, Charlie L.; Heggestad, Eric D.; Lievens, Filip – Intelligence, 2009
The assessment of cognitive abilities, whether it is for purposes of basic research or applied decision making, is potentially susceptible to both facilitating and debilitating influences. However, relatively little research has examined the degree to which these factors might moderate the criterion-related validity of cognitive ability tests. To…
Descriptors: Test Anxiety, Familiarity, Cognitive Tests, Test Validity
Garrett, Phyllis – ProQuest LLC, 2009
The use of polytomous items in assessments has increased over the years, and as a result, the validity of these assessments has been a concern. Differential item functioning (DIF) and missing data are two factors that may adversely affect assessment validity. Both factors have been studied separately, but DIF and missing data are likely to occur…
Descriptors: Sample Size, Monte Carlo Methods, Test Validity, Effect Size
Peer reviewed Peer reviewed
Direct linkDirect link
Yoo, Jin Eun – Educational and Psychological Measurement, 2009
This Monte Carlo study investigates the beneficiary effect of including auxiliary variables during estimation of confirmatory factor analysis models with multiple imputation. Specifically, it examines the influence of sample size, missing rates, missingness mechanism combinations, missingness types (linear or convex), and the absence or presence…
Descriptors: Monte Carlo Methods, Research Methodology, Test Validity, Factor Analysis
Peer reviewed Peer reviewed
Hattie, John – Multivariate Behavioral Research, 1984
This paper describes a simulation that determines the adequacy of various indices as decision criteria for assessing unidimensionality. Using the sum of absolute residuals from the two-parameter latent trait model, indices were obtained that could discriminate between one latent trait and more than one latent trait. (Author/BW)
Descriptors: Achievement Tests, Latent Trait Theory, Mathematical Models, Monte Carlo Methods
Previous Page | Next Page »
Pages: 1  |  2