NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Researchers3
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 38 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Vinay Kumar Yadav; Shakti Prasad – Measurement: Interdisciplinary Research and Perspectives, 2024
In sample survey analysis, accurate population mean estimation is an important task, but traditional approaches frequently ignore the intricacies of real-world data, leading to biassed results. In order to handle uncertainties, indeterminacies, and ambiguity, this work presents an innovative approach based on neutrosophic statistics. We proposed…
Descriptors: Sampling, Statistical Bias, Predictor Variables, Predictive Measurement
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Donoghue, John R.; McClellan, Catherine A.; Hess, Melinda R. – ETS Research Report Series, 2022
When constructed-response items are administered for a second time, it is necessary to evaluate whether the current Time B administration's raters have drifted from the scoring of the original administration at Time A. To study this, Time A papers are sampled and rescored by Time B scorers. Commonly the scores are compared using the proportion of…
Descriptors: Item Response Theory, Test Construction, Scoring, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Xiao, Leifeng; Hau, Kit-Tai – Educational and Psychological Measurement, 2023
We examined the performance of coefficient alpha and its potential competitors (ordinal alpha, omega total, Revelle's omega total [omega RT], omega hierarchical [omega h], greatest lower bound [GLB], and coefficient "H") with continuous and discrete data having different types of non-normality. Results showed the estimation bias was…
Descriptors: Statistical Bias, Statistical Analysis, Likert Scales, Statistical Distributions
Peer reviewed Peer reviewed
Direct linkDirect link
Giada Spaccapanico Proietti; Mariagiulia Matteucci; Stefania Mignani; Bernard P. Veldkamp – Journal of Educational and Behavioral Statistics, 2024
Classical automated test assembly (ATA) methods assume fixed and known coefficients for the constraints and the objective function. This hypothesis is not true for the estimates of item response theory parameters, which are crucial elements in test assembly classical models. To account for uncertainty in ATA, we propose a chance-constrained…
Descriptors: Automation, Computer Assisted Testing, Ambiguity (Context), Item Response Theory
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Silva, R. M.; Guan, Y.; Swartz, T. B. – Journal on Efficiency and Responsibility in Education and Science, 2017
This paper attempts to bridge the gap between classical test theory and item response theory. It is demonstrated that the familiar and popular statistics used in classical test theory can be translated into a Bayesian framework where all of the advantages of the Bayesian paradigm can be realized. In particular, prior opinion can be introduced and…
Descriptors: Item Response Theory, Bayesian Statistics, Test Construction, Markov Processes
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Abed, Eman Rasmi; Al-Absi, Mohammad Mustafa; Abu shindi, Yousef Abdelqader – International Education Studies, 2016
The purpose of the present study is developing a test to measure the numerical ability for students of education. The sample of the study consisted of (504) students from 8 universities in Jordan. The final draft of the test contains 45 items distributed among 5 dimensions. The results revealed that acceptable psychometric properties of the test;…
Descriptors: Foreign Countries, Item Response Theory, Numeracy, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Bramley, Tom – Research in Mathematics Education, 2017
This study compared models of assessment structure for achieving differentiation across the range of examinee attainment in the General Certificate of Secondary Education (GCSE) examination taken by 16-year-olds in England. The focus was on the "adjacent levels" model, where papers are targeted at three specific non-overlapping ranges of…
Descriptors: Foreign Countries, Mathematics Education, Student Certification, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Al-Shara'H, Nayel Darweesh – Education, 2013
The study aimed at investigating Jordanian EFL teachers' self-reported frequencies of using the procedures of preparing, correcting, analyzing, interpreting an achievement test, and discussing its results with students. To achieve this, a 31-item questionnaire was used. The questionnaire was administered to 118 basic stage EFL teachers after…
Descriptors: Foreign Countries, English (Second Language), Second Language Instruction, Test Construction
Thompson, Bruce; Melancon, Janet G. – 1996
This study investigated the benefits of creating item "testlets" or "parcels" in the context of structural equation modeling confirmatory factor analysis (CFA). Testlets are defined as groups of items related to a single content area that is developed as a unit. The strategy is illustrated using data from the administration of…
Descriptors: Statistical Distributions, Structural Equation Models, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Kirnan, Jean Powell; Edler, Erin; Carpenter, Allison – International Journal of Testing, 2007
The range of response options has been shown to influence the answers given in self-report instruments that measure behaviors ranging from television viewing to sexual partners. The current research extends this line of inquiry to 36 quantitative items extracted from a biographical inventory used in personnel selection. A total of 92…
Descriptors: Personnel Selection, Biographical Inventories, Testing, Self Disclosure (Individuals)
Matthews-Lopez, Joy L.; Hombo, Catherine M. – 2001
The purpose of this study was to examine the recovery of item parameters in simulated Automatic Item Generation (AIG) conditions, using Markov chain Monte Carlo (MCMC) estimation methods to attempt to recover the generating distributions. To do this, variability in item and ability parameters was manipulated. Realistic AIG conditions were…
Descriptors: Estimation (Mathematics), Monte Carlo Methods, Statistical Distributions, Test Construction
Peer reviewed Peer reviewed
de Gruijter, Dato N. M. – Journal of Educational Measurement, 1997
K. May and W. A. Nicewander recently concluded (1994) that percentile ranks are inferior or raw scores as indicators of latent ability. It is argued that their conclusions are incorrect, and an error in their derivation is identified. The incorrect equation results in an incorrect conclusion, as work by F. M. Lord (1980) also indicates.…
Descriptors: Equations (Mathematics), Estimation (Mathematics), Raw Scores, Statistical Distributions
Peer reviewed Peer reviewed
Emons, Wilco H. M.; Meijer, Rob R.; Sijtsma, Klaas – Applied Psychological Measurement, 2002
Studied whether the theoretical sampling distribution of the U3 person-fit statistic is in agreement with the simulated sampling distribution under different item response theory models and varying item and test characteristics. Simulation results suggest that the use of standard normal deviates for the standardized version of the U3 statistic may…
Descriptors: Item Response Theory, Sampling, Simulation, Statistical Distributions
Peer reviewed Peer reviewed
Enders, Craig K.; Bandalos, Deborah L. – Applied Measurement in Education, 1999
Examined the degree to which coefficient alpha is affected by including items with different distribution shapes within a unidimensional scale. Computer simulation results indicate that reliability does not increase dramatically as a result of using differentially shaped items within a scale. Discusses implications for test construction. (SLD)
Descriptors: Computer Simulation, Reliability, Scaling, Statistical Distributions
Peer reviewed Peer reviewed
Piedmont, Ralph L.; Hyland, Michael E. – Educational and Psychological Measurement, 1993
The use of mean inter-item correlation as a technique for examining homogeneity is proposed as a descriptive tool that can orient researchers to salient aspects of their scales. A study of 341 undergraduates who completed the NEO Personality Inventory illustrates the technique. (SLD)
Descriptors: Correlation, Evaluation Methods, Higher Education, Personality Measures
Previous Page | Next Page ยป
Pages: 1  |  2  |  3