NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Yunxiao; Lee, Yi-Hsuan; Li, Xiaoou – Journal of Educational and Behavioral Statistics, 2022
In standardized educational testing, test items are reused in multiple test administrations. To ensure the validity of test scores, the psychometric properties of items should remain unchanged over time. In this article, we consider the sequential monitoring of test items, in particular, the detection of abrupt changes to their psychometric…
Descriptors: Standardized Tests, Test Items, Test Validity, Scores
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mahmud, Jumailiyah; Sutikno, Muzayanah; Naga, Dali S. – Educational Research and Reviews, 2016
The aim of this study is to determine variance difference between maximum likelihood and expected A posteriori estimation methods viewed from number of test items of aptitude test. The variance presents an accuracy generated by both maximum likelihood and Bayes estimation methods. The test consists of three subtests, each with 40 multiple-choice…
Descriptors: Maximum Likelihood Statistics, Computation, Item Response Theory, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Scherer, Ronny; Meßinger-Koppelt, Jenny; Tiemann, Rüdiger – International Journal of STEM Education, 2014
Background: Complex problem-solving competence is regarded as a key construct in science education. But due to the necessity of using interactive and intransparent assessment procedures, appropriate measures of the construct are rare. This paper consequently presents the development and validation of a computer-based problem-solving environment,…
Descriptors: Computer Assisted Testing, Problem Solving, Chemistry, Science Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Young, Charles; Campbell, Megan – British Journal of Guidance & Counselling, 2014
This article provides GP-CORE norms for a South African university sample, which are compared to published data obtained from a United Kingdom university sample. The measure appears to be both reliable and valid for this multilingual and multicultural South African sample. The profiles of the psychological distress reported by white South African…
Descriptors: Foreign Countries, Well Being, Comparative Analysis, Psychological Needs
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zwick, Rebecca – ETS Research Report Series, 2012
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. The goal of this project was to review the status of ETS DIF analysis procedures, focusing on three aspects: (a) the nature and stringency of the statistical rules used to flag items, (b) the minimum sample size…
Descriptors: Test Bias, Sample Size, Bayesian Statistics, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Skorupski, William P.; Carvajal, Jorge – Educational and Psychological Measurement, 2010
This study is an evaluation of the psychometric issues associated with estimating objective level scores, often referred to as "subscores." The article begins by introducing the concepts of reliability and validity for subscores from statewide achievement tests. These issues are discussed with reference to popular scaling techniques, classical…
Descriptors: Testing Programs, Test Validity, Achievement Tests, Scores
Martin, John T.; And Others – 1983
A conventional verbal ability test and a Bayesian adaptive verbal ability test were compared using a variety of psychometric criteria. Tests were administered to 550 Marine recruits, half of whom received two 30-item alternate forms of a conventional test and half of whom received two 30-item alternate forms of a Bayesian adaptive test. Both types…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Individual Testing
Pine, Steven M.; Weiss, David J. – 1978
This report examines how selection fairness is influenced by the characteristics of a selection instrument in terms of its distribution of item difficulties, level of item discrimination, degree of item bias, and testing strategy. Computer simulation was used in the administration of either a conventional or Bayesian adaptive ability test to a…
Descriptors: Adaptive Testing, Bayesian Statistics, Comparative Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Wilcox, Rand R. – Journal of Educational Statistics, 1977
False-positive and false-negative decisions are the two possible errors committed with a mastery test; yet the estimation of the likelihood of committing these errors has not been investigated. Two methods of this type of estimation are presented and discussed. (Author/JKS)
Descriptors: Bayesian Statistics, Hypothesis Testing, Mastery Tests, Measurement Techniques
Peer reviewed Peer reviewed
Jensema, Carl J. – Educational and Psychological Measurement, 1974
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Instruction, Computer Programs
Peer reviewed Peer reviewed
Helvey, T. Charles – Journal of Experimental Education, 1975
This article describes a new testing method which can be used to screen learning-deficient children fast, reliably, and inexpensively out of any population of public school systems. (Editor)
Descriptors: Bayesian Statistics, Electroencephalography, Error of Measurement, Intelligence Tests
Weiss, David J., Ed. – 1977
This symposium consists of five papers and presents some recent developments in adaptive testing which have applications to several military testing problems. The overview, by James R. McBride, defines adaptive testing and discusses some of its item selection and scoring strategies. Item response theory, or item characteristic curve theory, is…
Descriptors: Ability, Achievement Tests, Adaptive Testing, Bayesian Statistics
Peer reviewed Peer reviewed
Jensema, Carl J. – Applied Psychological Measurement, 1977
Owen's Bayesian tailored testing method is introduced along with a brief review of its derivation. The characteristics of a good item bank are outlined and explored in terms of their influence on the Bayesian tailoring process. (Author/RC)
Descriptors: Ability, Adaptive Testing, Bayesian Statistics, Computer Oriented Programs
Peer reviewed Peer reviewed
Braun, Henry I.; And Others – Psychometrika, 1983
Empirical Bayes methods are shown to provide a practical alternative to standard least squares methods in fitting high dimensional models to sparse data. An example concerning prediction bias in educational testing is presented as an illustration. (Author)
Descriptors: Bayesian Statistics, Educational Testing, Goodness of Fit, Mathematical Models
PDF pending restoration PDF pending restoration
Civil Service Commission, Washington, DC. Personnel Research and Development Center. – 1976
This pamphlet reprints three papers and an invited discussion of them, read at a Division 5 Symposium at the 1975 American Psychological Association Convention. The first paper describes a Bayesian tailored testing process and shows how it demonstrates the importance of using test items with high discrimination, low guessing probability, and a…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Oriented Programs, Computer Programs