Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 8 |
Descriptor
Probability | 14 |
Scores | 14 |
Test Reliability | 14 |
Multiple Choice Tests | 5 |
Statistical Analysis | 5 |
Test Construction | 4 |
Test Validity | 4 |
Classification | 3 |
College Students | 3 |
Correlation | 3 |
Criterion Referenced Tests | 3 |
More ▼ |
Source
Author
Publication Type
Reports - Research | 11 |
Journal Articles | 8 |
Collected Works - Proceedings | 1 |
Reports - Evaluative | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 5 |
Postsecondary Education | 4 |
Elementary Secondary Education | 1 |
Secondary Education | 1 |
Audience
Location
Asia | 1 |
Australia | 1 |
Brazil | 1 |
Colorado | 1 |
Connecticut | 1 |
Denmark | 1 |
Egypt | 1 |
Estonia | 1 |
Florida | 1 |
Germany | 1 |
Greece | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Metsämuuronen, Jari – Practical Assessment, Research & Evaluation, 2022
The reliability of a test score is usually underestimated and the deflation may be profound, 0.40 - 0.60 units of reliability or 46 - 71%. Eight root sources of the deflation are discussed and quantified by a simulation with 1,440 real-world datasets: (1) errors in the measurement modelling, (2) inefficiency in the estimator of reliability within…
Descriptors: Test Reliability, Scores, Test Items, Correlation
Anderson, Darcie L.; Hooks, Tisha – Journal of College Student Retention: Research, Theory & Practice, 2022
With limited budgets and increasing enrollment demands, colleges need fast, free, and practical solutions supporting academic success and retention. The Academic Reality Check (ARC) tool helps to predict traditional freshmen awareness of their own academic expectations in college quickly, supporting the financial investment being made by all…
Descriptors: College Freshmen, Expectation, Predictor Variables, Academic Achievement
Kalinowski, Steven T.; Willoughby, Shannon – Journal of Research in Science Teaching, 2019
We present a multiple-choice test, the Montana State University Formal Reasoning Test (FORT), to assess college students' scientific reasoning ability. The test defines scientific reasoning to be equivalent to formal operational reasoning. It contains 20 questions divided evenly among five types of problems: control of variables, hypothesis…
Descriptors: Science Tests, Test Construction, Science Instruction, Introductory Courses
Wells, Kevin Eugene; Morgan, Grant; Worrell, Frank C.; Sumnall, Harry; McKay, Michael Thomas – International Journal of Behavioral Development, 2018
The goal of the present study is to examine the stability of time attitudes profiles across a one-year period as well as the association between time attitudes profiles and several variables. These variables include attitudes towards alcohol, context of alcohol use, consumption of a full drink, and subjective life expectancy. We assessed the…
Descriptors: Time, Attitude Measures, Drinking, Context Effect
Sadaghiani, Homeyra R.; Pollock, Steven J. – Physical Review Special Topics - Physics Education Research, 2015
As part of an ongoing investigation of students' learning in first semester upper-division quantum mechanics, we needed a high-quality conceptual assessment instrument for comparing outcomes of different curricular approaches. The process of developing such a tool started with converting a preliminary version of a 14-item open-ended quantum…
Descriptors: Science Instruction, Quantum Mechanics, Mechanics (Physics), Multiple Choice Tests
Stewart, Jeffrey; White, David A. – TESOL Quarterly: A Journal for Teachers of English to Speakers of Other Languages and of Standard English as a Second Dialect, 2011
Multiple-choice tests such as the Vocabulary Levels Test (VLT) are often viewed as a preferable estimator of vocabulary knowledge when compared to yes/no checklists, because self-reporting tests introduce the possibility of students overreporting or underreporting scores. However, multiple-choice tests have their own unique disadvantages. It has…
Descriptors: Guessing (Tests), Scoring Formulas, Multiple Choice Tests, Test Reliability
Muller, Jorg M. – Educational and Psychological Measurement, 2006
A new test index is defined as the probability of obtaining two randomly selected test scores (PDTS) as statistically different. After giving a concept definition of the test index, two simulation studies are presented. The first analyzes the influence of the distribution of test scores, test reliability, and sample size on PDTS within classical…
Descriptors: Test Reliability, Probability, Scores, Item Response Theory

Wilcox, Rand R. – Educational and Psychological Measurement, 1979
The classical estimate of a binomial probability function is to estimate its mean in the usual manner and to substitute the results in the appropriate expression. Two alternative estimation procedures are described and examined. Emphasis is given to the single administration estimate of the mastery test reliability. (Author/CTM)
Descriptors: Cutting Scores, Mastery Tests, Probability, Scores

Hansen, Richard – Journal of Educational Measurement, 1971
The relationship between certain personality variables and the degree to which examines display certainty in their responses was investigated. (Author)
Descriptors: Guessing (Tests), Individual Characteristics, Multiple Choice Tests, Personality Assessment
Dimitrov, Dimiter M. – 1996
A Monte Carlo approach is proposed, using the Statistical Analysis System (SAS) programming language, for estimating reliability coefficients in generalizability theory studies. Test scores are generated by a probabilistic model that considers the probability for a person with a given ability score to answer an item with a given difficulty…
Descriptors: Classification, Criterion Referenced Tests, Cutting Scores, Estimation (Mathematics)
Livingston, Samuel A. – 1976
A distinction is made between reliability of measurement and reliability of classification; the "criterion-referenced reliability coefficient" describes the former. Application of this coefficient to the probability distribution of possible scores for a single student yields a meaningful way to describe the reliability of a single score. (Author)
Descriptors: Classification, Criterion Referenced Tests, Error of Measurement, Measurement
Levine, Michael V.; Rubin, Donald B. – 1976
Appropriateness indexes (statistical formulas) for detecting suspiciously high or low scores on aptitude tests were presented, based on a simulation of the Scholastic Aptitude Test (SAT) with 3,000 simulated scores--2,800 normal and 200 suspicious. The traditional index--marginal probability--uses a model for the normal examinee's test-taking…
Descriptors: Academic Ability, Aptitude Tests, College Entrance Examinations, High Schools
Millman, Jason – 1974
This chapter should not only acquaint the reader with the present state of the art on Criterion-Referenced (CR) measurement but also suggest possible directions for further inquiry. The goal of the first part of this chapter is to deal with the definitional dilemma of CR measurement by proceeding from the more traditional view of CR measurement to…
Descriptors: Analysis of Variance, Bayesian Statistics, Behavioral Objectives, Comparative Analysis
International Association for Development of the Information Society, 2012
The IADIS CELDA 2012 Conference intention was to address the main issues concerned with evolving learning processes and supporting pedagogies and applications in the digital age. There had been advances in both cognitive psychology and computing that have affected the educational arena. The convergence of these two disciplines is increasing at a…
Descriptors: Academic Achievement, Academic Persistence, Academic Support Services, Access to Computers