Publication Date
| In 2026 | 2 |
| Since 2025 | 188 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2889 |
| Since 2007 (last 20 years) | 6174 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Teachers | 480 |
| Practitioners | 358 |
| Researchers | 152 |
| Administrators | 122 |
| Policymakers | 51 |
| Students | 44 |
| Parents | 32 |
| Counselors | 25 |
| Community | 15 |
| Media Staff | 5 |
| Support Staff | 3 |
| More ▼ | |
Location
| Australia | 183 |
| Turkey | 157 |
| California | 133 |
| Canada | 124 |
| New York | 118 |
| United States | 112 |
| Florida | 107 |
| China | 103 |
| Texas | 72 |
| United Kingdom | 72 |
| Japan | 70 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 5 |
| Meets WWC Standards with or without Reservations | 11 |
| Does not meet standards | 8 |
Peer reviewedAllen, Layman E.; And Others – Simulation and Games, 1978
Studies the extent to which the effectiveness of instructional gaming in learning specific mathematical ideas can be increased by incorporating devices that channel learners' attention upon those ideas. In particular, the effect of channeling attention by changing the method of scoring is explored. (Author/CMV)
Descriptors: Achievement Gains, Educational Games, Educational Research, Game Theory
Peer reviewedDuncan, George T.; Milton, E. O. – Psychometrika, 1978
A multiple-answer multiple-choice test is one which offers several alternate choices for each stem and any number of those choices may be considered to be correct. In this article, a class of scoring procedures called the binary class is discussed. (Author/JKS)
Descriptors: Answer Keys, Measurement Techniques, Multiple Choice Tests, Scoring Formulas
Peer reviewedCohen, Arie; Farley, Frank H. – Educational and Psychological Measurement, 1977
Cross-cultural validity studies for psychological instruments may result in overestimation of structure invariance due to some items being scored on more than one scale. This problem, called the common-item effect, is investigated with some data from the literature. (JKS)
Descriptors: Cross Cultural Studies, Factor Analysis, Item Sampling, Multidimensional Scaling
Peer reviewedEvans, Roy; And Others – Educational Research, 1975
Examines three reliability characteristics of the Harris revision of the Goodenough "Draw-a-man" Test when used with five-year-old school entrants. (Author/RK)
Descriptors: Data Analysis, Educational Research, Educational Testing, Freehand Drawing
Peer reviewedBailey, Kent G.; And Others – Journal of Consulting and Clinical Psychology, 1977
Two measures of "breadth" and three measures of "depth" were derived, based on special administration and scoring of the Wechsler Similarities subscale. Factor analyses indicated that breadth and depth can be distinguished statistically, and multiple regression analyses revealed that the derived measures contributed to improved…
Descriptors: Factor Analysis, Intelligence Tests, Multiple Regression Analysis, Predictor Variables
Peer reviewedPryzwansky, Walter B. – Psychology in the Schools, 1977
This article investigates teacher use of the Developmental Test of Visual-Motor Integration (VMI) as a screening instrument with groups of young school-age children. Findings argue for some refinement in the scoring system in order to improve consistency in scoring. (Author)
Descriptors: Group Testing, Individual Testing, Kindergarten Children, Learning Disabilities
Peer reviewedSlogoff, Stephen; Hughes, Francis P. – Journal of Medical Education, 1987
A study of the use of "dangerous answers" as a scoring method for certification examinations in anesthesiology concluded that selection of dangerous answers in multiple-choice tests results from lack of information rather than purposeful action, and that implementation of the scoring method is unjustified and unfairly punitive. (MSE)
Descriptors: Anesthesiology, Certification, Higher Education, Malpractice
Noonan, John V.; Sarvela, Paul D. – Performance and Instruction, 1988
Identifies a number of practical decisions that must be made when designing and developing criterion referenced tests as part of a larger system of computer assisted instruction. The topics discussed include test construction, test security, item presentation, and response capturing and scoring. (19 references) (CLB)
Descriptors: Computer Assisted Instruction, Computer Assisted Testing, Criterion Referenced Tests, Item Banks
Peer reviewedWilcox, Rand R.; Wilcox, Karen Thompson – Journal of Educational Measurement, 1988
Use of latent class models to examine strategies that examinees (92 college students) use for a specific task is illustrated, via a multiple-choice test of spatial ability. Under an answer-until-correct scoring procedure, models representing an improvement over simplistic random guessing are proposed. (SLD)
Descriptors: College Students, Decision Making, Guessing (Tests), Multiple Choice Tests
Peer reviewedMeyers, G. Douglas – Bulletin of the Association for Business Communication, 1988
Claims the primary trait scoring method offers an efficient and effective way to grade student papers. Also suggests primary trait scoring can be useful for commenting on rough drafts, and for evaluating assignments throughout the business communication curriculum. (MS)
Descriptors: Business Communication, Evaluation Criteria, Evaluation Methods, Grading
Peer reviewedLuecht, Richard M. – Educational and Psychological Measurement, 1987
Test Pac, a test scoring and analysis computer program for moderate-sized sample designs using dichotomous response items, performs comprehensive item analyses and multiple reliability estimates. It also performs single-facet generalizability analysis of variance, single-parameter item response theory analyses, test score reporting, and computer…
Descriptors: Computer Assisted Testing, Computer Software, Computer Software Reviews, Item Analysis
Peer reviewedGruijter, Dato N. M. – Journal of Educational Measurement, 1985
To improve on cutoff scores based on absolute standards which may produce an unacceptable number of failures, a compromise is suggested. The compromise draws on the information in the observed score distribution to adjust the standard. Three compromise models developed by Hofstee, Beuk, and De Gruijter are compared. (Author/GDC)
Descriptors: Academic Standards, Comparative Testing, Cutting Scores, Mastery Tests
Peer reviewedSilverstein, A. B. – American Journal of Mental Deficiency, 1986
The means and standard deviations of standard scores on the new Vineland Adaptive Behavior Scales vary considerably from age group to age group in the standardization sample. Thus, different standard scores may reflect the same levels of performance in terms of distance from the mean. (Author/CL)
Descriptors: Adaptive Behavior (of Disabled), Mental Retardation, Scoring, Standardized Tests
Peer reviewedJannarone, Robert J. – Psychometrika, 1986
Conjunctive item response models are introduced such that: (1) sufficient statistics for latent traits are not necessarily additive in item scores; (2) items are not necessarily locally independent; and (3) existing compensatory (additive) item response models including the binomial, Rasch, logistic, and general locally independent model are…
Descriptors: Cognitive Processes, Hypothesis Testing, Latent Trait Theory, Mathematical Models
Peer reviewedDorans, Neil J. – Journal of Educational Measurement, 1986
The analytical decomposition demonstrates how the effects of item characteristics, test properties, individual examinee responses, and rounding rules combine to produce the item deletion effect on the equating/scaling function and candidate scores. The empirical portion of the report illustrates the effects of item deletion on reported score…
Descriptors: Difficulty Level, Equated Scores, Item Analysis, Latent Trait Theory


