Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 38 |
| Since 2007 (last 20 years) | 132 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 12 |
| Practitioners | 10 |
| Community | 5 |
| Parents | 5 |
| Teachers | 3 |
| Policymakers | 2 |
Location
| Florida | 7 |
| United Kingdom | 6 |
| United Kingdom (England) | 6 |
| Australia | 5 |
| Canada | 5 |
| United States | 5 |
| Georgia | 3 |
| New York | 3 |
| North Carolina | 3 |
| Turkey | 3 |
| California | 2 |
| More ▼ | |
Laws, Policies, & Programs
| Elementary and Secondary… | 3 |
| No Child Left Behind Act 2001 | 3 |
| Education for All Handicapped… | 1 |
| Individuals with Disabilities… | 1 |
| Serrano v Priest | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedPoizner, Sharon B.; And Others – Applied Psychological Measurement, 1978
Binary, probability, and ordinal scoring procedures for multiple-choice items were examined. In two situations, it was found that both the probability and ordinal scoring systems were more reliable than the binary scoring method. (Author/CTM)
Descriptors: Confidence Testing, Guessing (Tests), Higher Education, Multiple Choice Tests
Herwick, Mary Jo; Regennitter, John F. – Journal of Business Education, 1977
When applicants for the cooperative office education program exceed space and funds, the coordinator must have a sound basis for selecting students. This article presents both fixed and variable selection criteria to enable the coordinator to evaluate applicants with minimum subjectivity. (MF)
Descriptors: Admission Criteria, Computer Oriented Programs, Cooperative Education, Cooperative Programs
Tsujimoto, Richard N.; Berger, Dale E. – Child Abuse and Neglect: The International Journal, 1988
Two criteria are discussed for determining cutting scores on a predictor variable for identifying cases of likely child abuse--utility maximizing and error minimizing. Utility maximizing is the preferable criterion, as it optimizes the balance between the costs of incorrect decisions and the benefits of correct decisions. (Author/JDD)
Descriptors: Child Abuse, Cost Effectiveness, Cutting Scores, Error of Measurement
Peer reviewedLord, Frederic M. – Journal of Educational Measurement, 1984
Four methods are outlined for estimating or approximating from a single test administration the standard error of measurement of number-right test score at specified ability levels or cutting scores. The methods are illustrated and compared on one set of real test data. (Author)
Descriptors: Academic Ability, Cutting Scores, Error of Measurement, Scoring Formulas
Peer reviewedNapier, John D. – Theory and Research in Social Education, 1976
The study examined (1) whether 60 elementary school teachers could score moral thought statements into Kohlberg's moral stages by receiving special training and using a rater manual, and (2) what factors were related to their stage-scoring ability. Major conclusion was that the rater manual and training were ineffective. (Author/ND)
Descriptors: Achievement Rating, Cognitive Processes, Elementary Education, Elementary School Teachers
Peer reviewedStocker, Leonard P. – Reading Improvement, 1971
Suggests approximately 200 words of religious denotation that should be added to the Dale-Chall List of 3,000 Familiar Words when assessing the readability of materials that contain a Catholic vocabulary." (VJ)
Descriptors: Catholic Schools, Readability, Reading Level, Reading Materials
Willingness to Answer Multiple-Choice Questions as Manifested Both in Genuine and in Nonsense Items.
Peer reviewedFrary, Robert B.; Hutchinson, T.P. – Educational and Psychological Measurement, 1982
Alternate versions of Hutchinson's theory were compared, and one which implies the existence of partial knowledge was found to be better than one which implies that an appropriate measure of ability is obtained by applying the conventional correction for guessing. (Author/PN)
Descriptors: Guessing (Tests), Latent Trait Theory, Multiple Choice Tests, Scoring Formulas
Peer reviewedBirenbaum, Menucha; Fatsuoka, Kikumi K. – Journal of Educational Measurement, 1983
The outcomes of two scoring methods (one based on an error analysis and the second on a conventional method) on free-response tests, compared in terms of reliability and dimensionality, indicates the conventional method is inferior in both aspects. (Author/PN)
Descriptors: Achievement Tests, Algorithms, Data, Junior High Schools
Peer reviewedPlake, Barbara S.; And Others – Journal of Experimental Education, 1981
Number right and elimination scores were analyzed on a college level mathematics exam assembled from pretest data. Anxiety measures were administered along with the experimental forms to undergraduates. Results suggest that neither test scores nor attitudes are influenced by item order knowledge thereof, or anxiety level. (Author/GK)
Descriptors: College Mathematics, Difficulty Level, Higher Education, Multiple Choice Tests
Peer reviewedBliss, Leonard B. – Journal of Educational Measurement, 1980
A mathematics achievement test with instructions to avoid guessing wildly was given to 168 elementary school pupils who were later asked to complete all the questions using a differently colored pencil. Results showed examinees, particularly the more able students, tend to omit too many items. (CTM)
Descriptors: Anxiety, Guessing (Tests), Intermediate Grades, Multiple Choice Tests
Bonett, Douglas G. – Applied Psychological Measurement, 2006
Comparing variability of test scores across alternate forms, test conditions, or subpopulations is a fundamental problem in psychometrics. A confidence interval for a ratio of standard deviations is proposed that performs as well as the classic method with normal distributions and performs dramatically better with nonnormal distributions. A simple…
Descriptors: Intervals, Mathematical Concepts, Comparative Analysis, Psychometrics
Ben-Simon, Anat; Bennett, Randy Elliott – Journal of Technology, Learning, and Assessment, 2007
This study evaluated a "substantively driven" method for scoring NAEP writing assessments automatically. The study used variations of an existing commercial program, e-rater[R], to compare the performance of three approaches to automated essay scoring: a "brute-empirical" approach in which variables are selected and weighted solely according to…
Descriptors: Writing Evaluation, Writing Tests, Scoring, Essays
Suhadolnik, Debra; Weiss, David J. – 1983
The present study was an attempt to alleviate some of the difficulties inherent in multiple-choice items by having examinees respond to multiple-choice items in a probabilistic manner. Using this format, examinees are able to respond to each alternative and to provide indications of any partial knowledge they may possess concerning the item. The…
Descriptors: Confidence Testing, Multiple Choice Tests, Probability, Response Style (Tests)
Bishop, Arthur, Ed. – 1978
The decline in writing ability among students entering college can be attributed to such factors as the absence of verbal precision in television, advertising, and political language, the lack of writing practice in high school English classes, overloaded and unprepared English teachers, and the difficulties involved in evaluating student writing.…
Descriptors: College Freshmen, Evaluation Criteria, Holistic Evaluation, National Competency Tests
Jaeger, Richard M. – 1980
Five statistical indices are developed and described which may be used for determining (1) when linear equating of two approximately parallel tests is adequate, and (2) whan a more complex method such as equipercentile equating must be used. The indices were based on: (1) similarity of cumulative score distributions; (2) shape of the raw-score to…
Descriptors: College Entrance Examinations, Difficulty Level, Equated Scores, Higher Education

Direct link
