Publication Date
| In 2026 | 3 |
| Since 2025 | 190 |
| Since 2022 (last 5 years) | 1069 |
| Since 2017 (last 10 years) | 2891 |
| Since 2007 (last 20 years) | 6176 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Teachers | 481 |
| Practitioners | 358 |
| Researchers | 153 |
| Administrators | 122 |
| Policymakers | 51 |
| Students | 44 |
| Parents | 32 |
| Counselors | 25 |
| Community | 15 |
| Media Staff | 5 |
| Support Staff | 3 |
| More ▼ | |
Location
| Australia | 183 |
| Turkey | 157 |
| California | 134 |
| Canada | 124 |
| New York | 118 |
| United States | 112 |
| Florida | 107 |
| China | 103 |
| Texas | 72 |
| United Kingdom | 72 |
| Japan | 70 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 5 |
| Meets WWC Standards with or without Reservations | 11 |
| Does not meet standards | 8 |
Peer reviewedAiken, Lewis R. – Educational and Psychological Measurement, 1987
Formulas for transforming scores and statistics on a rating scale having any number of categories to a scale having a different number of categories are described. To illustrate the use of one of the formulas, the means and variances of items on six forms of a course evaluation questionnaire were compared. (Author/LMO)
Descriptors: Mathematical Models, Rating Scales, Scoring, Statistical Studies
Peer reviewedKolen, Michael J. – Journal of Educational Measurement, 1988
Linear and nonlinear methods for incorporating score precision information when the score scale is established for educational tests are compared. Examples illustrate the methods, which discourage overinterpretation of small score differences and enhance score interpretability by equalizing error variance along the score scale. Measurement error…
Descriptors: Error of Measurement, Measures (Individuals), Scaling, Scoring
Peer reviewedBraun, Henry I. – Journal of Educational Statistics, 1988
A statistical experiment was conducted in an operational setting to determine the contributions of different sources of variability to the unreliability scoring of essays and other free-response questions. Partially balanced incomplete block designs facilitated the unbiased estimation of certain main effects without requiring readers to assess the…
Descriptors: Essay Tests, Grading, Reliability, Scoring
Peer reviewedJohnson, William L. – Journal of Experimental Education, 1986
Simplified and complex methods of scoring discrepancy format needs assessment instruments were compared. It was shown that the use of the simplified procedure yielded statistically equivalent results as compared to the use of the more complex procedure. The importance of the format and the scoring techniques is discussed. (Author/JAZ)
Descriptors: Administrators, Correlation, Measurement Techniques, Needs Assessment
Peer reviewedNorton, Linda S.; Hartley, James – Higher Education, 1986
A study of college student study skills and note sources and their relationship to test performance found that the more sources students used, the higher their test scores would be. In addition, students made choices about the sources they use, and some choices led to higher scores, especially those reflectng scorer. (MSE)
Descriptors: Higher Education, Notetaking, Scores, Scoring
Peer reviewedAnastasi, Anne – Journal of Counseling & Development, 1985
Describes the role of information on score reliabilities, significance of score differences, intercorrelations of scores, and differential validity of score patterns on the interpretation of results from multiscore batteries. (Author)
Descriptors: Psychological Testing, Scoring, Test Interpretation, Test Reliability
Peer reviewedNilsson, Ingvar; Wedman, Ingemar – Scandinavian Journal of Educational Research, 1976
States that, due to confusion of concepts and lack of systemization, "previous studies are often difficult to interpret and consequently...afford little possibility of formulating more precise statements about those errors the concepts represent...." A proposal for systematization is presented. (Author/RW)
Descriptors: Bias, Response Style (Tests), Scoring, Test Reliability
Ferdous, Abdullah; Plake, Barbara – 2003
In the Angoff standard setting procedure, subject matter experts (SMEs) estimate the probability that a hypothetical randomly selected minimally competent candidate will answer correctly each item comprising the test. In many cases, these item performance estimates are made twice, with information shared with SMEs between estimates. This…
Descriptors: Cost Effectiveness, Estimation (Mathematics), Standard Setting (Scoring), Test Items
Rudner, Lawrence M. – 2001
This paper describes and evaluates the use of decision theory as a tool for classifying examinees based on their item response patterns. Decision theory, developed by A. Wald (1947) and now widely used in engineering, agriculture, and computing, provides a simple model for the analysis of categorical data. Measurement decision theory requires only…
Descriptors: Classification, Mathematical Models, Measurement Techniques, Responses
Marzano, Robert J. – 2000
There has been little discussion of two conventions common within classroom assessment: the convention of representing student's performance on an assessment using a single score; and the convention of using the average score to summarize a student's performance over a set of assessments. This paper attempts to demonstrate that the assumptions…
Descriptors: Elementary Secondary Education, Scoring, Teacher Made Tests, Test Theory
Moskal, Barbara M. – 2003
This Digest draws from the current literature and the author's experience to identify suggestions for developing performance assessments and their accompanying scoring rubrics. This Digest, part 1, addresses writing goals and objectives and developing performance assessments. Before a performance assessment or scoring rubric is written or…
Descriptors: Measurement Techniques, Performance Based Assessment, Scoring Rubrics, Test Construction
A Comparison of Scoring Modalities for Performance Assessment: A Case of Constructed Response Items.
Lee, Guemin; Sykes, Robert C. – 2000
This study was designed to investigate several scoring modalities in terms of the accuracy of scores and the efficiency of scoring procedures. Five scoring modalities (SM) are conceptualized by considering tasks and raters as sources of error where "p,""t," and "r" represent person, task, and rater, respectively.…
Descriptors: Comparative Analysis, Constructed Response, Performance Based Assessment, Scores
Gerl, Sister Marion Joseph – Journal of Business Education, 1974
The procedures and advantages of the gross speed--two percent-of-error method in scoring typewriting timed writings are presented. The method makes allowance for errors according to the number of opportunities for error. A mailing address for the typing scoring chart and further information on the method is included. (AG)
Descriptors: Business Education, Business Skills, Scoring Formulas, Timed Tests
Peer reviewedHalpin, Gerald; Halpin Glennelle – Psychology in the Schools, 1974
This study attempts to determine the extent of agreement on the Torrance Tests of Creative Thinking scores among independent scorers with no training other than individual study of the scoring guides, and a scorer trained by Torrance. (Author)
Descriptors: Creativity Research, Creativity Tests, Scoring, Statistical Studies
Peer reviewedLord, Frederic M. – Educational and Psychological Measurement, 1973
A group of 21 students was tested under a time limit considerably shorter than should have been allowed. This report describes a tryout of a method for estimating the power'' scores that would have been obtained if the students had had enough time to finish. (Author/CB)
Descriptors: Mathematical Models, Scoring Formulas, Statistical Analysis, Theories


