Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 10 |
| Since 2017 (last 10 years) | 49 |
| Since 2007 (last 20 years) | 145 |
Descriptor
Source
Author
Publication Type
Education Level
Location
| Canada | 10 |
| Australia | 8 |
| Tennessee | 8 |
| United Kingdom | 7 |
| California | 4 |
| Kansas | 4 |
| Massachusetts | 4 |
| New Jersey | 4 |
| United States | 4 |
| Illinois | 3 |
| Michigan | 3 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Melican, Gerald; Thomas, Nancy – 1984
Setting standards for the purpose of certification is frequently performed using judgmental techniques such as the Angoff method. This study was performed to identify types of items that judges find hard to rate accurately, that is, types of items on which examinees perform differently than predicted by the judges. Once identified these item types…
Descriptors: Certification, Cutting Scores, Difficulty Level, Minimum Competency Testing
Peer reviewedKane, Michael T. – Journal of Educational Measurement, 1987
The use of item response theory models for analyzing the results of judgmental standard setting studies (the Angoff technique) for establishing minimum pass levels is discussed. A comparison of three methods indicates the traditional approach may not be best. A procedure based on generalizability theory is suggested. (GDC)
Descriptors: Comparative Analysis, Cutting Scores, Generalizability Theory, Latent Trait Theory
Plake, Barbara S.; Impara, James C.; Irwin, Patrick – 1999
Judgmental standard setting methods, such as the Angoff method (W. Angoff, 1971), use item performance estimates as the basis for determining the minimum passing score (MPS). Therefore the accuracy of these item performance estimates is crucial to the validity of the resulting MPS. Recent researchers (L. Shepard, 1994; J. Impara, 1997) have called…
Descriptors: Cutting Scores, Estimation (Mathematics), Judges, Performance Factors
Peer reviewedMills, Craig N. – Journal of Educational Measurement, 1983
This study compares the results obtained using the Angoff, borderline group, and contrasting groups methods of determining performance standards. Congruent results were obtained from the Angoff and contrasting groups methods for several test forms. Borderline group standards were not similar to standards obtained with other methods. (Author/PN)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Cutting Scores, Standard Setting (Scoring)
Peer reviewedLivingston, Samuel A. – Journal of Educational Measurement, 1982
To set a standard on the "beardedness" test (see TM 507 062) the probability that a student with a specific score will be judged as bearded must be estimated for each test score. To get an unbiased estimate of that probability, a representative sample of students at each test score level must be chosen. (BW)
Descriptors: Cutting Scores, Evaluation Methods, Graduation Requirements, Minimum Competency Testing
Peer reviewedRowley, Glenn L. – Journal of Educational Measurement, 1982
Livingston's (TM 507 218) response to Rowley (TM 507 062) is compared with the original Zieky and Livingston formulation of the Contrasting Groups Method of setting standards. (BW)
Descriptors: Cutting Scores, Evaluation Methods, Graduation Requirements, Minimum Competency Testing
Peer reviewedHamilton, J. S.; McLone, R. R. – Studies in Educational Evaluation, 1989
Influences on the educational validity of examinations are reviewed. Changes occurring in approaches to standard setting are traced. A view of reliability is presented, with emphasis on assessment of project work, which often involves individual investigation and design by students. A consistency index formula for grading standards is presented.…
Descriptors: Cutting Scores, Educational Assessment, Elementary Secondary Education, Standard Setting (Scoring)
Peer reviewedCahan, Sorel; Cohen, Nora – Educational and Psychological Measurement, 1990
A solution is offered to problems associated with the inequality in the manipulability of probabilities of classification errors of masters versus nonmasters, based on competency test results. Eschewing the typical arbitrary establishment of observed-score standards below 100 percent, the solution incorporates a self-correction of wrong answers.…
Descriptors: Classification, Error of Measurement, Mastery Tests, Minimum Competency Testing
Peer reviewedBourque, Mary Lyn; Hambleton, Ronald K. – Measurement and Evaluation in Counseling and Development, 1993
Notes that the methods used to set standards for National Assessment of Education Progress (NAEP) tests suggest recommendations for state-level policymakers. Explains the national assessment, basic assumptions in setting performance standards on NAEP, selection of judges, standard-setting methodology for NAEP, and measurement issues in setting…
Descriptors: Elementary Secondary Education, National Norms, Standard Setting (Scoring), Standards
Peer reviewedGoodwin, Laura D. – Applied Measurement in Education, 1999
The relations between Angoff ratings (minimum passing levels) and the actual "p" values for borderline examinees were studied with 115 examinees taking the Certified Financial Planner examination. Findings do not suggest that the Angoff judges' task is nearly impossible, but they do suggest the need to improve standard-setting…
Descriptors: Cutting Scores, Difficulty Level, Judges, Licensing Examinations (Professions)
Peer reviewedReckase, Mark D. – Applied Measurement in Education, 1998
The process used to identify the provisional locations of points that would inform the National Assessment Governing Board in setting achievement levels for the science National Assessment of Educational Progress (NAEP) is described. This process includes identification of panelists for achievement-level setting and the method for converting…
Descriptors: Academic Achievement, Achievement Tests, Elementary Secondary Education, National Surveys
National High School Center, 2007
This issue brief outlines the immediate need for more accurate dropout and graduation data, while providing a snapshot of work currently underway. By drawing on two prominent methods for calculating graduation rates: the National Governors Association's endorsed longitudinal approach and the Averaged Freshman Graduation Rate (AFGR), this brief…
Descriptors: Graduation Rate, Dropouts, Dropout Rate, Data Collection
Lunz, Mary E. – 1997
This paper explains the multifacet technology for analyzing performance examinations and the fair average method of setting criterion standards. The multidimensional nature of performance examinations requires that multiple and often different facets elements of a candidate's examination form be accounted for in the analysis. After this is…
Descriptors: Ability, Computer Assisted Testing, Criteria, Educational Technology
Hambleton, Ronald K. – 1983
This paper offers answers to nine important questions concerning standard-setting issues and methods: (1) Should normative or content-referenced standards be used? (2) Different standard-setting methods yield different results. What is your reaction to this finding? That of your clients? Does this finding present a problem for the application of…
Descriptors: Academic Standards, Decision Making, Evaluation Methods, Evaluation Needs
Meskauskas, John A. – 1983
This paper presents the author's responses to nine questions addressed to the standard-setting symposium participants. A central theme in this assessment is that one cannot clearly demonstrate, a priori, that either a normative or a content-referenced approach is superior. Most methods used to date have been insufficiently grounded in…
Descriptors: Evaluation Criteria, Evaluation Methods, Research Methodology, Research Needs


