Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 4 |
Descriptor
Mathematics Tests | 11 |
Standard Setting (Scoring) | 7 |
Cutting Scores | 4 |
Standards | 4 |
Grade 8 | 3 |
Higher Education | 3 |
Item Analysis | 3 |
Probability | 3 |
Test Construction | 3 |
Test Items | 3 |
Algebra | 2 |
More ▼ |
Author
Publication Type
Reports - Research | 6 |
Journal Articles | 5 |
Reports - Evaluative | 4 |
Speeches/Meeting Papers | 3 |
Collected Works - General | 1 |
Non-Print Media | 1 |
Education Level
Grade 8 | 3 |
Elementary Secondary Education | 2 |
Elementary Education | 1 |
Grade 12 | 1 |
Grade 4 | 1 |
Grade 9 | 1 |
High Schools | 1 |
Secondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
General Educational… | 1 |
National Assessment of… | 1 |
What Works Clearinghouse Rating
Ferdous, Abdullah A.; Plake, Barbara S. – Educational and Psychological Measurement, 2008
Even when the scoring of an examination is based on item response theory (IRT), standard-setting methods seldom use this information directly when determining the minimum passing score (MPS) for an examination from an Angoff-based standard-setting study. Often, when IRT scoring is used, the MPS value for a test is converted to an IRT-based theta…
Descriptors: Standard Setting (Scoring), Scoring, Cutting Scores, Item Response Theory
Rodeck, Elaine M.; Chin, Tzu-Yun; Davis, Susan L.; Plake, Barbara S. – Journal of Applied Testing Technology, 2008
This study examined the relationships between the evaluations obtained from standard setting panelists and changes in ratings between different rounds of a standard setting study that involved setting standards on different language versions of an exam. We investigated panelists' evaluations to determine if their perceptions of the standard…
Descriptors: Mathematics Tests, Standard Setting (Scoring), French, Evaluation Research
Ferdous, Abdullah A.; Plake, Barbara S. – Educational and Psychological Measurement, 2007
In an Angoff standard setting procedure, judges estimate the probability that a hypothetical randomly selected minimally competent candidate will answer correctly each item in the test. In many cases, these item performance estimates are made twice, with information shared with the panelists between estimates. Especially for long tests, this…
Descriptors: Test Items, Probability, Item Analysis, Standard Setting (Scoring)
Impara, James C.; Plake, Barbara S. – 2000
This paper reports the results of using several alternative methods of setting cut scores. The methods used were: (1) a variation of the Angoff method (1971); (2) a variation of the borderline group method; and (3) an advanced impact method (G. Dillon, 1996). The results discussed are from studies undertaken to set the cut scores for fourth grade…
Descriptors: Cutting Scores, Intermediate Grades, Mathematics Tests, Scoring Formulas

Plake, Barbara S.; And Others – Educational and Psychological Measurement, 1988
The effect of item context on differential item performance based on gender on mathematics test items was studied, using 404 male and 375 female adults. The analyses were based on a modified one-parameter item response theory methodology. Gender differences emerged; however, they may be due to chance. (TJH)
Descriptors: Achievement Tests, Adults, Latent Trait Theory, Mathematics Tests
Buckendahl, Chad W.; Davis, Susan L.; Plake, Barbara S.; Sireci, Stephen G.; Hambleton, Ronald K.; Zenisky, April L.; Wells, Craig S. – Office of Planning, Evaluation and Policy Development, US Department of Education, 2009
The "Evaluation of the National Assessment of Educational Progress: Study Reports" describes the special studies that comprised the design of the evaluation. In the Final Report, the authors presented a practical discussion of the evaluation studies to its primary, intended audience, namely policymakers. On this accompanying CD, readers…
Descriptors: National Competency Tests, Program Evaluation, Educational Research, Program Effectiveness
Buckendahl, Chad W.; Smith, Russ W.; Impara, James C.; Plake, Barbara S. – 2000
This paper presents a comparison of two commonly used methods, Angoff (W. Angoff, 1971) and Bookmark (D. Lewis, H. Mitzel, and D. Green, 1996), for setting cut scores on selected response tests. This comparison is presented through an application to a grade 7 mathematics assessment in a suburban Midwestern school district. Training and operational…
Descriptors: Comparative Analysis, Cutting Scores, Junior High School Students, Junior High Schools
Plake, Barbara S.; Impara, James C.; Hertzog, Melody; Giraud, Gerald; Spies, Robert – 1997
Judgmental standard setting approaches rely on the perceptions of experts about examinee performance on a test. Traditional standard setting methods ask panelists to predict the probability that a randomly selected, hypothetical minimally competent candidate (MCC) will correctly answer test questions. Item performance predictions are difficult for…
Descriptors: Concept Formation, High School Students, High Schools, Judges
Plake, Barbara S.; Melican, Gerald J. – 1985
A methodology for investigating the influence of correction-for-guessing directions and formula scoring on test performance was studied. Experts in the test content field used a judgmental item appraisal system to estimate the knowledge of the minimally competent candidate (MCC) and to predict those items that the MCC would omit on the test under…
Descriptors: College Students, Guessing (Tests), Higher Education, Mathematics Tests

Plake, Barbara S.; Melican, Gerald J. – Educational and Psychological Measurement, 1989
The impact of overall test length and difficulty on the expert judgments of item performance by the Nedelsky method were studied. Five university-level instructors predicting the performance of minimally competent candidates on a mathematics examination were fairly consistent in their assessments regardless of length or difficulty of the test.…
Descriptors: Difficulty Level, Estimation (Mathematics), Evaluators, Higher Education
Plake, Barbara S.; Wise, Steven L. – 1986
One question regarding the utility of adaptive testing is the effect of individualized item arrangements on examinee test scores. The purpose of this study was to analyze the item difficulty choices by examinees as a function of previous item performance. The examination was a 25-item test of basic algebra skills given to 36 students in an…
Descriptors: Adaptive Testing, Algebra, College Students, Computer Assisted Testing