NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)1
Since 2006 (last 20 years)2
Laws, Policies, & Programs
Education Consolidation…1
Goals 20001
What Works Clearinghouse Rating
Showing 1 to 15 of 34 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Leonard, Jack – Education Policy Analysis Archives, 2018
This paper introduces the new Massachusetts Performance Assessment for Leaders (PAL) and uses critical policy analysis to re-examine the validity evidence (using the 2014 Standards for Educational and Psychological Testing and a theory of multicultural validity) for the use and interpretation of the PAL in regards to emerging school leadership.…
Descriptors: Performance Based Assessment, Test Validity, High Stakes Tests, School Administration
Peer reviewed Peer reviewed
Direct linkDirect link
Falk, Beverly; Ort, Suzanne Wichterle; Moirs, Katie – Educational Assessment, 2007
This article describes the findings of studies conducted on a large-scale, classroom-based performance assessment of literacy for the early grades designed to provide information that is useful for reporting, as well as teaching. Technical studies found the assessment to be a promising instrument that is reliable and valid. Follow-up studies of…
Descriptors: Program Effectiveness, Performance Based Assessment, Student Evaluation, Evaluation Research
Peer reviewed Peer reviewed
Guion, Robert M. – Educational Measurement: Issues and Practice, 1995
This commentary discusses three essential themes in performance assessment and its scoring. First, scores should mean something. Second, performance scores should permit fair and meaningful comparisons. Third, validity-reducing errors should be minimal. Increased attention to performance assessment may overcome these problems. (SLD)
Descriptors: Educational Assessment, Performance Based Assessment, Scores, Scoring
Messick, Samuel – 1996
The concept of "washback," especially prominent in the field of applied linguistics, refers to the extent to which a test influences teachers and learners to do things they would not otherwise necessarily do. Some writers invoke the notion of washback validity, holding that a test's validity should be gauged by the degree to which it has…
Descriptors: Applied Linguistics, Construct Validity, Criteria, Language Tests
Moore, Alan D.; Young, Suzanne – 1997
As schools move toward performance assessment, there is increasing discussion of using these assessments for accountability purposes. When used for making decisions, performance assessments must meet high standards of validity and reliability. One major source of unreliability in performance assessments is interrater disagreement. In this paper,…
Descriptors: Accountability, Correlation, Elementary Secondary Education, Generalizability Theory
Peer reviewed Peer reviewed
Sykes, Robert C.; Ito, Kyoko; Fitzpatrick, Anne R.; Ercikan, Kadriye – Journal of Educational Measurement, 1997
The five chapters of this report provide resources that deal with the validity, generalizability, comparability, performance standards, and fairness, equity, and bias of performance assessments. The book is written for experienced educational measurement practitioners, although an extensive familiarity with performance assessment is not required.…
Descriptors: Educational Assessment, Measurement Techniques, Performance Based Assessment, Standards
Reckase, Mark D. – 1997
This paper argues that special procedures for constructing assessment tools containing performance assessment tasks are unnecessary and that current test methodology can easily be generalized to complex performance assessment tasks without destroying the desirable characteristics of those tasks. Reasonable statistical requirements for sound…
Descriptors: Educational Assessment, Generalizability Theory, High Stakes Tests, Interrater Reliability
Wiggins, Grant – 1990
Based on material prepared for the California Assessment Program, an argument in favor of authentic assessment is presented, and authentic assessment is contrasted with traditional standardized tests. An assessment is authentic when student performance on intellectual tasks is directly examined. Comparatively, traditional assessment relies on…
Descriptors: Comparative Testing, Cost Effectiveness, Educational Assessment, Elementary Secondary Education
Peer reviewed Peer reviewed
Bullock, Cheryl Davis; DeStefano, Lizanne – Educational Evaluation and Policy Analysis, 1998
The usefulness of the 1992 TSA in reading was studied using interviews from 26 state directors of assessment. Perceptions about TSA credibility and orientation of test components and current use of TSA results were examined. The directors suggested involving more teachers in assessment, modifying descriptors, disseminating results quicker, and…
Descriptors: Achievement Tests, Educational Assessment, Elementary Secondary Education, Performance Based Assessment
Reber, Anne M. – 1995
The Wechsler Intelligence Scale for Children-Third Edition (WISC-III) is an individually administered test of intelligence for assessing children aged 6 through 16 years, 11 months. The WISC-III consists of several subtests, each classified into a verbal or performance scale. The child's performance on these measures is summarized in three…
Descriptors: Children, Intelligence Quotient, Intelligence Tests, Performance Based Assessment
Nweke, Winifred C. – 1993
A crucial, but often forgotten, role of educational assessment is to enhance students' learning. This author advocates that an assessment program designed for student learning differs from assessment for accountability in purpose, test format, measurement type, number, and spread of tests, use of test results, and amount of interval between…
Descriptors: Accountability, Correlation, Educational Assessment, Educational Improvement
Kaufman, Alan S.; And Others – 1994
The reliability and validity of three short forms of the Wechsler Intelligence Scale for Children III (WISC-III) were compared. Each of the short forms was a tetrad composed of two verbal and two performance subtests. The first tetrad was selected based primarily on practical considerations, particularly its brevity to administer and score. The…
Descriptors: Adolescents, Age Differences, Children, Clinical Diagnosis
Lyman, Howard B. – 1998
The first edition of this book was written to give information about testing to people whose work gave them access to test results, but whose training included little or nothing about the use and interpretation of tests. Later editions have been intended for a broader audience as the need for understanding what test scores really mean has…
Descriptors: Educational Testing, Norm Referenced Tests, Performance Based Assessment, Psychometrics
Tobias, Sigmund; Everson, Howard T.; Laitusis, Vytas – 1999
A knowledge monitoring assessment (KMA) was developed and evaluated. The KMA, which evaluates how well students distinguish between what they know and do not know by comparing their knowledge estimates to test performance, is partially performance based and may be group or computer administered and objectively scored. Participants were 462…
Descriptors: Aptitude, Educational Assessment, Gifted, High School Students
Peer reviewed Peer reviewed
Burger, Susan E.; Burger, Donald L. – Educational Measurement: Issues and Practice, 1994
One means of establishing the validity of performance-based assessments is to determine whether the criteria for basic competencies on performance tests are reasonable in relation to expected (nationally normed) age-level performance. Links between performances on three assessments in Michigan and the Comprehensive Tests of Basic Skills illustrate…
Descriptors: Achievement Tests, Age Differences, Competence, Criteria
Previous Page | Next Page ยป
Pages: 1  |  2  |  3