Descriptor
Test Validity | 19 |
Elementary Secondary Education | 11 |
Validity | 10 |
Evaluation Methods | 9 |
Test Use | 8 |
Test Reliability | 7 |
Reliability | 6 |
Test Bias | 5 |
Testing Problems | 5 |
Performance Based Assessment | 4 |
Standardized Tests | 4 |
More ▼ |
Source
Author
Rudner, Lawrence M. | 3 |
Austin, James T. | 1 |
Beswick, Richard | 1 |
Brescia, William | 1 |
Brualdi, Amy | 1 |
Carlson, Janet F. | 1 |
Childs, Ruth A. | 1 |
Coburn, Louisa | 1 |
Crafts, Jennifer | 1 |
Dayton, C. Mitchell | 1 |
Dunkel, Patricia A. | 1 |
More ▼ |
Publication Type
ERIC Digests in Full Text | 30 |
ERIC Publications | 30 |
Guides - Non-Classroom | 3 |
Reports - Evaluative | 2 |
Opinion Papers | 1 |
Education Level
Audience
Practitioners | 3 |
Administrators | 1 |
Location
Canada | 1 |
South Carolina | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Rafilson, Fred – 1991
An important issue in educational and employment settings is the degree to which evidence of validity obtained in one situation can be generalized to another situation without further study of validity in the new situation. Theory, procedures, and applications concerning validity generalization are addressed. Meta-analytic techniques make possible…
Descriptors: Analysis of Covariance, Concurrent Validity, Correlation, Educational Assessment
Rudner, Lawrence M. – 1994
The "Standards for Educational and Psychological Testing" of the American Educational Research Association, the American Psychological Association, and the National Council on Measurement in Education are intended to provide a comprehensive basis for evaluating tests. This digest identifies key standards applicable to most test…
Descriptors: Ability, Academic Achievement, Evaluation Methods, Norms
Dayton, C. Mitchell – 2002
This Digest, intended as an instructional aid for beginning research students and a refresher for researchers in the field, identifies key factors that play a critical role in determining the credibility that should be given to a specific research study. The needs for empirical research, randomization and control, and significance testing are…
Descriptors: Credibility, Data Analysis, Reliability, Research
Childs, Ruth A.; Jaciw, Andrew P. – 2003
Matrix sampling of test items, the division of a set of items into different versions of a test form, is used by several large-scale testing programs. This Digest discusses nine categories of costs associated with matrix sampling. These categories are: (1) development costs; (2) materials costs; (3) administration costs; (4) educational costs; (5)…
Descriptors: Costs, Matrices, Reliability, Sampling
Brualdi, Amy – 1999
Test validity refers to the degree to which the inferences based on test scores are meaningful, useful, and appropriate. Thus, test validity is a characteristic of a test when it is administered to a particular population. This article introduces the modern concepts of validity advanced by S. Messick (1989, 1996, 1996). Traditionally, the means of…
Descriptors: Criteria, Data Interpretation, Elementary Secondary Education, Reliability
Coburn, Louisa – 1984
Research on student evaluation of college teachers' performance is briefly summarized. Lawrence M. Aleamoni offers four arguments in favor of student ratings: (1) students are the main source of information about the educational environment; (2) students are the most logical evaluators of student satisfaction and effectiveness of course elements;…
Descriptors: College Faculty, Evaluation Problems, Evaluation Utilization, Higher Education
Elliott, Stephen N. – 1995
This digest offers principles of performance assessment as an alternative to norm-referenced tests. The definition of performance assessment developed by the U.S. Congress's Office of Technology and Assessment is given, common features are listed, and the terms "performance" and "authentic" are defined. Suggested guidelines for…
Descriptors: Definitions, Elementary Secondary Education, Evaluation Methods, Guidelines
Powell, Janet L. – 1989
Despite a significant increase in test usage, numerous issues surrounding reading assessment remain unresolved. Construct validity--whether the test actually measures aspects of the behavior under consideration--is of particular importance if one is to rely on test scores to direct instruction, predict performance, or determine accountability. A…
Descriptors: Construct Validity, Elementary Education, Metacognition, Reading Comprehension
Purves, Alan C. – 1990
A synthesis of the report, "The Current State of Assessment in Literature," (produced by the Center for the Learning and Teaching of Literature), this digest discusses methods of evaluating students' knowledge of literature. The digest argues that, by and large, the tests that now exist in the United States do not live up to the…
Descriptors: Critical Reading, Evaluation Criteria, Evaluation Methods, Evaluation Problems
Beswick, Richard – 1990
In this digest, readers are introduced to the scope of instructional program evaluation and evaluators' changing roles in school districts. A program evaluation measures outcomes based on student-attainment goals, implementation levels, and external factors such as budgetary restraints and community support. Instructional program evaluation may be…
Descriptors: Administrator Role, Consultants, Curriculum Evaluation, Elementary Secondary Education
Hills, Tynette Wilson – 1987
Screening programs are now widely used with children who are age-eligible for school entry. Screening is used to identify children who may be at risk of future difficulty in school (e.g., inability to meet academic expectations) and those who may have special needs in learning (e.g., extraordinary abilities and talents or handicapping conditions).…
Descriptors: Early Childhood Education, Identification, Kindergarten, School Entrance Age
Lomawaima, K. Tsianina; McCarty, Teresa L. – 2002
The constructs used to evaluate research quality--valid, objective, reliable, generalizable, randomized, accurate, authentic--are not value-free. They all require human judgment, which is affected inevitably by cultural norms and values. In the case of research involving American Indians and Alaska Natives, assessments of research quality must be…
Descriptors: Action Research, American Indian Education, Educational Research, Indigenous Knowledge
Scriven, Michael – 1995
Student ratings of instruction are widely used as a basis for personnel decisions and faculty development recommendations. This digest discusses concerns about the validity of student ratings and presents a case for their use in teacher evaluation. There are several strong arguments for using student ratings to evaluate teachers. Students are in a…
Descriptors: College Faculty, College Students, Data Collection, Decision Making
ERIC Clearinghouse on Tests, Measurement, and Evaluation, Princeton, NJ. – 1985
This Digest overviews legal challenges in five areas of test use for decision-making in schools: ability tracking, placement in special education classes, test scores as college admissions criteria, test disclosure, and teacher competency testing. Cases illustrating these challenges are described and include: Hobson v. Hansen (1967), Moses v.…
Descriptors: Court Litigation, Educational Testing, Intelligence Tests, Legal Problems
Merz, William R., Sr.; And Others – 1990
The nature of neuropsychological assessment and its application in the school environment are discussed. Neuropsychology is the study of how the brain and nervous system affect thinking and behavior. A complete neuropsychological assessment requires gathering and analyzing information on a child's: (1) physical, social, and psychological…
Descriptors: Children, Diagnostic Tests, Disabilities, Educational Diagnosis
Previous Page | Next Page ยป
Pages: 1 | 2