Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 42 |
Descriptor
Evaluation Research | 50 |
Program Validation | 50 |
Evaluation Methods | 20 |
Item Analysis | 17 |
Test Validity | 13 |
Psychometrics | 10 |
Research Methodology | 10 |
Test Reliability | 10 |
Foreign Countries | 7 |
Program Evaluation | 7 |
Intervention | 6 |
More ▼ |
Source
Author
Ahn, Jeehae | 1 |
Baartman, Liesbeth K. J. | 1 |
Babarovic, Toni | 1 |
Bacon, Don | 1 |
Benton, Tom | 1 |
Bham, Mohammed | 1 |
Bledsoe, Sarah E. | 1 |
Bogaerts, Stefan | 1 |
Boland, Joseph B. | 1 |
Borgmeier, Christopher J. | 1 |
Bracken, Stacey Storch | 1 |
More ▼ |
Publication Type
Journal Articles | 44 |
Reports - Evaluative | 20 |
Reports - Research | 17 |
Reports - Descriptive | 8 |
Dissertations/Theses -… | 3 |
Information Analyses | 3 |
Books | 1 |
Guides - Non-Classroom | 1 |
Opinion Papers | 1 |
Education Level
Audience
Policymakers | 1 |
Practitioners | 1 |
Researchers | 1 |
Location
Australia | 3 |
California | 1 |
Croatia | 1 |
Hong Kong | 1 |
Pennsylvania | 1 |
United Kingdom | 1 |
United Kingdom (England) | 1 |
United States | 1 |
Laws, Policies, & Programs
Individuals with Disabilities… | 1 |
Assessments and Surveys
Minnesota Multiphasic… | 2 |
California Psychological… | 1 |
Graduate Record Examinations | 1 |
What Works Clearinghouse Rating
What Works Clearinghouse, 2017
The What Works Clearinghouse (WWC) evaluates research studies that look at the effectiveness of education programs, products, practices, and policies, which the WWC calls "interventions." Many studies of education interventions make claims about impacts on students' outcomes. Some studies have designs that enable readers to make causal…
Descriptors: Program Design, Program Development, Program Effectiveness, Program Evaluation
Richer, Amanda; Charmaraman, Linda; Ceder, Ineke – Afterschool Matters, 2018
Like instruments used in afterschool programs to assess children's social and emotional growth or to evaluate staff members' performance, instruments used to evaluate program quality should be free from bias. Practitioners and researchers alike want to know that assessment instruments, whatever their type or intent, treat all people fairly and do…
Descriptors: Cultural Differences, Social Bias, Interrater Reliability, Program Evaluation
Farid, Alem – Electronic Journal of e-Learning, 2014
Although there are tools to assess student's readiness in an "online learning context," little is known about the "psychometric" properties of the tools used or not. A systematic review of 5107 published and unpublished papers identified in a literature search on student online readiness assessment tools between 1990 and…
Descriptors: Online Courses, Electronic Learning, Learning Readiness, Psychometrics
Elbeck, Matt; Bacon, Don – Journal of Education for Business, 2015
The absence of universally accepted definitions for direct and indirect assessment motivates the purpose of this article: to offer definitions that are literature-based and theoretically driven, meeting K. Lewin's (1945) dictum that, "There is nothing so practical as a good theory" (p. 129). The authors synthesize the literature to…
Descriptors: Definitions, Evaluation Methods, Global Approach, Evidence
Ho, Andrew D. – Teachers College Record, 2014
Background/Context: The target of assessment validation is not an assessment but the use of an assessment for a purpose. Although the validation literature often provides examples of assessment purposes, comprehensive reviews of these purposes are rare. Additionally, assessment purposes posed for validation are generally described as discrete and…
Descriptors: Elementary Secondary Education, Standardized Tests, Measurement Objectives, Educational Change
Royal, Kenneth D.; Gilliland, Kurt O.; Kernick, Edward T. – Anatomical Sciences Education, 2014
Any examination that involves moderate to high stakes implications for examinees should be psychometrically sound and legally defensible. Currently, there are two broad and competing families of test theories that are used to score examination data. The majority of instructors outside the high-stakes testing arena rely on classical test theory…
Descriptors: Item Response Theory, Scoring, Evaluation Methods, Anatomy
Johnson, Jeremiah; Hall, Jori; Greene, Jennifer C.; Ahn, Jeehae – American Journal of Evaluation, 2013
Evaluators have an obligation to present clearly the results of their evaluative efforts. Traditionally, such presentations showcase formal written and oral reports, with dispassionate language and graphs, tables, quotes, and vignettes. These traditional forms do not reach all audiences nor are they likely to include the most powerful presentation…
Descriptors: Evaluation Research, Change Strategies, Research Reports, Usability
Mo, Lun; Yang, Fang; Hu, Xiangen – Educational Research and Evaluation, 2011
School climate surveys are widely applied in school districts across the nation to collect information about teacher efficacy, principal leadership, school safety, students' activities, and so forth. They enable school administrators to understand and address many issues on campus when used in conjunction with other student and staff data.…
Descriptors: Evidence, Academic Achievement, Questionnaires, Item Response Theory
Ramineni, Chaitanya; Trapani, Catherine S.; Williamson, David M.; Davey, Tim; Bridgeman, Brent – ETS Research Report Series, 2012
Automated scoring models for the "e-rater"® scoring engine were built and evaluated for the "GRE"® argument and issue-writing tasks. Prompt-specific, generic, and generic with prompt-specific intercept scoring models were built and evaluation statistics such as weighted kappas, Pearson correlations, standardized difference in…
Descriptors: Scoring, Test Scoring Machines, Automation, Models
Baartman, Liesbeth K. J.; Prins, Frans J.; Kirschner, Paul A.; van der Vleuten, Cees P. M. – Evaluation and Program Planning, 2011
The goal of this article is to contribute to the validation of a self-evaluation method, which can be used by schools to evaluate the quality of their Competence Assessment Program (CAP). The outcomes of the self-evaluations of two schools are systematically compared: a novice school with little experience in competence-based education and…
Descriptors: Educational Innovation, Competency Based Education, Self Evaluation (Groups), Program Validation
Garb, Howard N.; Wood, James M.; Fiedler, Edna R. – Assessment, 2011
Using 65 items from a mental health screening questionnaire, the History Opinion Inventory-Revised (HOI-R), the present study compared three strategies of scale construction--(1) internal (based on factor analysis), (2) external (based on empirical performance) and (3) intuitive (based on clinicians' opinion)--to predict whether 203,595 U.S. Air…
Descriptors: Opinions, Mental Health, Test Validity, Measures (Individuals)
Porter, Jennifer Marie – ProQuest LLC, 2010
This research evaluated the inter-rater reliability of the Performance Assessment for California Teachers (PACT). Multiple methods for estimating overall rater consistency include percent agreement and Cohen's Kappa (1960), which yielded discrepancies between rater agreement in terms of whether candidates passed or failed particular PACT rubrics.…
Descriptors: Interrater Reliability, Program Effectiveness, Scoring Rubrics, Item Analysis
Donaldson, Linda Plitt; Shields, Joseph – Research on Social Work Practice, 2009
Contemporary trends in social service delivery systems require human service agencies to engage in greater levels of advocacy to reform structures and protect programs that serve vulnerable populations. Objective: The purpose of this study was to develop an instrument to measure the policy advocacy behavior of nonprofit human service agencies.…
Descriptors: Human Services, Delivery Systems, Measures (Individuals), Social Work
Gilbreath, Brad; Rose, Gail L.; Dietrich, Kim E. – Mentoring & Tutoring: Partnership in Learning, 2008
The purpose of this article is to inform readers about the types of instruments available for assessing and improving mentoring in organizations. Extensive review of the psychological, business and medical literature was conducted to identify commercially published, practitioner-oriented instruments. All of the instruments that were…
Descriptors: Mentors, Psychometrics, Literature Reviews, Evaluation Methods
Bradshaw, Catherine P.; Debnam, Katrina; Koth, Christine W.; Leaf, Philip – Journal of Positive Behavior Interventions, 2009
Schoolwide positive behavioral interventions and supports (SWPBIS) are becoming increasingly popular with schools across the country to help create safer learning environments for students. An important aspect of SWPBIS is the ongoing monitoring and evaluation of implementation fidelity. Although a few measures have been created to assess the…
Descriptors: Interrater Reliability, Positive Reinforcement, Behavior Modification, Program Validation