Evaluation: Where We’ve Been and Where We’re Going 2014 AmeriCorps State and National Symposium Presenters • Carla Ganiel, Senior Program & Project Specialist, AmeriCorps State and National • Emily Steinberg, Associate Director, AmeriCorps*Texas, OneStar Foundation • Stephen Plank, Director of Research and Evaluation, CNCS • Bethanne Barnes, Program Examiner, Office of Management and Budget AmeriCorps State and National: Lessons Learned Carla Ganiel Senior Program & Project Specialist [email protected] Strategies for Building Evidence • Include evidence in assessment criteria for competitive applicants • Ensure that grantees/subgrantees are on track with evaluation plans • Build evaluation capacity through training and technical assistance for national grantees, state commissions and state subgrantees Grant Application Review Process (GARP) • Levels of evidence outlined in NOFO • Points awarded based on strength of evidence – Strong evidence = 8 points – No evidence = 0 points • Evaluation reports reviewed concurrent with staff review and informed level of evidence assessment Competitive Grantees by Level of Evidence Level Number of New/Recompete Grantees Percent of New/Recompete Grantees Strong 7 7% Moderate 21 22% Preliminary 26 28% Pre-Preliminary 39 41% No Data 1 1% Evaluation Reports • Reviewed by CNCS Office of Research & Evaluation • Staff did not review reports from applicants that were not required to submit an evaluation • Evaluations were not scored • Evaluation results were considered in determining an applicant’s level of evidence Evaluation Review Reviewers assessed: • Alignment of Models • Methodological Quality • Strength of Findings • Recency • Compliance with ASN evaluation requirements Large Applicants ($500,000 +) Large Applicants Total evaluations reviewed Number 10 Met evaluation requirements (type of study) 3 Rated satisfactory quality 4 Met evaluation requirements AND rated satisfactory quality 2 Small Applicants (Less than $500,000) Small Applicants Total evaluations reviewed Number 31 Met evaluation requirements (type of study) 31 Rated satisfactory quality 12 Met evaluation requirements AND rated satisfactory quality 12 Evaluation Reports Submitted Type of Study # Large # Small Applicants Applicants Total Experimental 3 4 7 Quasi-Experimental 0 4 4 Non-Experimental 5 6 11 Process/Implementation 2 17 19 10 31 41 Total High Quality Evaluations Quality Evaluations by Type Experimental 43% Quasi-Experimental 21% Non-Experimental 7% Process/Implementation 29% Lessons Learned • Less than a third of the portfolio has moderate or strong evidence • Few evaluations met both ASN evaluation requirements and quality standards • Some small applicants are conducting impact evaluations • More in-depth review of evidence by external experts, focused on quality rather than quantity of studies Evaluation Technical Assistance • One-on-one technical assistance provided to large ($500,000+) national grantees and state subgrantees • TA focused on developing a plan to conduct an evaluation that would meet CNCS requirements – Quasi-Experimental Design (Comparison Group) – Randomized Control Trial (Control Group) Lessons Learned • Lack of clarity about the difference between performance measurement and evaluation • Many grantees/subgrantees did not understand CNCS evaluation requirements • Some grantees did not take full advantage of TA offered Lessons Learned • Grantees/subgrantees need: – Clear timeline for completing an evaluation with in a three-year cycle – Guidance on budgeting for evaluation – Guidance on how to hire an external evaluator Lessons Learned • At the time of grant award, most grantees have considerably more planning to do before they can meet CNCS evaluation requirements • Certain program designs face challenges when using a quasi-experimental or experimental design • Alternative evaluation approaches may be acceptable for some grantees (Pilot process) • Conversation must shift from “evaluation as compliance” to “evaluation to strengthen program and build evidence.” Evaluation Core Curriculum • Completed courses: – – – – – – – AmeriCorps Evaluation Requirements Evaluation 101 Basic Steps in Conducting an Evaluation Overview of Evaluation Designs How to Develop a Program Logic Model How to Write an Evaluation Plan How to Budget for an Evaluation https://www.nationalserviceresources.gov/evaluationamericorps Next Steps • Alternative Evaluation Approach Pilot Process • Feedback to all small 2014 competitive grants who were required to submit evaluation plans • One on one TA for large 2014 competitive grantees who were required to submit evaluation plans • Dialogue with commissions on how to build capacity to work with subgrantees • Develop additional core curriculum courses • Identify needs of tribal grantees CNCS Grantee Symposium September 17, 2014 Emily Steinberg, Director of National Service Programs 20 • Why evaluation matters – Accountability – Evaluative Learning • Our evaluative evolution – Then & Now – Capacity Building Evaluations (Other State/Federal Projects) – Direct Service Evaluations (AmeriCorps) – Lessons Learned • Current philosophy and approach DUAL ROLE Funder Texas State Service Commission 22 Nonprofit Governor’s Office of FaithBased and Community Initiatives $8-13 million in AmeriCorps grants/year Think Tank/Incubator Responsibility to evaluate grantees and portfolio Responsibility to evaluate self and initiatives Federal reporting requirements State/private/internal reporting requirements What is evaluation? … a method for determining what change occurred, how, why, and to whom a given change has happened. 23 EVALUATION IS… Important for: • Accountability/Stewardship • Learning about Impact – Driving Program Design • Communicating Impact – to staff, board, volunteers, funders and community stakeholders • Multiplying your Impact – increasing the knowledge-base of the field ACCOUNTABILITY VS. LEARNING Accountability • Did what you said you’d do? • Implemented as designed? • Cost-benefit analysis – R.O.I. Evaluative Learning • Game film vs. scoreboard • What outcomes were achieved and for whom? Did outcomes differ across sub-groups? • What made the difference? Which specific program aspects achieved the most/least impact? Peter York, TCC Group LEARNING AND LEADERSHIP • Learning is key to leadership: • Good leaders are good learners:* – Only one in four nonprofit orgs are “well led” – Only one in four nonprofits are effective “learners” *York, P. (2010). “Evaluative Learning.” TCC Group: Philadelphia, PA. HISTORY Our Evaluation Trajectory • 2003-04: OneStar became 501(c)(3) commission • 2004-2008: Blissful ignorance / empty enforcement – Began implementing National PMs – Focused more on capacity building initiatives & evaluation – Did not have a framework for reviewing AmeriCorps subgrantees’ evaluation plans and final reports (honor system) HISTORY Our Evaluation Trajectory… • 2006-2007: Compassion Capital Fund (federal) – FBCI: Focused on grants to small FBCOs ($25K) – Results: Some orgs performed better than others – WHY? • 2007-2008: Rural Texas Demonstration Project (state) – Applied learning to small FBCOs in rural Texas ($10-12K) – Results: Developed indicators of success and critical milestones: readiness to change, trust/rapport, minimum capacity HISTORY Our Evaluation Trajectory • 2008: FBCI Capacity Building Project (State) – Statewide grantees, honed evaluation and success indicators – Studied Org Assessment Tools: • Selected Core Capacity Assessment Tool (CCAT) by TCC-Group • 2009-2012: AmeriCorps Statewide Evaluation (State) STATEWIDE EVALUATION OVERVIEW • Multi-year (2009-2012) • Competitive RFP – Selected external evaluator: UT-Austin RGK Center • Based on previous research – Longitudinal AmeriCorps study by Abt Associates – Organizational capacity assessments – TCC and RGK • Instruments – Program director surveys – AmeriCorps member surveys – Organizational capacity surveys EVALUATION DESIGN PROPOSAL: analyze the value added of AmeriCorps to TX assess organizational/management structures most commonly associated with impactful, value-adding programs assess the different kinds of impact that AmeriCorps programs have on communities EVALUATION MODEL Value Added of AmeriCorps in Communities Organizational Structure and Program Management Characteristics Member Assessment of Community Impact and Value EVALUATION COMPONENTS Surveys AmeriCorps Program Managers AmeriCorps Members Database Organizational & Financial Indicators • Organizational Capacity Surveys • Grantee Budgets • 990 Forms • Organizational Annual Budgets/Financial Reports Case Studies Conducted four (4) site visits FINDING #1: Overall, AmeriCorps Service Has a Positive Impact Percent survey respondents choosing top response Program Managers 2009-10 2010-11 Members Program/service is “very effective” 79% 83% 63% Members “make an important contribution” 96% 91% 75% “A lot of change” is observed in the clients served 67% 70% 51% Finding #2: Most Programs Achieve Goals Percent Meeting Reported Goals by Organization Type Met SOME Reported Goals Met MOST Reported Goals Met ALL Reported Goals 37 50 67 71 100 25 33 100 56 29 25 7 State Agency Local Education Other Local Gov't Community-based 4-Yr College/Univ Agency Agency Org Other • 83% of 2009-10 and 91% of 2010-11 respondents report meeting or exceeding most of the goals reported to OneStar • Statistically significant differences (p=0.045) exist between organization types FINDING #3: Most Members Believe in Long-Term Impact of Their Service What Did You Leave Behind as a Result of Your Service Experience? (Percent of members surveyed choosing answer) No trace (box 1) A drop in the 2% bucket (box 2) Part of a real 15% solution (box 4) 30% The start of something important (box 3) 53% FINDING #4: Three Program Characteristics Are Tied to Member Assessment of Service Impact Thought that service to the community was "very helpful" (question 39)! Program Characteristics 1.00! • Service Clarity • Orientation Quality Probability! 0.80! 0.60! 0.40! Service Clarity! Orientation Quality! Inadequate! Adequate! Excellent! Inadequate! Adequate! Excellent! Inadequate! 0.00! Adequate! • Communication Quality Excellent! 0.20! Communication Quality! Financial Impact on Most Organizations… Financial Impact on Organizations by Utilizing AmeriCorps Members % of program managers surveyed Don't Know 26% Saves Money 57% Loses Money 2% Breaks Even 15% …and a Positive Economic Impact on Communities $16 $14 $12 $10 $8 $6 $4 $2 $0 $13.40 $6.82 Total net value per year: Over $10 million 6.58 Program manager assessment of value* Average hourly cost Net value * Program manager assessment of value has been regionally adjusted. Numbers in chart represent averages for 2009-10 and 2010-11 terms LESSONS LEARNED • Evaluating more than 1 AmeriCorps program at a time is like apples and oranges. • You may pay a lot of $$$ to learn things that seem somewhat predictable or ordinary. • Be ready to edit the final product. Heavily. • Not all evaluative data is created equal. (‘Perceived’ vs. ‘Actual’ Value) • Case Studies are difficult to do well. CURRENT PHILOSOPHY + APPROACH • Back to basics – You have to get ‘down in the weeds’ at some point – we have to specialize in each grantee’s program design to a basic extent, to empower them to make smart evaluation decisions and to ask the right questions. • Evaluation begins at accountability, but is really what should really drive decision-making – both for funders AND grantees. • Good leadership is what drives a culture of evaluative learning—make it part of everything we do from the top down. • ‘Practice what you preach’ to grantees + ‘There’s always room to grow’ (Annual grantee survey, training event surveys, member experience and inclusion survey) CURRENT TOOLS + RESOURCES GRANTEE RESOURCE LIBRARY: http://onestarfoundation.org/americorpstexas/grantees/grantee-resources/ 42 CURRENT TOOLS + RESOURCES GRANTEE RESOURCE LIBRARY: http://onestarfoundation.org/americorpstexas/grantees/grantee-resources/ 43 Contact Information • Emily Steinberg – Director, National Service Programs [email protected] 44 www.onestarfoundation.org 45 CNCS Office of Research & Evaluation: Recent Path, and Coming Steps CNCS Office of Research & Evaluation: Recent Path, and Coming Steps Stephen Plank, Ph.D. Director of Research & Evaluation [email protected] Overview • Who we are • What we do • What we strive to do in support of AmeriCorps State and National grantees • Recent accomplishments • Next priorities & activities About Our Team • 8 full-time members, plus talented interns, onsite consultant, & contracted partners • Strong working relationships with AmeriCorps State and National program leadership • Expertise in program evaluation, measurement, experimental & quasi-experimental design, community psychology, sociology, economics, public policy Our Mission • To build the evidence base for national service programs • To facilitate the use of evidence within CNCS & among its grantees and subgrantees – Our vision of what evaluation & evidence can be at their best… – Our view of what evaluation should not be about… What We Strive to do in Support AmeriCorps Grantees • Modeling & supporting best practices in evidence & evaluation • Participating in dialog & problem-solving • Participating in the grant-making & evaluationplanning process, as appropriate • Communicating findings & lessons learned to multiple audiences – Clarity & practicality as watchwords Our Portfolio • R&E about, and directed toward, the national service field • Grantee support – Technical assistance, “evidence exchange,” capacitybuilding tools • Helping CNCS, as an agency, use evidence & evaluation more effectively – GARP, organizational accountability, learning strategies Highlights from the Past Year - R&E with AmeriCorps State and National • Minnesota Reading Corps evaluations • School Turnaround AmeriCorps evaluation – Design Phase • AmeriCorps Exit Survey – revisions, focus on response rates and utilization • Grantee Evaluation and Training and Technical Assistance • Evaluation Bundling Project – Pilot Phase For the Coming Year and More - R&E with AmeriCorps State and National • School Turnaround AmeriCorps evaluation • justice AmeriCorps evaluation • R & E in support of other national service partnerships • Grantee evaluation training & technical assistance • Evaluation bundling project • Additional projects – building on past lessons learned Bethanne Barnes Program Examiner Office of Management and Budget Q&A
© Copyright 2024