Title: A Conversation about Program Evaluation: Why, How and When
1A Conversation about Program EvaluationWhy, How
and When?
- Uzo Anucha
- Applied Social Welfare Research and Evaluation
Group - School of Social Work
- York University
2What is Program Evaluation?
- Program evaluation means taking a systematic
approach to asking and answering questions about
a program. - Program evaluation is not an assessment of
individual staff performance. The purpose is to
gain an overall understanding of the functioning
of a program. - Program evaluation is not an audit evaluation
does not focus on compliance with laws and
regulations. - Program evaluation is not research. It is a
pragmatic way to learn about a program.
3What is Program Evaluation?
- Program evaluation is not one method. It can
involve a range of techniques for gathering
information to answer questions about a program. - Most programs already collect a lot of
information that can be used for evaluation. Data
collection for program evaluation can be
incorporated in the ongoing record keeping of the
program.
4Definition of Program Evaluation
- Program evaluation is a collection of methods,
skills and sensitivities necessary to determine
whether a human service is needed and likely to
be used, whether it is sufficiently intensive to
meet the unmet needs identified, whether the
service is offered as planned, and whether the
human service actually does help people in need
at reasonable cost without undesirable side
effects (Posavac Carey, 2003. p.2)
5Why Evaluate?
- Verify that resources are devoted to meeting
unmet needs - Verify that planned programs do provide services
- Examine the results
- Determine which services produce the best results
- Select the programs that offer the most needed
types of services
6Why Evaluate?
- Provide information needed to maintain and
improve quality - Watch for unplanned side effects
- Create program documentation
- Help to better allocate program resources
- Assist staff in program development and
improvement
7Evaluation can.
- Increase our knowledge base
- Guide decision making
- Policymakers
- Administrators
- Practitioners
- Funders
- General public
- Clients
- Demonstrate accountability
- Assure that client objectives are being achieved
8Who is an evaluation for?
9Who is an evaluation for?
- What do they want to know?
- What do we want to tell them about the program?
- How can they contribute to the evaluation?
- Program participants?
- Family members and caregivers?
- Program staff?
- Volunteers?
- Partner agencies and professionals?
- Referral sources?
- Funders?
- Others?
10Being Smart About Evaluation
- Every evaluation happens in a political context.
Find out what it is. - Clarify your role in the evaluation. Let people
know what you can and can not do. - Be fair and impartial.
- Consider what is a reasonable and feasible
evaluation for the particular program.
11Challenging Attitudes toward Program Evaluation.
- Expectations of slam-bang effects
- Assessing program quality is unprofessional
- Evaluation might inhibit innovation
- Program will be terminated
- Information will be misused
- Qualitative understanding might be lost
- Evaluation drains resources
- Loss of program control
- Evaluation has little impact
12Types of evaluations
- Needs assessment
- Evaluability assessment
- Process evaluation
- Outcome evaluation
- Efficiency evaluation (cost evaluation)
13Needs Assessment/Community Capacity Mapping
- Prerequisite to program planning and development
- What is the community profile?
- What are the particular unmet needs of a target
population? - What forms of service are likely to be attractive
to the population? - Are existing services known or acceptable to
potential clients? - What barriers prevent clients from assessing
existing services?
14Evaluability Assessment
- Prerequisite to formal evaluation
- Are program goals articulated and measurable?
- Is the program model definable (flow diagram)?
- Are the goals and activities logically linked?
- Is there sufficient rigour and resources to merit
evaluation?
15Process Evaluation
- Verify program implementation
- Is the program attracting a sufficient number of
clients? - Are clients representative of the target
population? - How much does the staff actually contact the
client? - Does the workload of staff match that planned?
- Are there differences in effort among staff?
16Outcome Evaluation
- Describe program effects
- Is the desired outcome observed?
- Are program participants better off than
non-participants? - Is there evidence that the program caused the
observed changes? - Is there support for the theoretical foundations
underpinning the program? - Is there evidence that the program could be
implemented successfully elsewhere?
17Evaluation of Efficiency
- Effectiveness relative to cost
- Are funds spent for intended purposes?
- Are program outcomes achieved at a reasonable
cost? - Can dollar values be assigned to the outcomes?
- Is the outcome achieved greater than other
programs of similar costs?
18Process Evaluation.
19Process Evaluation
- Sometimes referred to as formative evaluation
- Looks at the approach to client service
delivery...day to day operations - Two major elements
- 1) how a programs services are delivered to
clients (what workers do including frequency and
intensity client characteristics satisfaction - 2) administrative mechanisms to support these
services (qualifications structures hours
support services supervision training)
20Process Evaluation
- Can occur concurrently with outcome evaluation
- Need to establish common program language
- Purpose of process evaluation improve service
generate knowledge estimate cost efficiency - May be essential component of organizational
accreditation
21Steps in Process Evaluation
- Deciding what questions to ask
- background client profile staff profile
nature, amount and duration of service provided
nature of interventions admin. supports
satisfaction of key stakeholders efficiency? - Developing data collection instruments
- ease of use consistency with program operation
and objectives user input)
22Steps in Process Evaluation
- Developing a data collection monitoring system
(unit of analysis number of units to include
e.g. sampling when and how to collect data - Scoring and analyzing data (categorize by client
groups, interventions, program display
graphically). - Developing a feedback system (clients, workers,
supervisors, administrators). - Disseminating and communicating results
23Sources of Process Evaluation Data
- Funder/agency/program documents (eg. Model
rationalefunding agreement) - Key informant interviews with service delivery
and admin. personnel or key collateral agencies
(e.g. Referral source) - Service utilization statistics
- Management Information Systems (M.I.S.)
- Surveys/interviews with consumers (e.g. client
satisfaction
24Outcome Evaluation.
25Outcome Evaluation
- Outcomes are benefits or changes for individuals
or populations during or after participating in
program activities. Outcomes may relate to
behavior, skills, knowledge, attitudes, values,
condition, or other attributes. - They are what participants know, think, or can
do or how they behave or what their condition
is, that is different following the program. - Outcome evaluation helps us to demonstrate the
nature of change that took place
26Outcome Evaluation
- Outcome evaluation tests hypotheses about how we
believe that clients will change after a period
of time in our program. - Evaluation findings are specific to a specific
group of clients experiencing the specific
condition of one specific program over a specific
time frame at a specific time.
27For example
- A program to counsel families on financial
management, outputs--what the service
produces--include the number of financial
planning sessions and the number of families
seen. The desired outcomes--the changes sought in
participants' behavior or status--can include
their developing and living within a budget,
making monthly additions to a savings account,
and having increased financial stability.
28Uses of Outcome Evaluation
- Improving program services to clients
- Generating knowledge for the profession
- Estimating costs
- Demonstrate nature of change...evaluation of
program objectives e.g. what we expect clients to
achieve - Guide major program decisions and program
activities
29Program-Level Evaluations
- Program level evaluations vary on a continuum and
are fundamentally made up of three levels - Exploratory
- Descriptive
- Explanatory
30The Continuum
- Qualitative ------------Quantitative
- Exploratory----Descriptive----Explanatory
31Exploratory Outcome Evaluation Designs
- Questions here include
- Did the participants meet a criterion (e.g.
Treated vs. Untreated)? - Did the participants improve (e.g. appropriate
direction)? - Did the participants improve enough (e.g.
statistical vs. meaningful difference)? - Is there a relation between change and service
intensity and participant characteristics?
32Exploratory Designs
- One group post test only
- Multi-group post test only
- Longitudinal case study
- Longitudinal survey
33Strengths of Exploratory Designs
- Less intrusive and inexpensive
- Assess the usefulness and feasibility of further
evaluations - Can correlate improvement with other variables.
34Descriptive Designs
- To show that something causes something else, it
is necessary to demonstrate - That the cause precedes the supposed effects in
time e.g. that an intervention precedes the
change - That the cause covaries with the effect the
change covaries with the intervention the more
the intervention, the more the change. - That no viable explanation of the effect can be
found except for the assumed cause e.g. there can
be no other explanation for the change except the
intervention. - Both 1 and 2 can be achieved with exploratory
designsbut not 3.
35Descriptive Designs
- Randomized one-group posttest only
- Randomized cross-sectional and longitudinal
survey - One-group pretest-posttest
- Comparison group posttest only
- Comparison group pretest-posttest
- Interrupted time series
36Explanatory Designs
- Defining characteristic is observation of people
randomly assigned to either a program or control
condition. - Considered much better at addressing threats to
internal validity - Program group vs. Control group if groups are
formed randomly there is no reason to believe
they differ in rate of maturation no self
selection into groups groups did not begin at
different levels
37Explanatory Designs
- Classical experimental
- Solomon four group
- Randomized posttest only control group
38Explanatory Designs
- Strengths/Limitations
- counter threats to internal validity
- allow interpretations of causation
- expensive and difficult to implement
- frequently resistance from practitioners who
already know what is best - Suggested Times to Use
- when new program is introduced
- when stakes are high
- when there is controversy over efficacy
- when policy change is desired
- when program demand is high
39Internal Validity (causality)
- Mortality
- Reactive effects
- Interaction effects
- Relations between experimental and control groups
(e.g. rivalry)
- History
- Maturation
- Testing
- Instrumentation error
- Statistical regression
- Differential selection
40External Validity (generalizability)
- Pretest-treatment interaction
- Selection-treatment interaction
- specificity of variables and settings
- reactive effects
- multiple treatment interference
- researcher bias
41Steps in Outcome Evaluation
- Step 1 Operationalizing program objectives
- Step 2 Selecting the measurements and stating
the outcomes - psychometrics, administration, measurement burden
- Step 3 Designing a monitoring system
- sampling ( clients to include - min. 30 per
group of interest sampling strategies missing
data)
42Steps in Outcome Evaluation
- Step 3 (contd) Designing a monitoring system
deciding when how data will be collected will
depend on question we are trying to answer - What degree is program achieving its objectives
e.g. how much change? - Differences between program participants and
non-participants - Question of causality
- Longevity of client changes
- Deciding how data will be collected
- By telephone, mail or in person.
43Steps in Outcome Evaluation
- Step 4 Analysing and displaying data
- Can also present outcome data according to
subgroups by using demographics - most useful when when data can be aggregated and
summarized to provide an overview on client
outcomes. - Step 5 Developing a feedback system
- Step 6 Disseminating results
44Ready, Set, Go?
- Some things to consider..
45Important to consider
- Internal or external evaluators?
- Scope of evaluation?
- Boundary
- Size
- Duration
- Complexity
- Clarity and time span of program objectives
- Innovativeness
46Sources of Data for Evaluation
- Intended beneficiaries of the program
- Program participants
- Artifacts
- Community indexes
- Providers of service
- Program staff
- Program records
- Observers
- Expert observers
- Trained observers
- Significant others
- Evaluation staff
47Good Assessment Procedures
- Multiple Sources triangulation and corroborating
evidence - Multiple Variables focus on single variable can
corrupt evaluation different variables affected
by different sources of error - Non-Reactive Measures measures which do not
themselves effect respondents - Important Variables politically, conceptually
and methodologically important variables
48Good Assessment Procedures
- Valid Measures instrument measures what it is
supposed to measure (face, criterion, construct)
more focused on objective behaviour...more likely
to be valid - Reliable Measures consistent measure of
construct (stable e.g test-retest recognizable
e.g. inter-rater homogeneity e.g. split-half) - Sensitivity to Change able to detect small
changes pre-test scores should be scrutinized
49Good Assessment Procedures
- Cost-effectiveness length and ease and cost of
production and distribution - Grounded in existing research and Experiential
relevance use within literature published
psychometric data and population norms
pre-tested with relevant population
50Ideal Program Evaluation Characteristics
- Counter threats to internal/external validity
- established time ordering (intervention precedes
effect) - intervention is manipulated (admin. to at least
one group) - relations between intervention and program must
be established - design must control for rival hypotheses
- one control group must be used
- random assignment
51Planning an Evaluation
- Identify the program and stakeholders
- Identify information needs of evaluation
- Examine the literature
- Determine the methodology (sample, design, data
collection, procedures, analysis, time lines,
budget) - Prepare a written proposal
52Preparing for an Evaluation
- Obtain a complete program description
(newsletters, annual reports, etc). - Identify meet with stakeholders program
director, staff, funders/program sponsors and
clients/program participants. - Identify the information needs of evaluation
- Who wants an evaluation?
- What should evaluation focus on?
- Why is an evaluation needed?
- When is an evaluation wanted?
- What resources are needed?
- What is the evaluability of the program?
53Things to Consider..
- Planning an evaluation follows similar steps to
the conduct of more basic research with some
additional considerations - More effort needs to be expended in engaging and
negotiating with stakeholder groups (e.g.
interviews, or research study steering committee) - There needs to be a keener awareness of the
social/political context of the evaluation (e.g.
differing and competing interests)
54Things to Consider..
- Greater effort needs to be expended in becoming
familiar with and articulating program evaluation
criteria including goals, objectives and
implementation of model and theory - The choice of measurement needs to be grounded in
a detailed understanding of program goals,
objectives and service delivery in addition to
the qualities of ideal assessment measures (e.g.
reliability and validity) - All stages of planning must take into
consideration practical limitations (eg. time,
budget, resources)
55Program Evaluation Exercise..
- Consider a social service setting with which you
are familiar and illustrate how program
evaluation activities could be applied to it. - What needs to be done if you were to contemplate
evaluating this program? - Who are the different stakeholders who should be
involved in an evaluation and how? - What are some evaluation questions that could be
asked and what methods could one use? - What are some criteria of program success that
can be easily measured but miss the central point
of the program (measurable but irrelevant)? - What are some measurable and relevant criteria?
- What else needs to be considered?