Capturing Elusive Level 3 Data: The Secrets Of Survey Design - PowerPoint PPT Presentation

1 / 38
About This Presentation
Title:

Capturing Elusive Level 3 Data: The Secrets Of Survey Design

Description:

Title: Capturing Elusive Level 3 Data: The Secrects of Survey Design Subject: Measurement, Evaluation, ROI Author: Ken Phillips Description [Session SU116] In this ... – PowerPoint PPT presentation

Number of Views:56
Avg rating:3.0/5.0
Slides: 39
Provided by: KenPhi
Category:

less

Transcript and Presenter's Notes

Title: Capturing Elusive Level 3 Data: The Secrets Of Survey Design


1
  • Capturing Elusive Level 3 Data The Secrets Of
    Survey Design
  • Session SU116

Presented by Ken Phillips Phillips
Associates May 19, 2013
2
AGENDA
  • Review Kirkpatrick/Phillips 5 Level Evaluation
    Model
  • Examine Level 3 evaluation facts
  • Analyze survey creation errors in a sample Level
    3 evaluation
  • Apply 12 tips for creating valid, scientifically
    sound Level 3 evaluations

3
KIRKPATRICK/PHILLIPS EVALUATION MODEL
Levels of Evaluation Measurement Focus Time Frame
Level 1 Reaction Participant favorable reaction to a learning program Conclusion of learning program
Level 2 Learning Degree to which participants acquired new knowledge, skills or attitudes Conclusion of learning program or within 6 to 8 weeks after
Level 3 Behavior Degree to which participants applied back-on-the-job what was learned 2 to 12 months
Level 4 Results Degree to which targeted business outcomes were achieved 9 to 18 months
Level 5 ROI Degree to which monetary program benefits exceed program costs 9 to 18 months
4
LEVEL 3 EVALUATION FACTS
55 of organizations evaluate at least some
learning programs at Level 3
Organizations that use Level 3s on average
evaluate 25 of all programs
75 of organizations view data collected as
having high or very high value
ASTD Research Study, The Value of Evaluation
Making Training Evaluations More Effective, 2009
5
DATA COLLECTION METHODS
Donald James Kirkpatrick, Evaluating Training
Programs The Four Levels, 2006.
6
LEVEL 3 EVALUATION FACTS
  • Most common methods for evaluating behavior
  • Participant surveys (31)
  • Action planning (27)
  • Performance records monitoring (24)
  • On job observation (24)

ASTD Research Study 2009
7
Possible survey respondents
  • Learners
  • Peers/colleagues
  • Direct reports
  • Managers
  • External customers

8
HOW TO DECIDE
Who has first-hand knowledge of learners
behavior?
How disruptive/costly is data collection method?
How credible do results need to be?
What are business executive/stakeholder
expectations?
9
REQUIRED DRIVERS
  • Planning to measure behavior? Include required
    drivers
  • Reinforcements follow-up modules, job aids
  • Encouragement coaching, mentoring
  • Rewards recognition,
  • Monitoring surveys, focus groups,
  • On-the-job behavior change is both our
    managements responsibility

Kirkpatrick Four Levels Evaluation Certification
Program participant workbook
10
SAMPLELEVEL 3PARTICIPANT SURVEY
11
WHATS WRONG WITH THESE?
8. Before providing employees with feedback about
their job performance, my manager considers
whether or not he or she is knowledgeable about
their job. 25. When giving feedback to an
employee my manager considers whether it should
be done privately or in the presence of others.
12
TIP 1 (CONTENT)
Focus on observable behavior not thoughts or
motives
Palmer Morrel-Samuels, Getting the Truth into
Workplace Surveys, Harvard Business Review, 2002.
13
WHATS WRONG WITH THESE?
  • 14. My manager gives his or her employees
    feedback just as soon as possible after an event
    has happened and avoids getting emotional or
    evaluative.
  • 18. My manager provides employees with regular
    ongoing feedback about their job performance and
    speaks in a normal conversational tone or manner
    when delivering the feedback.

14
TIP 2 (CONTENT)
Limit each item to a single description of
behavior
Palmer Morrel-Samuels, 2002
15
WHATS WRONG WITH THESE?
  • 2. My manager doesnt get to know his or her
    employees as individuals before providing them
    with feedback about their job performance.
  • 7. When giving employees feedback about their
    job performance, my manager doesnt distinguish
    between patterns of behavior and random one-time
    events.

16
TIP 3 (CONTENT)
Word about 1/3 of the survey items so that the
desired answer is negative
Palmer Morrel-Samuels, 2002.
17
WHATS WRONG WITH THESE?
  • Building Trust
  • Credibility
  • Feedback Sign
  • Feedback Timing
  • Feedback Frequency
  • Message Characteristics

18
TIP 4 (FORMAT)
Keep sections of the survey unlabeled
Palmer Morrel-Samuels, 2002.
19
TIP 5 (FORMAT)
Design sections to contain a similar number of
items and questions to contain a similar number
of words
Palmer Morrel-Samuels, 2002.
20
TIP 6 (FORMAT)
Place questions regarding respondent demographics
(e.g. name, title, department, etc.) at end of
survey, make completion optional and keep
questions to a minimum
Palmer Morrel-Samuels, 2002.
21
TIP 7 (MEASUREMENT)
Collect data from multiple observers or a single
observer multiple times
22
WHATS WRONG WITH THIS?
Strongly Agree Agree Disagree Strongly Disagree N/A
4 3 2 1
23
TIP 8 (MEASUREMENT)
Create a response scale with numbers at regularly
spaced intervals and words only at each end
Palmer Morrel-Samuels, 2002.
24
EXAMPLES
This
Not at all True Completely True
1 2 3 4 5 6 7
Not This
Not at all True RarelyTrue Occasionally True SomewhatTrue MostlyTrue FrequentlyTrue Completely True
1 2 3 4 5 6 7


Never This
Not at all True Rarely True Occasionally True Somewhat True Mostly True Frequently True Completely True

25
TIP 9 (MEASUREMENT)
Use only one response scale with an odd number of
points (7, 9 11 point scales are best)
Palmer Morrel-Samuels, 2002.
26
ODD vs EVEN SCALE
27
TIP 10 (MEASUREMENT)
Use a response scale that measures frequency not
agreement or effectiveness
Palmer Morrel-Samuels, 2002.
28
EXAMPLES
This
Never Always
1 2 3 4 5 6 7
Or this
Not at all True Completely True
1 2 3 4 5 6 7
29
TIP 11 (MEASUREMENT)
Place small numbers at left or low end of scale
and large numbers at right or high end of scale
30
EXAMPLES
This
Not at all True Completely True
1 2 3 4 5 6 7
Not this
Completely True Not at all True
7 6 5 4 3 2 1
31
TIP 12 (MEASUREMENT)
Include a Not Applicable or Did Not Observe
response choice and make it different
Palmer Morrel-Samuels, 2002.
32
EXAMPLE
Not at all True Completely True NA
1 2 3 4 5 6 7


33
SUMMARY (CONTENT)
Focus on observable behavior
Limit ideas to a single description of behavior
Word 1/3 of items as reverse score
34
SUMMARY (FORMAT)
Keep survey sections unlabeled
Design sections to contain similar number of
items questions similar number of words
Place questions regarding respondent demographics
at end of survey, make completion optional and
keep questions to a minimum
35
SUMMARY (measurement)
Collect data from multiple observers or multiple
times
Create a response scale that
  • Has words only at each end
  • Has an odd number of points
  • Measures frequency
  • Has small numbers at left and large numbers at
    right
  • Includes a Not Applicable that is different

36
Free Articles
  • Phillips, Ken, Eight Tips on Developing Valid
    Level 1 Evaluation Forms, Training Today, Fall
    2007, pps. 8 14.
  • Phillips, Ken, Developing Valid Level 2
    Evaluations, Training Today, Fall 2009.
  • Phillips, Ken, Capturing Elusive Level 3 Data
    The Secrets of Participant Survey Design,
    Unpublished article, 2013.
  • Phillips, Ken, Do Level 1 Evaluations Have a
    Role in Organizational Learning Strategy?,
    Unpublished article, 2013.

37
YOUR FEEDBACK COUNTS!
  • Your Feedback Counts!
  • Evaluation forms for this session are available
    NOW via the mobile app and at the following link
    www.astdconference.org

38
  • Ken Phillips
  • Phillips Associates
  • 34137 N. Wooded Glen Drive
  • Grayslake, Illinois 60030
  • (847) 231-6068
  • www.phillipsassociates.com
  • ken_at_phillipsassociates.com
Write a Comment
User Comments (0)
About PowerShow.com