6 Evaluating HRD3e Chapter09 Slides

download 6 Evaluating HRD3e Chapter09 Slides

of 55

Transcript of 6 Evaluating HRD3e Chapter09 Slides

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    1/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 1

    Evaluating HRD Programs

    Chapter 9

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    2/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 2

    Deliver theHRD

    Programme

    or intervention

    Assess Needs

    PrioritizeNeeds

    Define

    Objectives

    Develop

    Lesson Plans

    Develop/acquire materials

    select

    trainer/leader

    Select

    Methods & Techniques

    Schedule the

    programme/intervention

    Assessment Design Implementation Evaluation

    Select

    evaluation Criteria

    Determine

    evaluation design

    Conductevaluation

    of programme or

    intervention

    Interpret results

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    3/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 3

    Effectiveness The degree to which a training (or

    other HRD program) achieves its

    intended purpose. Measures are relative to some starting

    point.

    Measures how well the desired goal isachieved.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    4/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 4

    HRD EvaluationTextbook definition:

    The systematic collection ofdescriptive and judgmental informationnecessary to make effective trainingdecisions related to the selection,

    adoption, value, and modification ofvarious instructional activities.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    5/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 5

    In Other WordsAre we training:

    the right people

    the right stuff

    the right way

    with the right materials at the right time?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    6/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 6

    Evaluation Needs Descriptive and judgmental

    information needed.

    Objective and subjective data

    Information gathered according to a

    plan and in a desired format.

    Gathered to provide decision making

    information.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    7/553/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 7

    Purposes of Evaluation Determine whether the program is

    meeting the intended objectives.

    Identify strengths and weaknesses. Determine cost-benefit ratio.

    Identify who benefited most or least.

    Determine future participants. Provide information for improving HRD

    programs.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    8/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 8

    Purposes of Evaluation-2 Reinforce major points to be made.

    Gather marketing information.

    Determine if training program isappropriate.

    Establish management database.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    9/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 9

    Evaluation Bottom Line Is HRD a revenue contributor or a

    revenue user?

    Is HRD credible to line and upper-levelmanagers?

    Are benefits of HRD readily evident to

    all?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    10/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 10

    How Often are HRD

    Evaluations Conducted? Not often enough!!!

    Frequently, only end-of-courseparticipant reactions are collected.

    Transfer to the workplace is evaluatedless frequently.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    11/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 11

    Why HRD Evaluations are

    Rare Reluctance to having HRD programs

    evaluated.

    Evaluation needs expertise and resources. Factors other than HRD cause performance

    improvements, e.g.,

    Economy Equipment

    Policies, etc.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    12/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 12

    Need for HRD Evaluation Shows the value of HRD.

    Provides metrics for HRD efficiency.

    Demonstrates value-added approachfor HRD.

    Demonstrates accountability for HRDactivities.

    Everyone else has it why not HRD?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    13/553/15/2013

    HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 13

    Make or Buy EvaluationI bought it, therefore it is good.

    Since its good, I dont need to post-

    test. Who says its:

    Appropriate?

    Effective? Timely?

    Transferable to the workplace?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    14/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 14

    Evolution of Evaluation Efforts1. Anecdotal(short interesting story) approach: Talk to

    other users.

    2. Try before buy: Borrow and usesamples.

    3. Analyticalapproach: Match research

    data to training needs.4. Holisticapproach: Look at overall HRD

    process, as well as individual training.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    15/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 15

    Models and Frameworks of

    Evaluation Table 7-1 lists nine frameworks for

    evaluation.

    The most popular is that of D.Kirkpatrick:

    Reaction

    Learning Job Behavior

    Results

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    16/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 16

    Kirkpatricks Four Levels Reaction

    Focus on trainees reactions

    Learning Did they learn what they were supposed to?

    Job Behavior

    Was it used on job?

    Results

    Did it improve the organizations effectiveness?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    17/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 17

    Issues Concerning

    Kirkpatricks Framework Most organizations dont evaluate at

    all four levels.

    Focuses only on post-training.

    Doesnt treat inter-stageimprovements.

    WHAT ARE YOUR THOUGHTS?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    18/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 18

    Other Frameworks/Models 1 CIPP: Context, Input, Process, Product

    CIRO: Context, Input, Reaction, Outcome

    Brinkerhoff: Goal setting

    Program design

    Program implementation

    Immediate outcomes Usage outcomes

    Impacts and worth

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    19/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 19

    Other Frameworks/Models 2 Kraiger, Ford, & Salas:

    Cognitive outcomes

    Skill-based outcomes Affective outcomes

    Phillips: Reaction

    Learning Applied learning on the job

    Business results

    ROI

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    20/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 20

    A Suggested Framework 1 Reaction

    Did trainees like the training?

    Did the training seem useful?

    Learning

    How much did they learn?

    Behavior

    What behavior change occurred?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    21/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 21

    Suggested Framework 2 Results

    What were the tangible outcomes?

    What was the return on investment(ROI)?

    What was the contribution to the

    organization?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    22/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 22

    Data Collection for HRD

    EvaluationPossible methods:

    Interviews

    Questionnaires

    Direct observation

    Written tests Simulation/Performance tests

    Archival performance information

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    23/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 23

    InterviewsAdvantages:

    Flexible

    Opportunity forclarification

    Depth possible

    Personal contact

    Limitations:

    High reactive effects

    High cost Face-to-face threat

    potential

    Labor intensive Trained observers

    needed

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    24/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 24

    QuestionnairesAdvantages:

    Low cost to

    administer Honesty increased

    Anonymity possible

    Respondent sets thepace

    Variety of options

    Limitations:

    Possible inaccurate

    data Response conditions

    not controlled

    Respondents set

    varying paces Uncontrolled return

    rate

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    25/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 25

    Direct ObservationAdvantages:

    Non-threatening

    Excellent way tomeasure behaviorchange

    Limitations:

    Possibly disruptive

    Reactive effects arepossible

    May be unreliable

    Need trainedobservers

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    26/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 26

    Written TestsAdvantages:

    Low purchase cost

    Readily scored Quickly processed

    Easily administered

    Wide samplingpossible

    Limitations:

    May be threatening

    Possibly no relationto job performance

    Measures onlycognitive learning

    Relies on norms Concern for racial/

    ethnic bias

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    27/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 27

    Simulation/Performance TestsAdvantages:

    Reliable

    Objective Close relation to job

    performance

    Includes cognitive,psychomotor andaffective domains

    Limitations:

    Time consuming

    Simulations oftendifficult to create

    High costs to

    development anduse

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    28/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 28

    Archival Performance DataAdvantages:

    Reliable

    Objective Job-based

    Easy to review

    Minimal reactiveeffects

    Limitations:

    Criteria for keeping/

    discarding records Information system

    discrepancies

    Indirect

    Not always usable Records prepared

    for other purposes

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    29/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 29

    Choosing Data Collection Methods Reliability

    Consistency of results, and freedom from

    collection method bias and error. Validity

    Does the device measure what we want tomeasure?

    Practicality

    Does it make sense in terms of the resourcesused to get the data?

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    30/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 30

    Type of Data Used/Needed Individual performance

    System-wide performance

    Economic

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    31/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 31

    Individual Performance Data Individual knowledge

    Individual behaviors

    Examples: Test scores

    Performance quantity, quality, and

    timelinessAttendance records

    Attitudes

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    32/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 32

    System-Wide Performance Data Productivity

    Scrap/rework rates

    Customer satisfaction levels

    On-time performance levels

    Quality rates and improvement rates

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    33/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 33

    Economic Data Profits

    Product liability claims

    Avoidance of penalties Market share

    Competitive position

    Return on Investment (ROI) Financial utility calculations

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    34/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 34

    Use of Self-Report Data Most common method

    Pre-training and post-training data

    Problems: Mono-method bias

    Desire to be consistent between tests

    Socially desirable responses

    Response Shift Bias:

    Trainees adjust expectations to training

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    35/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 35

    Research DesignSpecifies in advance:

    the expected results of the study.

    the methods of data collection to be

    used.

    how the data will be analyzed.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    36/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 36

    Research Design Issues Pretest and Posttest

    Shows trainee what training has

    accomplished. Helps eliminate pretest knowledge bias.

    Control Group

    Compares performance of group withtraining against the performance of asimilar group without training.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    37/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 37

    Recommended Research

    Design Pretest and posttest with control group.

    Whenever possible:

    randomly assign individuals to the testgroup and the control group to minimizebias.

    Use time-series approach to datacollection to verify performanceimprovement is due to training.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    38/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 38

    Ethical Issues Concerning

    Evaluation Research Confidentiality

    Informed consent

    Withholding training from controlgroups

    Use of deception

    Pressure to produce positive results

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    39/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 39

    Assessing the Impact of HRD Money is the language of business.

    You MUST talk dollars, not HRD

    jargon. No one (except maybe you) cares

    about the effectiveness of training

    interventions as measured by andanalysis of formal pretest, posttestcontrol group data.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    40/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 40

    HRD Program Assessment HRD programs and training are

    investments.

    Line manager often see HR and HRD ascosts, i.e., revenue users, not revenueproducers.

    You must prove your worth to the

    organization Or youll have to find another

    organization.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    41/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 41

    Two Basic Methods for

    Assessing Financial Impact Evaluation of training costs

    Utility analysis

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    42/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 42

    Evaluation of Training Costs Cost-benefit analysis

    Compares cost of training to benefits

    gained such as attitudes, reduction inaccidents, reduction in employee sick-days, etc.

    Cost-effectiveness analysis Focuses on increases in quality, reduction

    in scrap/rework, productivity, etc.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    43/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 43

    Return on Investment Return on investment = Results/Costs

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    44/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 44

    Types of Training Costs Direct costs

    Indirect costs

    Development costs

    Overhead costs

    Compensation for participants

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    45/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 45

    Direct Costs Instructor

    Base pay

    Fringe benefits Travel and per diem

    Materials

    Classroom and audiovisual equipment Travel

    Food and refreshments

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    46/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 46

    Indirect Costs Training management

    Clerical/Administrative

    Postal/shipping, telephone,computers, etc.

    Pre- and post-learning materials

    Other overhead costs

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    47/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 47

    Development Costs Fee to purchase program

    Costs to tailor program to organization

    Instructor training costs

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    48/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 48

    Overhead Costs General organization support

    Top management participation

    Utilities, facilities

    General and administrative costs,such as HRM

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    49/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 49

    Compensation for Participants Participants salary and benefits for

    time away from job

    Travel, lodging and per-diem costs

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    50/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 50

    Measuring Benefits Change in quality per unit measured in

    dollars

    Reduction in scrap/rework measured indollar cost of labor and materials

    Reduction in preventable accidentsmeasured in dollars

    ROI = Benefits/Training costs

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    51/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 51

    Utility Analysis Uses a statistical approach to support

    claims of training effectiveness: N = Number of trainees

    T = Length of time benefits are expected to last

    dt = True performance difference resulting fromtraining

    SDy = Dollar value of untrained job performance (instandard deviation units)

    C = Cost of training

    U = (N)(T)(dt)(Sdy) C

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    52/55

    3/15/2013 HRD3e Contributed by WellsDoty, Ed.D. Clemson Univ. 52

    Critical Information for Utility

    Analysis dt = difference in units between

    trained/untrained, divided by

    standard deviation in units producedby trained.

    SDy = Standard deviation in dollars,

    or overall productivity oforganization.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    53/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 53

    Ways to Improve HRD

    Assessment Walk the walk, talk the talk: MONEY.

    Involve HRD in strategic planning.

    Involve management in HRD planning andestimation efforts.

    Gain mutual ownership

    Use credible and conservative estimates. Share credit for successes and blame for

    failures.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    54/55

    3/15/2013HRD3e Contributed by Wells

    Doty, Ed.D. Clemson Univ. 54

    HRD Evaluation StepsAnalyze needs.

    Determine explicit evaluation strategy.

    Insist on specific and measurabletraining objectives.

    Obtain participant reactions.

    Develop criterion measures/instrumentsto measure results.

    Plan and execute evaluation strategy.

  • 7/29/2019 6 Evaluating HRD3e Chapter09 Slides

    55/55

    HRD3 C t ib t d b W ll

    Summary Training results must be measured

    against costs.

    Training must contribute to thebottom line.

    HRD must justify itself repeatedly as

    a revenue enhancer, not a revenuewaster.