How to Develop a Project Evaluation Plan Pat Gonzalez Office of Special Education Programs...

26
How to Develop a Project Evaluation Plan Pat Gonzalez Office of Special Education Programs [email protected] 202-245-7355

Transcript of How to Develop a Project Evaluation Plan Pat Gonzalez Office of Special Education Programs...

How to Develop a Project Evaluation

Plan

Pat GonzalezOffice of Special Education Programs

[email protected]

202-245-7355

Initial Steps in Developing the Outcome Evaluation Plan

Identify the program or project’s mission and/or goals and objectives

Identify all relevant and important outcomes that should be evaluated

Select outcome indicators Identify data sources and data collection

procedures

Step 1: Identify mission and/or goals and objectives

Clarify the expectations and priorities of key stakeholders/collaborators

Get a reasonable level of agreement on goals, strategies or activities and outcomes

Develop Goals: Broad statements generally describing desired outcomes

Develop Objectives: Measurable statements about outcomes (target performance) expected to be accomplished in a given time frame

More on Objectives…

Objectives require detail and must include a target group (who), what is to be done (activities), a time frame (when), and a target performance (how much).

80% (how much) of the 300 participating teachers (who) will indicate that the Transition Toolkit is useful, relevant and of high quality (what) on the second year follow-up survey (when).

Step 2: Identify relevant and important project outcomes

Short-term Outcomes typically involve learning: awareness, knowledge, attitudes, skills

Intermediate Outcomes typically involve action: behavior, practice, policies

Long-term Outcomes typically involve conditions: social, economic, civic, environmental

REMEMBER!

Focus on Short and Intermediate-term outcomes that can be

completed within the grant period

Outcomes are not Outputs!

Outputs are the direct products of program activities, usually measured by “volume”, such as the number of classes taught or number of participants served

Sources of information on Program Outcomes

Legislation and regulations Purpose statements contained in the RFP Strategic plans, SPPs or APRs State data systems Program descriptions or annual reports Discussions with stakeholders or collaborators Complaint information Performance measures from government agencies

or other programs

Step 3: Select outcome indicators

Each identified outcome needs to be translated into one or more outcome indicators that state specifically what is to be measured (e.g., the # of teachers passing a test)

Checklist for Outcome Indicators

Does each indicator measure some important aspect of the outcome?

Does each indicator start with a numerical designation such as : incidence, percentage, rate, or proportion of?

Does your list of indicators cover all your outcomes?

Checklist for Outcome Indicators-continued

Does your list indicators cover “quality” characteristics, such as timeliness of services?

Is the wording of your indicator sufficiently specific? Avoid terms like “appropriate”.

What is the feasibility and cost of collecting the indicator? Note that sometimes costly indicators are the most important and should be retained.

Step 4: Identify data sources and data collection procedures

Determine if a research design can be used to evaluate effectiveness. There are several quasi-experimental designs that can be readily applied to program evaluation.

Identify data sources, such as: extant agency/program records, performance assessments, surveys, observer ratings, and interview data.

Considerations in Determining Data Collection Procedures

When will data be collected? Consider your design… When entering the program When completing the program Fixed interval after entering Fixed interval after completing Combination of above

Considerations in Determining Data Collection Procedures-continued

Who is considered a participant?

Include all participants or a sample?

Who will collect the data?

How will the data be analyzed?

The Evaluation PlanImplementation Questions

Using the information concerning, goals, objectives, strategies/activities and outcomes, develop evaluation questions on implementation:

Were the activities completed as intended, on time and did they result in the planned outputs? These questions provide a feedback loop for the purposes of on-going project monitoring.

The Evaluation PlanOutcome Questions

Using the information concerning, goals, objectives, strategies/activities and outcomes, develop evaluation questions on impact/effectiveness:

How well did the activities address the objectives as measured by the indicators? What changed for the target group either over time or in comparison to another group?

The Evaluation PlanMethods

For each evaluation question:

Is a research design feasible/which ones? What are the data sources? What methods will be used to collect the data? How might the data be analyzed and

reported?

The Evaluation PlanTimelines

For each evaluation question:

When will data be collected?

When will data be reported or used?

The Evaluation PlanPersonnel Responsible

Who is responsible for data collection, analysis and reporting at each point in the timeline?

SPDG Evaluation Criteria

The extent to which the methods of evaluation are thorough, feasible, and appropriate to the goals, objectives and outcomes of the proposed project.

The extent to which the methods of evaluation provide for examining the effectiveness of project implementation strategies.

SPDG Evaluation Criteria - continued

The extent to which the methods of the evaluation include the use of the objective performance measures that are clearly related to intended outcomes of the project and will produce quantitative and qualitative data to the extent possible.

SPDG Evaluation Criteria - continued

The extent to which the methods of evaluation will provide performance feedback and permit assessment of progress toward achieving intended outcomes

Performance Measures

The percent of personnel receiving professional development through the SPDG Program based on scientific- or evidence-based instructional practices.

The percent of SPDG projects that have implemented personnel development/training activities that are aligned with improvement strategies identified in their State Performance Plan.

Performance MeasuresContinued

The percent of professional development/training activities provided through the SPDG Program based on scientific- or evidence-based instructional/behavioral practices.

The percent of professional development/training activities based on scientific- or evidence-based instructional/behavioral practices, provided through the SPDG Program, that are sustained through ongoing and comprehensive practices (e.g., mentoring, coaching, structured guidance, modeling, continuous inquiry, etc.).

Performance MeasuresContinued

In States with SPDG projects that have special education teacher retention as a goal, the Statewide percent of highly qualified special education teachers in State-identified professional disciplines (e.g., teachers of children with emotional disturbance, deafness, etc.), consistent with sections 602(a)(10) and 612(a)(14) of IDEA, who remain teaching after the first three years of employment.

Performance MeasuresContinued

The percent of SPDG projects that successfully replicate the use of scientific or evidence-based instructional/behavioral practices in schools.

The percent of SPDG projects whose cost per personnel receiving professional development on scientific or evidence-based practices is within a specified range.