42
Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into Evaluation Science April 2005

Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Evaluation DialogueBetween OMB Staff and Federal Evaluation Leaders

Digging a Bit Deeper intoEvaluation Science

April 2005

Page 2: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Evaluation Dialogue

Session II

Examples of Evaluation Approaches and Discussion

Page 3: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Example 1: Evaluation of Impact of Agency Actions on Baby Walker Injuries

Evaluation Question: Has the safety standard for baby walkers, designed to prevent falls down stairs, been effective in reducing injuries?

Page 4: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Logic Model for Reducing Baby-Walker Related Injuries

How?Why?

ResourcesFTEs$

Inputs

Activities• Collect baby-walker related cases at NEISS hospitals• Complete in-depth investigations• Analyze data • Conduct technical reviews• Conduct labora-tory tests•Work with industry

Outputs Intermediate Outcomes

•Safety StandardsMake voluntary standard recommenda-tions or develop a mandatory standard for baby walkers•Compliance –obtain recalls of defective baby walkers• Consumer Information –publicize dangers of baby walkers

Mission-related outcome: Reduce baby-walker related injuries

Final Outcomes (Impact)

External Factors – Alternative explanations

Page 5: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Evaluation Design:Interrupted Time Series

• Voluntary standard applied to all baby walkers sold or imported to U.S.

• Few alternative explanations for effects.

• Random assignment to groups hasethical implications.

Page 6: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Baby Walker-Related Injury Rate: 1981 to 2001

012345678

1980 1983 1986 1989 1992 1995 1998 2001

Inju

ry R

ate

Per

100

0 L

ive

Bir

ths

Page 7: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Possible alternative explanations for impact of safety standard for baby walkers

• A secular safety trend?

• Change in sales of baby walkers?

• The probability sample for estimating injuries changed?

• The population changed?

Page 8: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Example 2: Evaluation of Montgomery GI Bill Program (at VA)

Evaluation Question: To what extent has the program met• its statutory intent, • the educational needs of

beneficiaries, and • the expectations of stakeholders?

Page 9: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Logic Model for Assessing Educational Needs Attainment

How?Why?

Resources$ and FTEsData•Requirement projections of labor market•User profiles•Educational status leaving military•Future military needs•YATS data

Inputs

•Trainees

•Program Drop-out

•Certificate/License

Outputs Intermediate Outcomes

•Degree/Certificate or Professional License

•Income

•Education and Career Goals

•Full Adjustment to Civilian Life

•Youth Attracted to Military Service

Final Outcomes (Impact)

External Factors – Alternative explanations

Page 10: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Evaluation Design:Quasi-Experimental

• Stratified survey (mock longitudinal)– 10 cohorts by year entered military– user, non-user, non participants

• Comparisons among cohort groups• Compared to general population

educational statistics

Page 11: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Ability to find Work: Users and Non-Users

17%14%

21% 20%

0%

5%

10%

15%

20%

25%

Perc

ent

Problems finding a job Unemployed and looking for work for >=1 month (past yr)

Page 12: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Possible alternative explanations for impact of GI Bill program success

• Age at completing military service

• Impact of family responsibilities

• Program’s adaptation to new educational delivery models

Page 13: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Example 3: Evaluation of the impact of CR Packaging Requirements on the

Child Death Rate

Evaluation Question: Have the mandatory standards requiring child resistant (CR) packaging for oral prescription drugs been effective in reducing the death rate from unintentional poisonings to children under 5 years old?

Page 14: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Logic Model for Reducing the Death Rate from Unintentional Poisonings to Children Under 5 Years Old

How?Why?

ResourcesFTEs$Partners

Inputs

Activities• Monitor ingestion incident data•Identify products with chemicals known to be potent-ially hazardous to children.• conduct human performance testing for compliance with CR requirements.

Outputs Intermediate Outcomes

•Safety StandardsDevelop mandatory standards for CR packaging for oral prescription drugs•Compliance –obtain recalls for products that violate CR packaging regulations• Consumer Information –promote CR packaging with partners

Mission-related outcome: Reduce the death rate for children under 5 associated with oral prescription drugs.

Final Outcomes (Impact)

External Factors – Alternative explanations

Page 15: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Evaluation Design: Interrupted Time Series

• Mandatory standard applied to all oralprescription drugs sold or imported toU.S.

• Few alternative explanations foreffects

• Random assignment to groups hasethical implications

Page 16: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Estimated Child Death Rates Associated with Accidental Medicine-Related Poisonings

0

1

2

3

4

65 70 75 80 85 90Year

Dea

ths

per M

illio

n C

hild

ren

Und

er A

ge F

ive

Predicted

Predicted, in Absenceof CR Packaging Reqs.

Predicted with CR Packaging

Page 17: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Possible alternative explanations for impact of CR Packaging Requirements on

the Child Death Rate

• A secular safety trend?

– Health care

– Poison Control Centers

– Parental awareness

• Change in sales of prescription drugs?

Page 18: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

OSHA Example

Evaluation of alternative strategies to meet strategic goal to reduce rate of workplace injuries and illnesses

Page 19: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Budget Activities Major Outputs Performance Goals

Safety & Health Standards

Federal Enforcement

State Programs

Technical Support

Compliance Assistance

Safety & Health Statistics

Inspections & Discrimination Investigations

Construction Standards

Construction Guidance

Economic Analysis & Rulemaking

Technical Support Services, including Construction Engineering Services

Construction Services

Education Centers Programs

Reduce the rate of workplace& work-related

fatalities, injuries & illnesses

Page 20: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Budget Activities Major Outputs Performance Goals

Safety & Health Standards

Federal Enforcement

State Programs

Technical Support

Compliance Assistance

Safety & Health Statistics

Inspections & Discrimination Investigations

Construction Standards

Construction Guidance

Economic Analysis & Rulemaking

Technical Support Services, including Construction Engineering Services

Construction Services

Education Centers Programs

Reduce the rate of workplace& work-related

fatalities, injuries & illnesses

Evaluation Question:

Which type of post-inspection “settlement agreement” is most cost effective for improving workplace safety & health?

Page 21: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Study DesignPOPULATION In-scope establishments that were inspected in the study time period with citations

TREATMENT GROUPS

Samples of in-scope establishments with:–Informal Settlement Agreements–Formal Settlement Agreements–Corporate-Wide Settlement Agreements–Contested Citations Settled Through Litigation

CONTROL GROUPS

• Each treatment groups serves as a control for other treatment groups.• In-scope establishments with citations, but no settlement agreements or litigated case decisions by an administrative law judge or the Review Commission

OUTCOMES changes in injury/illness rates types of hazards abatedTypes of training conducted number of employees trainedOther safety/health improvements Implementation & improvements in safety & health programsreduction of violations in subsequent inspections

COSTS Federal government hours expendedOutcomes compared to other types of citation resolution

Page 22: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

DATA SOURCES• OSHA’s Integrated Management Information

System (IMIS)• Records on corporate-wide settlement

agreements• OSHA Data Initiative (ODI)• Area office case files• Review Commission Docket• Interviews with DOL staff

Page 23: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

DATA ANALYSIS1. Descriptive & comparative statistics analysis.

Determine similarities & differences in the characteristics of each type of settlement agreement & litigation by:

• Characteristics of the establishments involved• Types of inspections• Types of violations

2. Process analysis.Identify factors associated with each type of agreement & litigation.

3. Primary outcome analysis.Estimate changes in injury/illness rates by type of agreement & litigation.

4. Cost effectiveness analysis.Compare costs & outcomes, both primary & intermediate, for each type ofagreement & litigation.

Page 24: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Data Collection SummaryCategory Inf. Sett.

Agree.Formal Sett.

Agree.Corporate

Sett. Agree.Litigated

CasesControl Group

Source •IMIS files•Case files

•IMIS files•Case files

OSHA records •Review Com. files•IMIS files

IMIS files

Coverage Sample Sample Population Sample Population

Type of Data

•Injury/illness

•Types of violations

•Types of training

•Improv. to health plans

•Govt. hours

•Penalties

•Injury/illness

•Types of violations

•Types of training

•Improv. to health plans

•Govt. hours

•Penalties

•Injury/illness

•Types of violations

•Types of training

•Improv. to health plans

•Govt. hours

•Penalties•Affected Estab.

•Injury/illness

•Types of violations

•Types of training

•Improv. to health plans

•Govt. hours

•Penalties

•Injury/illness

•Types of violations

•Govt. hours

Page 25: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Inspected Establishments

Meeting In-Scope Criteria [a]

IMIS

ODI

Inspected Establishments

Meeting In-Scope Criteria With

Citations

Sample List:In-Scope Establishments

Chosen in Stratified Random Sample With Associated IMIS

Log Data and ODI Data

In-Scope Establishments That Are Not Chosen

For The Sample

Sample: In-Scope Establishments Chosen

in Stratified Random Sample With Associated IMIS Log Data

and ODI Data and With Additional Outcome and Cost Data From Case File Review

Injury/Illness and Violations Reduction Outcomes Analysis:

Statistically Compare Violation and Injury/Illness Outcomes Between

Treatment Groups (Informal SA, Formal SA, Corporate SA, and Litigated Cases) and Control Group (Just Citations) For

Sample of Establishments

Cost-Effectiveness Analysis: Compile and Compare Outcomes and Costs For Each Treatment Group and

Control Group

Sampling Plan

Info

rmat

ion

Sour

ces

In-s

cope

In

spec

tions

/Cas

esA

naly

ses

Out

-Of-

Scop

e In

spec

tions

/Cas

esStep 1: Develop Sampling Frame, Characterize

Groups, and Analyze ProcessStep 2: Choose Sample of Establishments and

Analyze Violation and Injury/Illness OutcomesStep 3: Obtain Detailed Outcome

Information and Analyze Cost-Effectiveness

OSHRC Docket

OSHA Records on Corporate Settlements

Sampling Frame:In-Scope Establishments

With:-Informal SA-Formal SA

-Corporate SA-Litigated Cases-Just Citation (no

agreement or litigation)

Recommendations For Improving the Cost-Effectiveness of Programs

IMIS Log Data

Approach For Estimating The Relative Cost Effectiveness of Settlement Agreements and Litigated Cases

Descriptive and Comparative Statistics For Each Group

In-Scope Criteria (Inspections/Cases Not Meeting Criteria are out of scope)

Jurisdiction: Federal jurisdiction.Sector: General industry (non-construction).Time frame criteria: Calendar year 2000-2003 except for corporate

settlement agreements which will use a 1991/1992-2003 time frame.Case status criteria: Cases opened in the time period with a disposition.

Process Analysis: Identification of Factors Associated With Entering Into Each Type of Agreement or

Litigation

Inspected establishments without citations Uninspected establishments

Area Director and Solicitors

Interviews

Case File Review

Page 26: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Analytic Step 1

• Develop checklist/data collection form for case file review.

• Extract data:• Implementation of training programs• Number of workers trained• Frequency of training• Implementation or improvements of safety/health

program• Details of safety & health program• Hiring of a consultant• Inclusion of other establishments owned by same

firm• Other concessions

Page 27: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Analytic Step 2

Conduct interviews to collect information not available in databases

• To systematically gather information about the processesinvolved with the different types of citation resolution

• To understand regional variations in processes

• To identify “process patterns”

• To collect “examples of what works well”

Page 28: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Analytic Step 3

Develop analysis data fileThe database will begin with IMIS data on in-scopeInspections. This will be augmented with data from other sources:

• Case file reviews

• OSHA and Solicitor time sheet data

• ODI data on injuries and illnesses for before & after comparisons

Data validation checks will look for and resolve inconsistencies across and within various data sources.

Page 29: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Analytic Step 4Conduct analyses (descriptive, comparative, multivariate)1. Descriptive and comparative statistical analysis.

• Similarities & differences in the characteristics of each type of citation• Characteristics of the establishments involved• Types of inspections• Types of violations

2. Process analysis. Identify factors associated with entering into each type of citation resolution.

3. Primary outcome analysis. Estimate changes in injury/illness rates in establishments with each type of citation resolution.

4. Cost effectiveness analysis. Compare costs & outcomes. This will also address the additional costs of litigation compared to settlement agreements.

Page 30: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Program Evaluation of Law Enforcement Program

US Coast Guard Undocumented Migrant Interdiction Program

LCDR Eric Bernholz, CG-812

Page 31: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Purpose and Design of Program

• Purpose– Interdict alien migrants seeking to enter US

via sea IAW E.O. 12807, PDD-9, E.O. 13276• Focus on interdiction/repatriation at sea

– Multi-mission assets operating forward for both “layered defense” and safety of life at sea

– No other USG agency with appropriate capability, authority, and presence

Page 32: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Pieces of the Process

• Primarily maritime interdiction (“Patrolling), but…– Humanitarian (“Holding”)– Repatriation (“Returning”)– Transit to and from

Page 33: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

What we’re facing

Page 34: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Holding, Returning, Transiting• Tens to hundreds migrants involved in each event

• Basic sanitation, health care, and feeding• Interview process, consider what to do with migrants• Repatriate as appropriate – steam to next location

Page 35: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

D7 “area of responsibility”

AOPS “D7/LANT AMIO”

Cubans

Haitians Dominicans

Page 36: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Effectiveness

• In all these steps, the cutter is busy…– Can only hold so many migrants– Can’t patrol when full– Awaiting direction on repatriation

• …but not “effective” WRT stated measures– Not patrolling– Actions not directly affecting stated

performance goal of “Interdicting 87%* of migrants attempting to enter the US via maritime means”

Page 37: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Accounting for the Mission

AOPS “MIGRANT” Hours

Transit Repat. Holding Patrol

ToothTail

Page 38: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Program Performance

70

75

80

85

90

95

100

% Undocumented Migrants

Interdicted or Deterred

Target 87 87 87 87 87 87 88 89 91 93 95

USCG Actual 91.5 94.4 91.1 86.7 89 82.5 88.3 85.3 87.1

1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009

Note: Prior to 1999, the Migrant program set explicit performance standards, but did not utilize explicit performance targets.

Page 39: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Results of Program Evaluation• Program is largely effective

– Good program & management design– Flexible, decentralized execution– Tactical creativity, strong partnerships– Meets or nearly meets targets

• But– Late, informal planning & little or no reporting– Measures incomplete, potentially misleading– Weak management information systems (MIS),

inadequate data collection– No tools for resource & performance tradeoffs

Page 40: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Next Steps

• Improve execution of the current metric via regular updates of source countries and threat estimates.

• Tie goals explicitly to threat and resource levels, and update them regularly as levels change

• Define & track program efficiency measures

Page 41: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Next Steps• Implement new capabilities

– Better data collection and improved reporting MIS– Improved metrics– Resource-constrained, threat-dependent goals– Efficiency and proxy measures– Quantitative tools linking performance to threats and

resources– Regularly-updated goals, metrics, and planning

guidance, incorporating MPAR feedback– “What-if” capability for changing threats, resources,

policy, doctrine (within and across programs)

Page 42: Evaluation Dialogue Between OMB Staff and Federal Evaluation … · 2018. 8. 10. · Evaluation Dialogue Between OMB Staff and Federal Evaluation Leaders Digging a Bit Deeper into

Contributor Acknowledgements• LCdr Eric Bernholz, USCG• Alan Ginsburg, ED• Marcelle Habibion, VA• John Heffelfinger, EPA• David Introcaso, HHS• Cheryl Oros, USDA• N.J. Scheers, CPSC• Stephanie Shipman, GAO• Linda Stinson, DOL• Bill Valdez, DOE