61
1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health Authority (“FH”) authorizes the use, reproduction and/or modification of this publication for purposes other than commercial redistribution. In consideration for this authorization, the user agrees that any unmodified reproduction of this publication shall retain all copyright and proprietary notices. If the user modifies the content of this publication, all FH copyright notices shall be removed, however FH shall be acknowledged as the author of the source publication. Reproduction or storage of this publication in any form by any means for the purpose of commercial redistribution is strictly prohibited. This publication is intended to provide general information only, and should not be relied on as providing specific healthcare, legal or other professional advice. The Fraser Health Authority, and every person involved in the creation of this publication, disclaims any warranty, express or implied, as to its accuracy, completeness or currency, and disclaims all liability in respect of any actions, including the results of any actions, taken or not taken in reliance on the information contained herein.

1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Embed Size (px)

Citation preview

Page 1: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

11

Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation

and Performance

© Fraser Health Authority, 2009

The Fraser Health Authority (“FH”) authorizes the use, reproduction and/or modification of this publication for purposes other than commercial redistribution. In consideration for this authorization, the user agrees that any unmodified reproduction of this publication shall retain all copyright and proprietary notices. If the user modifies the content of this publication, all FH copyright notices shall be removed, however FH shall be acknowledged as the author of the source publication.

Reproduction or storage of this publication in any form by any means for the purpose of commercial redistribution is strictly prohibited.

This publication is intended to provide general information only, and should not be relied on as providing specific healthcare, legal or other professional advice. The Fraser Health Authority, and every person involved in the creation of this publication, disclaims any warranty, express or implied, as to its accuracy, completeness or currency, and disclaims all liability in respect of any actions, including the results of any actions, taken or not taken in reliance on the information contained herein.

Page 2: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Outline

Objectives Background Anatomy Good Indicators Variation

Page 3: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Objectives

Understand and appreciate the purpose of indicators

Identify criteria for designing, appraising or choosing indicators

Apply indicators for research, evaluation or quality improvement.

Page 4: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Background

What is an indicator? Succinct measures that describe as much as possible

about a system Why use indicators?

Understanding – how a system works and how it might be improved (research)

Performance monitoring – to assess if the system is performing to at an expected level, compare and improve a system (improvement)

Accountability – to inform assessment of effectiveness, efficiency and of responsibility (evaluation)

Page 5: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 6: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Pop Quiz

Page 7: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 8: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Considerations Indicators indicate:

Cannot capture and reflect the complexity of a system program or service

Must be considered within context Encourage explicitness:

Help to clarify our understanding of a system Facilitate communication of expectations

Numerically based: Rates, ratios etc.

Not specifically designed to detect fault: Can be used for interpreting both strengths and

weaknesses

Page 9: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Improvement

Indicators primarily used to measure systems and outcomes in health care. Goal - improvement Understand how things work leads to

understanding of how things can be done better

Measurement alone does not lead to improvement.

Page 10: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 11: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Anatomy

Basic construction How to deconstruct and assess

Metadata (indicator) Title, rationale, and how indicator is

defined/constructed Data

Information that is entered into the indicator

Page 12: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Metadata

Metadata will help asses if an indicator is: Important and relevant Able to be populated with reliable data Likely to have desired effect when

communicated well

Page 13: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Metadata

What is being measured? Why is it being measured? How is this indicator defined? Who/what does it measure? When does it measure? Absolute or proportion? What is the source of data? What is the accuracy/completeness of the data? Any considerations, warnings, caveats? Are specific tests needed to test the meaning of the

data?

Page 14: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Statistics Canada – Catalogue no. 82-221-X

Statistics Canada – Health Indicators

Page 15: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Definition: Wait time for hip fracture surgery (same/next day) Proportion with surgery same or next day: risk-adjusted

proportion of hip fracture patients aged 65 and older who underwent hip fracture surgery on the day of admission or the next day.

Wait time for surgery following hip fracture provides a measure of the access to care. While some hip fracture patients need medical treatment to stabilize their condition before surgery, research suggests patients typically benefit from timely surgery in terms of reduced morbidity, mortality, pain, length of stay in hospital, as well as improved rehabilitation.

Rates for Quebec are not available due to differences in data collection.

Source(s): Canadian Institute for Health Information, Discharge Abstract

Database.

Page 16: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Group Activity

Deconstruct the Wait Time for Hip Fracture Surgery indicator using the Metadata Assessment Questions in the handout.

Identify information that you would need to fully understand the composition of this indicator.

Report back.

Page 17: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Data

Is the indicator populated with the best available data? Reliable Valid Bias Completeness Error Convenience

Page 18: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Acronyms

Page 19: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Good Indicators

SMART

S SpecificM MeasureableA AchievableR RelevantT Time-bound

DUMB

D DoableU UseableM MeasurableB Believable

Page 20: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Criteria for Good Indicators

1. Importance and relevance

2. Validity

3. Feasibility

4. Meaning

5. Implications

Adapted from: The Good Indicators Guide: Understanding how to use and choose indicators. NHS Institute for Innovation and Improvement, 2008http://www.apho.org.uk/resource/item.aspx?RID=44584

Page 21: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

1. Importance and Relevance

Does the indicator measure what is relevant in the system?

Indicators must relate to objectives of the system Key parts of process and/or outcome should have

associated indicators Is there balance?

All important system components are covered No over or under-represented components

Will the indicators promote consensus? Well formed indicators can be helpful in developing

consensus on the objectives of the system

Page 22: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

2. Validity Does the indicator actually measure what it is

intended to measure? Accuracy and degree to which the indicator actually

measures the construct May require validation if existing valid indicators are

not available or appropriate• Translation validity

Face validity Content validity

• Criterion-related validity Predictive validity Concurrent validity Convergent validity Discriminant validity

http://www.socialresearchmethods.net/kb/measval.php

Page 23: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

3. Feasibility

Are reliable data available to populate the indicator?

Are data available at appropriate times? Are comparator data similarly available? Can the cost/resource allocation be

justified?

Page 24: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

4. Meaning

Will the indicator be sensitive to detect and demonstrate meaningful changes in the system? (it should not identify random variations)

Can high and low indicator values provide appropriate signals to take action?

Can the indicator be deconstructed to provide insight about specific results or patterns?

Will others accept the indicator as having meaning?

Page 25: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

5. Implications

What will you do if an indicator suggests further action?

Will indicator results induce perverse incentives and unintended uses?

Is there sufficient lag time in measurement so as to allow interventions to have an effect?

Page 26: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

MacNamara Fallacy

The MacNamara Fallacy is named after the Robert McNamara, the US Secretary of Defence in the 1960s who was obsessed with quantifying the Vietnam War in a way that tended to ignore what was truly going on.

MacNamara argued that the ratio of Viet Cong losses to US losses was an important measure of effectiveness

Page 27: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

MacNamara Fallacy

The first step is to measure whatever can be easily measured. This is OK as far as it goes.

The second step is to disregard that which can’t be easily measured or to give it an arbitrary quantitative value. This is artificial and misleading.

The third step is to presume that what can’t be measured easily really isn’t important. This is blindness.

The fourth step is to say that what can’t be easily measured really doesn’t exist. This is suicide.

Charles Handy, ‘The Empty Raincoat’, page 219.

Page 28: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Qualitative Indicators

Qualitative indicators such as perceptions are important.

Qualitative indicators may enable or hamper improvement/change.

It is important to balance quantitative with qualitative indicators to provide context

Page 29: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

SPICED IndicatorsSubjective: Informants have a special position or experience that gives

them unique insights which may yield a very high return on the investigators time.

Participatory: Indicators should be developed together with those best placed to assess them.

Interpreted and Communicable: Locally defined indicators may not mean much to other stakeholders, so they often need to be explained.

Cross-Checked and Compared: The validity of assessment needs to be cross-checked, by comparing different indicators and progress, and by using different informants, methods, and researchers.

Empowering: The process of setting and assessing indicators should be empowering in itself and allow groups and individuals to reflect critically on their changing situation.

Diverse and Disaggregated: There should be a deliberate effort to seek out different indicators from a range of groups. This information needs to be recorded in such a way that differences can be assessed over time.

Roche (2002)

Page 30: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 31: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Program Theory

A statement of goals accompanied by the

underlying assumptions that guide a

system/program/service delivery strategy

and are believed to be critical to producing

the desired outcomes.

Page 32: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Program Theory Considerations

Who are you serving? What are you striving to accomplish for the

population that you are serving? What strategies do you believe will help

you successfully accomplish your goals?

Page 33: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Logic Model

A program logic model links outcomes with program activities … and the theoretical principles of the program” (Kellogg, 2001)

Thus, logic models set up both formative and summative evaluation questions

Evaluative answers are “useful” when they reduce the risks of making the wrong decision

Page 34: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Types of Evaluation

Formative “Improve” Periodic and timely Focus on program

activities and outputs Leads to early

recommendations for program improvement

Summative “Prove” Were resources

committed worthwhile Focus on outcomes

and impact Measures value of

program based on impact

* Kellogg logic model development guide

Page 35: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 36: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

ResourcesDental Clinic Coordinator

Community Health Director

Staff dentist

Staff pediatrician

Medical providers

Money for supplies

ActivitiesTraining•Develop curriculum•Two one-hour didactic trainings to medical providers in oral health assessment•One-on-one training to medical providers on oral health

Outreach•Order dental supplies for packets•Make up packets•Distribute to parents at end of each visit

Outputs

Training# of two-hour trainings held# of one-on-one trainings held# of medical providers trained

Outreach# of parents/children receiving packets

Outcomes

Medical providers demonstrate accurate oral health assessment, education and prevention activities

More children receive high-quality oral health assessment, education and prevention activities during well-child visits

Parents/children are more knowledgeable about oral health and caring for children’s teeth

Reduced incidence of caries in children at the community health center

Program Goal: To improve the oral health of low-income children who receive primary care in a community health center

Example Logic Model

Page 37: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Example – Logic Model Linked with Indicators

Page 38: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Framework

Page 39: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Indicators Linked with Program Framework

Page 40: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Group Activity

Using the DoctorDad logic model: Develop one indicator for an anticipated output Develop one indicator for an anticipated outcome

Create a definition and identify sources of data Consider your indicator with respect to Indicator

Quality and Planning checklist. Report back.

Page 41: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health
Page 42: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Variation

indicators indicate – variation informs action

Objectives: Understanding variation. Know when to investigate. How to respond to variation.

Page 43: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Statistical Process Control

Common Cause Variation Everyday and inevitable variation which tends

to be small, with observed values close to the average.

Special Cause Variation Variation outside the historical experience

base. Signal of some important change in the

system.

Page 44: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Statistical Process Control

Statistical Process Control (SPC) is an effective method of monitoring a process through the use of control charts.

Control charts enable the use of objective criteria for distinguishing background variation from events of significance based on statistical techniques.

Page 45: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Statistical Process Control

Allows you to determine: That the system is working with an acceptable level of

performance and there are no outliers• No action needed

That the system is working with an acceptable level of performance and there are outliers

• Address outliers That the system’s average level of performance is not

acceptable• Address entire system

Page 46: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

SPC Run Chart

Time ordered presentation of observations

A centre line (average or median) of all the observations plotted.

Page 47: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

SPC Control Chart

three basic components:

a centre line, usually the average of all the samples plotted.

upper and lower statistical control limits that define the constraints of common cause variations.

performance data plotted over time.

Page 48: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

SPC Run Chart Example

Steps to create a Run Chart Ideally, there should be a minimum of 15 data points. Draw a horizontal line (the x-axis), and label it with the

unit of time. Draw a vertical line (the y-axis), and scale it to cover the

current data, plus sufficient room to accommodate future data points. Label it with the outcome.

Plot the data on the graph in time order and join adjacent points with a solid line.

Calculate the mean or median of the data (the centre line) and draw this on the graph.

Page 49: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart Run chart of the number of red beads drawn across 25

draws

Page 50: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart – What to look for:

Useful Observations – number of observations that do not fall directly on centre line.

Run – sequence of one or more consecutive useful observations on the same side of the centre line.

Trend – sequence of successive increases or decreases in useful observations

Page 51: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart RulesIdentifying Special Cause

VariationNumber of Runs If there are too few or too many runs in the process.

Page 52: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart RulesIdentifying Special Cause

VariationNumber of Runs If there are too few or too many runs in the process.

Page 53: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart RulesIdentifying Special Cause

VariationShift If the number of successive useful observations falling

on either side of the centre line is greater than 7.

Page 54: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart RulesIdentifying Special Cause

VariationTrend If the number of successive useful observations increasing or

decreasing is greater than 7.

Page 55: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Run Chart RulesIdentifying Special Cause

VariationZig-Zag If the number of useful observations increasing or decreasing

alternately (zig-zag pattern) is greater than 14.

Page 56: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Control Chart RulesIdentifying Special Cause

VariationControl Limits – 2SD above or below centre line If there is one or more observations beyond the control limits

Warning Limits – 3SD above or below centre line If there are two successive observations beyond the control limits

Control Limit2SD

Warning Limit3SD

Page 57: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Example – uniform system underperforming

Page 58: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Control Chart Resource

http://www.indicators.scot.nhs.uk/SPC/SPC.html

Page 59: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Addressing Special Cause Variation

When you detect a special cause: Control any damage or problems with an immediate, short-term fix. Once a quick fix is in place, search for the cause. Ask people in the process what was

different that time. What was out of the ordinary? It might not have been much – an unexpected emergency, a change in schedules, or new employees.

Once you have discovered the special cause, you can develop a longer-term remedy. Most special causes have a negative impact on the output of the process and need to be removed.

Occasionally, a special cause can have a positive impact depending on the nature of the process. If this is the case, finds ways to capture and integrate it into the system.

Avoid these mistakes: Be careful not to view short-term fixes as a permanent solution or the process will

never be improved. Changing the process to accommodate the special cause. This usually adds cost and

bureaucracy. Blaming individuals. Warn employees to do better. People can only do as well as the system allows.

Page 60: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health

Addressing Common Cause Variation

Common Cause Variation – ideal situation for quality improvement

If a process indicator is stable, or in statistical control, does not mean that its results are satisfactory.

An indicator may be very consistent, but not meeting an expected outcome

Variation can be systematically reduced, even in stable processes, enabling a gradual tightening of control limits, and an overall increase in quality.

Page 61: 1 1 Best Practices in Selecting, Using and Assessing Indicators for Research, Evaluation and Performance © Fraser Health Authority, 2009 The Fraser Health