101
Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Embed Size (px)

Citation preview

Page 1: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Evaluating Outcomes

Marshall (Mark) Smith, MD, PhDDirector, Simulation and Innovation

Banner Health

Page 2: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Lecture today is adapted from presentation given at last Laerdal SUN meeting by Geoffrey T. Miller, and some of the slides today are used/modified

from that presentation.

Geoffrey T. MillerAssociate Director, Research and Curriculum Development

Division of Pehospital and Emergency HealthcareGordon Center for Research in Medical Education

University of Miami Miller School of Medicine

Acknowledgement and Appreciation

Page 3: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

The wisdom of a

group is greater

than that of a

few experts…

Page 4: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Session Aims

• Discuss the importance and use of outcomes

evaluation and challenges to traditional assessments

• Discuss some learning models that facilitate

developing assessments

• Discuss the importance of validity, reliability and

feasibility as it relates to assessment

• Discuss types of assessments and their application in

healthcare education

Page 5: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Evaluation

• Systematic determination of merit, worth, and

significance of something or someone using

criteria against a set of standards.

– Wikipedia, 2009

Page 6: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment

• Educational Assessment is the process of

documenting, usually in measurable terms,

knowledge, skills, attitudes and beliefs.

Assessment can focus on the individual

learner,……– Wikipedia, 2009

Page 7: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment vs Evaluation

• Assessment is about the progress and

achievements of the individual learners

• Evaluation is about the learning program as a

whole

Tovey, 1997

Page 8: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

So why measure anything?

Why not just teach?

Page 9: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Measurement

• What is measured, improves

• You Can't Manage What You Don't Measure

• You tend to improve what you measure

Page 10: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Measurement

• Promotes learning

• Allows evaluation of individuals and learning

programs

• Basis of outcomes- or competency-based education

• Documentation of competencies

Page 11: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Measurements in Future

• Credentialing• Privileging • Licensure • Board certification • High stake assessments for practitioners

All involve assessment of competence

Page 12: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

What are the challenges today

of traditional methods of

measurement/assessment

for healthcare providers?

Page 13: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Challenges in traditional assessments

• Using actual (sick) patients for evaluation of skills

– Cannot predict nor schedule clinical training events

– Compromise of quality of patient care, safety

– Privacy concerns

– Patient modesty

– Cultural issues

– Prolongation of care (longer procedures, etc)

Page 14: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Challenges in traditional assessments

• Challenges with other models

– Cadaveric tissue models

– Animal models / labs

Page 15: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Challenges in traditional assessments

• Feasibility issues for large-scale examinations

• Standardized, perceived fairness issues in high-stakes

settings

• Standardized patients (SPs) improve reliability, but

validity issues exist: cannot mimic many physical

findings

Page 16: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Challenges in traditional assessments

• Wide range of clinical problems, including

rare and critical events

• Availability

• Financial cost

• Adequate resources

• Reliability, validity, feasibility

Page 17: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

• Reaction

• Learning

• Performance

• Results

Page 18: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

1. Reaction

• Measures only one thing – learners perception

• Not indicative of any skills, performance

• Success is critical to success of program

• Relevance to learner important

Page 19: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

2. Learning

• This is where learner changes

• Requires pre and post testing

• Evaluation at this step is through learner

assessment

• First level to measure change in learner!

Page 20: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

3. Performance (Behavior)

• Action that is performed• Consequence of behavior is performance• Traditionally involves measurement in the

workplace• Transfer of learning from classroom to work

environment

Page 21: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

4. Results

• Clinical and quality outcomes

• Difficult to measure in healthcare

• Perhaps better in team training

• Often ROI that management wants

Page 22: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

• Reaction

• Learning

• Performance

• Results

1. Increasing complexity

2. Increasing difficulty to

measure, time

consuming

3. Increasing value!

Page 23: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Kirkpatrick's Four Levels of Evaluation

• Reaction

• Learning

• Performance

• Results

Page 24: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

How do we start to develop outcome measurements

Page 25: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health
Page 26: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Learning Outcomes

Projected Outcomes

Course Objectives

Goals

Expected Outcomes

Page 27: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Development of Curricula

• Analysis– Clearly define and clarify desired outcomes*

• Design • Development• Implementation • Evaluation

ADDIE

Page 28: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Defining Assessments

• Outcomes are general, objectives are specific

and support outcomes

• If objectives are clearly defined and written,

questions and assessments nearly write

themselves

Page 29: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Defining Outcomes

• Learners are more likely to achieve competency and

mastery of skills if the outcomes are well defined and

appropriate for the level of skill training

• Define clear benchmarks for learners to achieve

• Clear goals with tangible, measurable objectives

• Start with the end-goal in mind and the assessment

metrics, then the content will begin to develop itself

Page 30: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Role of Assessment in Curricula Design

Course •Teaching and learning•Assessment and evaluation

Refine •Learner and Course Outcomes•Modify curricula/assessments

Course •Teaching and learning•Assess learners

Page 31: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Use of assessments in healthcare simulation

InformationInformation

DemonstrationDemonstration

PracticePractice

Rosen, MA et al. Measuring Team Performance in Simulation-Based Training: Adopting Best Practices for Healthcare.Simulation in Healthcare 3:2008;33–41.

FeedbackFeedback RemediationRemediation

MeasurementMeasurement DiagnosisDiagnosis

Page 32: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Preparing assessments

• What should be assessed?

– Any part of curriculum considered essential

and/or has significant designated teaching time

– Should be consistent with learning outcomes that

are established as the competencies for learners

– Consider weighted assessments

Page 33: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Clinical competence and performance

• Competent performance requires acquisition of basic

knowledge, skills & attitudes

• Competence =

– Application of specific KSAs

• Performance =

– Translation of competence into action

Page 34: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Three Types of Learning(Learning Domains)

Bloom's Taxonomy

• Cognitive: mental skills (Knowledge)

• Psychomotor: manual or physical skills (Skills)

• Affective: growth in feelings or emotional areas

(Attitude)

Page 35: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Three Types of Learning

Bloom's Taxonomy

• Cognitive = Knowledge K

• Psychomotor = Skills S

• Affective = Attitude A

Page 36: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Knowledge

Page 37: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Knowledge

Page 38: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Knowledge

Page 39: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Knowledge

Page 40: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Knowledge

Page 41: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

The Anti – Blooms…

Page 42: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Skills

• Bloom’s committee did not propose a

compilation of the psychomotor domain

model, but others have since.

Page 43: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Bloom’s Taxonomy – Attitude

Five Major Categories• Receiving phenomena• Responding to Phenomena• Valuing• Organization• Internalizing values

Page 44: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Knowledge Competencies

KnowledgeCognitive knowledge• (factual) Recall• Comprehension• Application• Analysis• Synthesis• Evaluation

Page 45: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Knowledge

Skills

Skills• Communication• Physical Exam• Procedures• Informatics• Self Learning• Time Management• Problem Solving

Skill competencies

Page 46: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Knowledge

Skills

Attitudes• Behavior• Teamwork• Professionalism• Key Personal

Qualities• MotivationAttitudes

XAttitude competencies

Page 47: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Continuous process

Knowledge

SkillsAttitudes

Page 48: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Possible Outcome Competencies(GME Based)

• Patient care• Medical knowledge• Practice-based

learning and improvement

• Interpersonal and communication skills

• Professionalism • Systems-Based

Practice

Knowledge

SkillsAttitudes

Page 49: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

So we know what we

want to measure, but

how do we do that?

Page 50: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health
Page 51: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Copyright ©2003 BMJ Publishing Group Ltd.

Norcini, J. J BMJ 2003;326:753-755

Miller’s Pyramid of Competence

Page 52: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Miller’s Pyramid

• Top two cells of the pyramid, in the domains

of action, or performance, reflect clinical

reality

• The professionalism and motivation required

to continuously apply these in the real setting

must be observed during actual patient care.

Page 53: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Miller’s Pyramid

• Top two levels most difficult to measure

• Quality of assessment in the clinical setting lags

far behind

• In training Evaluation Reports and Likert Scale

• Little value as formative, or feedback, instrument

that might contribute to the learner’s education

Page 54: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Miller’s Pyramid of

competence for

learning and assessment

Page 55: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Miller’s Pyramid of Competence

Miller GE. The Assessment of Clinical Skills / Competence / Performance, Academic Medicine, 65:9, S63-S67.

Does

Shows

Knows How

KnowsCognition

Behavior

Page 56: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

LearningOpportunity• Reading /

Independent Study

• Lecture• Computer-

based• Colleagues /

Peers

Teaching and Learning “Knows”

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance, Academic Medicine, 65:9, S63-S67.

Page 57: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment of “Knows”

Factual Tests

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 58: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment Tools for “Knows”

• Multiple Choice Questions (MCQs)

• Short Answer

• True / False

• Matching (extended)

• Constructed Response Questions

Page 59: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

LearningOpportunity• Problem-based

Ex.• Tabletop

Exercises• Direct

Observation• Mentors

Teaching and Learning - “Knows How”

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 60: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Clinical Context

Based Tests

Assessment of “Knows how”

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 61: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

• Multiple-choice question

• Essay• Short answer• Oral interview

The Tools of “Knows How”

Page 62: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

LearningOpportunity• Skill-based

Exercises• Repetitive practice

• Small Group• Role Playing

Teaching and Learning - “Shows”

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 63: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment of “Shows”

PerformanceAssessment

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 64: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

• Objective Structured Clinical

Examination (OSCE)

• Standardized Patient-based

The Tools of “Shows”

Page 65: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Variables in Clinical Assessment

ClinicalAssessment

Examiner

Patient

Student

CONTROL VARIABLES as much as possible…

Page 66: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

LearningOpportunity• Experience

Teaching and Learning - “Does”

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 67: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment of “Does”

PerformanceAssessment

Does

Shows

Knows How

Knows

Miller GE. The Assessment of Clinical Skills / Competence / Performance

Page 68: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

• Undercover / Stealth / Incognito

Standardized Patient-based

• Videos of performance

• Portfolio of learner

• Patient Satisfaction Surveys

The Tools of “Does”

Page 69: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Influences on clinical performance

Does

Performance

CompetenceSyste

m related Individual related

Cambridge Model for delineating performance and competence

Rethans JJ, et al. The relationship between competence and performance: implications for assessing practice performance, Medical Education, 36:901-909.

Page 70: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Reliability

and

Validation

Page 71: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health
Page 72: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - Reliability

• Does the test consistently measure what it is

supposed to be measuring?

– Types of reliability:

• Inter-rater (consistency over raters)

• Test-retest (consistency over time)

• Internal consistency (over different items/forms)

Page 73: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Inter-rater Reliability

• Multiple judges code independently using the same criteria

• Reliability = raters code same observations into same classification

• Examples• Medical record reviews• Clinical skills• Oral examinations

Page 74: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Factors Influencing Reliability

• Test length• Longer tests give more reliable scores

• Group homogeneity• The more heterogeneous the group, the higher

the reliability• Objectivity of scoring• The more objective the scoring, the higher the

reliability

Page 75: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - Validity

• Are we measuring what we are supposed to be measuring

• Use the appropriate instrument for the knowledge, skill, or attitude you are testing

• The major types of validity should be considered– Face– Content– Construct

Page 76: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Both archers are equally reliableValidity = quality of archer’s hits

Archer 1 hits bulls eye every time

Archer 2 hits outer ring in same spot every time

Validity is accuracy

Page 77: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Reliable and Valid

Reliable, not valid

Not reliable, not validNot reliable, not valid

Reliability and Validity

Page 78: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Improving reliability and validity

• Base assessment on outcome/objectives- event triggers- observable behavior- behavioral rating-assess

• Define:– Low-medium-high performance– Use of rubric or rating metric– Use (video) training examples of performance– Employ quality assurance/improvement system

Page 79: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments types

Choose the appropriate

assessment method:

– Formative

– Summative

– Self

– Peer

Page 80: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment• Formative Assessment

– Lower stakes – One of several, over time of course or program– May be evaluative, diagnostic, or prescriptive– Often results in remediation or progression to next level

• Summative Assessment– Higher stakes – Generally final of course or program – Primary purpose is performance measurement– Often results in a “Go, No-Go” outcome

Page 81: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - self

• Encourages responsibility for the learning process, fosters skills in making judgments as to whether work is of an acceptable standard – it improves performance.

• Most forms of assessment can be adapted to a self-assessment format (MCQs, OSCEs, and short answers)

• Students must be aware of standards required for competent performance.

Page 82: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - peer

• Enables learners to hone their skills in their ability to work with others and professional insight

• Enables faculty to obtain a view of students they do not see

• An important part of peer assessment is for students to justify the marks they award to others

• Justification can also be used as a component when faculty evaluates attitudes and professionalism.

Page 83: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Setting Standards

• Should be set to determine competence• Enables certification to be documented, accountable

and defensible• Appropriately set standards for an assessment will

pass those students who are truly competent• Standards should not be two low (false positives) to pass

those who are incompetent, nor too high (false negative) to fail those who are competent.

Page 84: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Setting Standards

• Those responsible in setting standards must

also have a direct role in teaching students at

the level being examined and assist in

providing examination material

Page 85: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Setting Standards• Standards should be set around a core curriculum that

includes the knowledge, skills and attitudes required of all students

• When setting a standard the following should be considered:– What is assessed must reflect the core curriculum– Students should be expected to reach a high standard in

the core components of the curriculum (For instance an 80-90% pass mark of for the important core and 60-80% for the less important aspects.)

– Students should be required to demonstrate mastery of the core in one phase of the curriculum before moving on to the next part of the curriculum

Page 86: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - Feasibility• Is the administration and taking of the assessment

instrument feasible in terms of time and resources• The following questions should be considered:

– How long will it take to construct the instrument?– How much time will be involved with the scoring process?– Will it be relatively easy to interpret the scores and

produce the results?– Is it practical in terms of organization?– Can quality feedback result from the instrument?– Will the instrument indicate to the students the important

elements within the course?– Will the assessment have a beneficial effect in terms of

student motivation, good study habits and positive career aspirations?

Page 87: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Practicality

• Number of students to be assessed

• Time available for the assessment

• Number of staff available

• Resources/equipment available

• Special accommodations

Page 88: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment

Instruments

Page 89: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments - Instruments

• Be aware of the types of assessment

instruments available as well as the

advantages and disadvantages of each

• Use more than one assessment instrument

and more than one assessor if possible

when looking at skills and attitudes

Page 90: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Choosing appropriate assessment methods

• When choosing the assessment instrument,

the following should be answered:

–Is it valid

–Is it reliable

–Is it feasible

Page 91: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Knowledge Instruments

• Objective tests (short answer,

true/false, matching, multiple choice)

• Objective Structured Clinical

Evaluations (OSCEs)

• Constructed response questions

• Rating scales (used on clerkships)

Page 92: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Skill Instruments

• Objective tests (Simulation based)• OSCEs• Constructed response questions• Critical reading papers (interpreting literature)• Checklists• Rating Scales• Portfolios (self-evaluation, time management)

Page 93: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Weighted Checklists

• List of items to measure

• Set of weights of each

• Summary score

Page 94: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessments – Attitude Instruments

• Portfolios• Essays / Modified essay questions• OSCEs• Checklists• Rating scales• Patient management problems• Short/long case assessments

Page 95: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment Metrics

• Procedural or Check List assessment

• Global Rating assessment

Page 96: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment Metrics• Procedural or Check List assessment

BCLS Y N

Open Airway

Check Breathing

BCLS Y N

Open Airway(< 5 sec of LOC)

Check Breathing(< 5 sec of Airway)

BCLS Y N

Open Airway

Check Breathing

A

Rating Score +1 -1 0 *Assist

Page 97: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Assessment Metrics• Global Rating assessment

Code Blue P F

CPR and

ACLS

Code Blue

CPR<1(low) - 5(Hi)> points

ACLS<1(low)- 5(Hi)> points

Code Blue H M

CPR

ACLS

L

Rating Score +1 0 -1

Pts.

Page 98: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Review

• Assessment drives learning• Clearly define the desired outcome,

ensure that it can be measured• Consider the effectiveness of the

measurement• Feedback to individual candidates• Feedback to training programs

Page 99: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Questions and discussion

Page 100: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

Good Luck

Page 101: Evaluating Outcomes Marshall (Mark) Smith, MD, PhD Director, Simulation and Innovation Banner Health

101

CompetencyCompetency is noted when a learner is observed performing a

task or function that has been established as a standard by the profession. The achievement of professional competency requires the articulation of learning objectives as observable, measurable outcomes for a specific level of learner performance. Such specific detailing of performance expectations defines educational competencies. They are verified on the basis of evidence documenting learner achievement, and must be clearly communicated to learners, faculty, and institutional leaders prior to assessment.

____________Identified by members of work group on competency-based women’s health education at APGO Interdisciplinary Women’s Health Education Conference in September, 1996, Chantilly, VA.