Upload
nihr-clahrc-west-midlands
View
136
Download
0
Embed Size (px)
Citation preview
Rapid qualitative analysis:
does it deliver?
Beck Taylor, Clinical Research Fellow
Theme 1
06/07/2016
The team
Beck Taylor
Cathy Henshall
Ian Litchfield
Louise Bentham
Sara Kenyon
Sheila Greenfield
Our partners
Birmingham Womens Hospital, particularly the Home Birth Team
Birmingham South Central Clinical Commissioning Group
Evaluation of a service innovation
Asked by NHS to evaluate dedicated Home Birth Service
Model and its implementation evolving, not well-defined
Designed project based on Evaluability Assessment
methodology
Interviews (n=21) and focus group (n=13) with key stakeholders, plus
documentary analysis (n=9)
Explores: programme theory and fidelity; barriers/facilitators; available
data; areas for evaluation; recommendations for changes and further
evaluation
My background
Public health physician, working in predominantly
qualitative research since 2008
Pre-2008 conducted pragmatic evaluations and
assessments with key stakeholders in short timescales,
not methodologically robust!
Considerations
Desire to inform practice in real time
Understanding that stakeholders want key findings
rather than fine detail
BUT not at expense of academic rigour
Was there a way to analyse data and
deliver findings more quickly?
Sharing of model/lack of model quickly
Inform ongoing development and decision-making
Information not out of date, useful
What we did
Data gathered predominantly by BT
Analysed rapidly by BT and CH with input from SK
No coding summary templates used to manage
data (Alison Hamilton. Qualitative Methods in Rapid Turn-Around Health Services Research. Presented online for the US Department of Veterans Affairs (2013) )
Our Rapid Analysis approach
Summary template example
Questions arising
How does this approach work in practice?
Does it deliver findings more quickly than traditional
qualitative analysis?
Does it elicit similar findings to traditional approaches?
If not, how do they differ?
What impact might any differences in findings have?
What might the applications of this approach be?
Cant we do it the normal way and
see how the two compare?
(Thank you Dr Kenyon!)
Comparative analysis project
Repeat analysis of data using in-depth analysis,
Framework method
Independent, blinded researcher (Ian Litchfield)
Input from second researcher (Louise Bentham)
Oversight and methodological support from Prof Sheila
Greenfield
Comparison research questions
Qualitative
data
Rapid
analysis
In-depth
analysis
Comparison research questions
Qualitative
data
Rapid
analysis
In-depth
analysis
? ?
Comparison research questions
Qualitative
data
Rapid
analysis
Findings and
recommendations
In-depth
analysis
? ?
Comparison research questions
Qualitative
data
Rapid
analysis
Findings and
recommendations
In-depth
analysis
Similarities / differences
Importance
?
? ?
?
Approach to comparison (not much in the literature)*
1) Time
Researcher timesheets
Total time + time for specific tasks
2) Findings and recommendations
Independent review match/partial match/no match
Check as a team
Quantitative and qualitative summaries of similarities/differences
*Burgess-Allen J, Owen-Smith V. Using mind mapping techniques for rapid qualitative data analysis in
public participation processes. Health expectations : an international journal of public participation in
health care and health policy. 2010;13(4):406-15.
*Putten JV, Nolen AL. Comparing Results from Constant Comparative and Computer Software Methods:
A Reflection about Qualitative Data Analysis. Journal of Ethnographic & Qualitative Research.
2010;5(2):99-112.
Results: time
Data management was much faster in rapid analysis (RA)
(In-depth analysis (IDA) 3x longer)
Interpretation took much longer in RA, but we think other
factors influenced this
(RA 6x longer than IDA)
Results: overlap in recommendations
and findings
SPECIFIC,
DETAILED
KEY
ISSUES
CONTEXT-
INFORMED,
INTERPRETIVE
In-depth
Analysis
Rapid
Analysis
Comparing the two methods
Rapid Analysis In Depth Analysis
Clinical Not clinical
Embedded in the field No prior exposure to field
BT collected the data Did not collect data
Using RA for first time learning time,
need to avoid usual practice
Experienced in method no method to
learn, doing what comes naturally
Shared office No space for informal reflection
Equal workload IL did the lions share of analysis
Main focus of work over short period Project squeezed in among other
commitments
Focused on producing and crafting
outputs for known stakeholders
Much less focused on the stakeholder
team
Reflections rapid analysis
Requires time discipline
Uncomfortable at times true to data?
How would this work for novice researchers?
How would this work in larger teams? Would it take longer
to synthesise data?
Would this work if we were not embedded/clinical?
Reflections comparing methods
Very time consuming
Defining and interpreting outcomes is not
straightforward what constitutes a finding or a
recommendation?
Comparing apples with oranges further work needed
Ideally need to compare two similar research teams
using different methods, but cost/capacity implications
Reflections applications of RA
Identifying headline/priority issues how much of the
detail do stakeholders actually use in practice?
Providing rapid findings where time of the essence
(provided it is truly rapid)
Identifying areas for more in-depth analysis
Further work
More comparable comparisons
Repeat comparisons in other contexts
Comparisons of other rapid approaches
Consultation with users of research outputs
Feedback and comments
very welcome
(feedback slips on your tables)