33
MSE Performance Metrics, Tentative Results and Summary Joint Technical Committee Northwest Fisheries Science Center, NOAA Pacific Biological Station, DFO School of Resource and Environmental Management, SFU

MSE Performance Metrics, Tentative Results and Summary

  • Upload
    jatin

  • View
    31

  • Download
    0

Embed Size (px)

DESCRIPTION

MSE Performance Metrics, Tentative Results and Summary. Joint Technical Committee Northwest Fisheries Science Center, NOAA Pacific Biological Station, DFO School of Resource and Environmental Management, SFU. Outline. Summarize the hake MSE Example simulations - PowerPoint PPT Presentation

Citation preview

Page 1: MSE Performance  Metrics,  Tentative  Results and Summary

MSE Performance Metrics, Tentative Results and Summary

Joint Technical CommitteeNorthwest Fisheries Science Center, NOAA

Pacific Biological Station, DFOSchool of Resource and Environmental Management, SFU

Page 2: MSE Performance  Metrics,  Tentative  Results and Summary

Outline

• Summarize the hake MSE • Example simulations • Performance metrics • Summary figures

Page 3: MSE Performance  Metrics,  Tentative  Results and Summary

Objectives of the MSE

• Use the 2012 base case as the operating model.

• As defined in May 2012– Evaluate the performance of the harvest control

rule– Evaluate the performance of annual, relative to

biennial survey frequency.

Page 4: MSE Performance  Metrics,  Tentative  Results and Summary

Organization of MSE Simulations

Operating Model* Stock dynamics* Fishery dynamics* True population

Management Strategy* Data choices* Stock Assessment* Harvest control rule

CatchData

Performance Statistics* Conservation

objectives* Yield objectives* Stability objectives

Feedback

Loop

Use the MPD (not posterior medians, or other quantiles) for applying the harvest control rule

Page 5: MSE Performance  Metrics,  Tentative  Results and Summary

1960 1970 1980 1990 2000 2010 2020 2030

0.0

0.5

1.0

1.5

2.0

2.5

3.0

Year

SS

Bt

Existing (2012) assessment MSE Simulations

Page 6: MSE Performance  Metrics,  Tentative  Results and Summary

Cases Considered

• No fishing• Perfect Information Case• Annual Survey • Biennial Survey

Page 7: MSE Performance  Metrics,  Tentative  Results and Summary

Perfect Information Case

• We created a reference, perfect information case where we simulated data with no error

• The purpose of the perfect information case was as follows:– To separate observation vs process error i.e. variable

data don’t affect management procedure performance

– to provide a standard relative to which a comparison of the test (biennial and annual) cases could be made

Page 8: MSE Performance  Metrics,  Tentative  Results and Summary

Perfect information case

• Every year operating model simulates dynamics of the stock (i.e. recruitments, stock size etc)

• No assessment model is fit, simulated catches come from the application of the control rule to the true stock

Page 9: MSE Performance  Metrics,  Tentative  Results and Summary

Biennial Survey Case• Every year operating model simulates dynamics of the stock (i.e.

recruitments, stock size etc)• Every odd year operating model simulates and assessment

model fits:– catch– survey age composition data– commercial age composition data– survey biomass

• In even years operating model simulates and assessment model fits– catch– commercial age composition data

Page 10: MSE Performance  Metrics,  Tentative  Results and Summary

Annual Survey Case

• Every year operating model simulates dynamics of the stock (i.e. recruitments, stock size etc)

• Every year operating model simulates and assessment model fits:– catch– survey age composition data– commercial age composition data– survey biomass

Page 11: MSE Performance  Metrics,  Tentative  Results and Summary
Page 12: MSE Performance  Metrics,  Tentative  Results and Summary

But remember – starting points are not the same for each MSE run

Page 13: MSE Performance  Metrics,  Tentative  Results and Summary
Page 14: MSE Performance  Metrics,  Tentative  Results and Summary
Page 15: MSE Performance  Metrics,  Tentative  Results and Summary
Page 16: MSE Performance  Metrics,  Tentative  Results and Summary

Measuring Performance• Choose metrics that capture the tradeoffs between conservation,

variability in catch and total yield for specific time periods.• Define short, medium and long time periods as Short=2013-2015,

Medium=2016-2020, Long=2021-2030.• The main conservation metric is the proportion of years depletion

is below 10%• The main variability in catch metric is the Average Annual

Variability in catch for a given time period.• For yield we used the median average catch• We’ve chosen what we think are the top six. We’d like to discuss

if others are needed.

Page 17: MSE Performance  Metrics,  Tentative  Results and Summary

Key Performance Statistics

Medium 2016-2020 Perfect Information Annual Biennial

Median average depletion 28% 27% 28%

Proportion of years below SB10% 1% 7% 6%Proportion of years between SB10% and SB40% 70% 61% 58%

Proportion of years above SB40% 29% 32% 36%

Median Average Annual Variability (AAV) in catch 23% 35% 36%

Median Average Catch 216 219 211

Page 18: MSE Performance  Metrics,  Tentative  Results and Summary

Other available options• First quartile depletion• Third quartile depletion• Median final depletion• Median of lowest depletion• Median of lowest perceived depletion• First quartile of lowest depletion• Third quartile of lowest depletion• First quartile of AAV in catch• Third quartile of AAV in catch• First quartile of average catch• Third quartile of average catch• Median of lowest catch levels• First quartile of lowest catch levels• Third quartile of lowest catch levels• Proportion with any depletion below SB10%• Proportion perceived to have any depletion below SB10%

Page 19: MSE Performance  Metrics,  Tentative  Results and Summary

Statistics Break - Medians vs Means

Page 20: MSE Performance  Metrics,  Tentative  Results and Summary

Average Annual Variability in Catch (illustration)

Page 21: MSE Performance  Metrics,  Tentative  Results and Summary

Comparisons of Depletion, Catch and AAV for All Cases

Page 22: MSE Performance  Metrics,  Tentative  Results and Summary
Page 23: MSE Performance  Metrics,  Tentative  Results and Summary
Page 24: MSE Performance  Metrics,  Tentative  Results and Summary

Summary for long-term depletion

Page 25: MSE Performance  Metrics,  Tentative  Results and Summary

Summary for long term AAV

Page 26: MSE Performance  Metrics,  Tentative  Results and Summary

Summary for long-term catch

Page 27: MSE Performance  Metrics,  Tentative  Results and Summary

Discussion

• Next steps

Page 28: MSE Performance  Metrics,  Tentative  Results and Summary

Alternative Analyses

Page 29: MSE Performance  Metrics,  Tentative  Results and Summary

Analysis of alternative target harvest rates

• The hake treaty doesn't specify a target depletion level, only a target harvest rate (F40%) and a control rule (40-10).

• This makes it difficult to evaluate the efficacy of the control rule (i.e. relative to what?)

• One additional curiosity that we considered was what would the target harvest rate have to be in order to achieve a range of target depletion levels

• The MSE can be used to explore how changes to the target harvest rate might affect depletion, AAV, and average catch.

• This is an exploration of trade-offs, not a proposal to change the hake treaty.

Page 30: MSE Performance  Metrics,  Tentative  Results and Summary

Alternative target harvest rates

Page 31: MSE Performance  Metrics,  Tentative  Results and Summary
Page 32: MSE Performance  Metrics,  Tentative  Results and Summary
Page 33: MSE Performance  Metrics,  Tentative  Results and Summary

Discussion

• Does the groups want alternative performance statistics considered

• Progress and next steps