Seasonal Forecast Verification What Why and How Simon

  • Slides: 39
Download presentation
Seasonal Forecast Verification: What, Why and How? Simon J. Mason simon@iri. columbia. edu International

Seasonal Forecast Verification: What, Why and How? Simon J. Mason simon@iri. columbia. edu International Research Institute for Climate and Society The Earth Institute of Columbia University Cari. COF Seasonal Forecast Verification Training Workshop Port of Spain, Trinidad, 22 – 25 May, 2013

What makes a good forecast? Forecast: West Indies will win the 2012 ICC Twenty

What makes a good forecast? Forecast: West Indies will win the 2012 ICC Twenty 20 World Cup Verification: Correct! Correctness is only one aspect of forecast quality – how well do the forecasts correspond with the observations? 2

What makes a good forecast? Forecast: West Indies will win the 2012 ICC Twenty

What makes a good forecast? Forecast: West Indies will win the 2012 ICC Twenty 20 World Cup Final by 30 runs Verification: They won by 36 runs Accuracy may be more appropriate than correctness. 3

Definitions “I don’t like definitions” Mark Knopfler Correctness – did what the forecast say

Definitions “I don’t like definitions” Mark Knopfler Correctness – did what the forecast say would happen actually happen? Accuracy – is the forecast close to the observation? Precision – what level of accuracy is being implied? Sharpness – do the forecasts differ sometimes? 4

What makes a good forecast? Forecast: Sri Lanka will lose the 2012 ICC Twenty

What makes a good forecast? Forecast: Sri Lanka will lose the 2012 ICC Twenty 20 World Cup Final When: In the seventeenth over of the second innings Timeliness Example: Are seasonal climate forecasts issued too late for important decisions? 5

What makes a good forecast? Forecast: Trinidad and Tobago will do well in the

What makes a good forecast? Forecast: Trinidad and Tobago will do well in the 2012 Olympics Verification: ? 4 medals Ambiguity Example: Is the precise meaning of the forecasts clear? 6

What is the predictand? What is the seasonal forecast for Port of Spain? Does

What is the predictand? What is the seasonal forecast for Port of Spain? Does the forecast apply to: • Individual locations? • Regions, and if so what size? 7

“Good” forecasts Forecast: Laos will not top the 2012 Olympic medal table. Verification: Who

“Good” forecasts Forecast: Laos will not top the 2012 Olympic medal table. Verification: Who needs to ask? (Laos has never won a medal). Example: Uncertainty Short-term forecasts of the SPI during a drought. Temperature forecasts. 8

Seasonal temperature forecasts The convention is to issue temperature forecasts using categories defined with

Seasonal temperature forecasts The convention is to issue temperature forecasts using categories defined with a 1971 – 2000 climatology. 2009 2008 2007 2006 2005 2004 2003 2002 9

What makes a good forecast? Forecast: Usain Bolt will win gold in the 2012

What makes a good forecast? Forecast: Usain Bolt will win gold in the 2012 Olympics 100 m But: I lied – I thought he would come second. Example: Consistency Preference to forecast high probabilities of normal rainfall? Is the forecast what we really think, or what we are most comfortable with saying? 70% - 80% of all African RCOF forecasts have the highest probability on “normal”. 10

What makes a good forecast? Forecast: Phil Taylor and Adrian Lewis (England) will successfully

What makes a good forecast? Forecast: Phil Taylor and Adrian Lewis (England) will successfully defend the 2013 PDC World Cup of Darts Verification: Who cares? (They did, defeating the Belgians) Example: Relevance Are tercile forecasts really what users want to know? 11

Seasonal rainfall forecasts • Looming drought in Guyana • Will the drought in Georgetown

Seasonal rainfall forecasts • Looming drought in Guyana • Will the drought in Georgetown alleviate? • Does the forecast answer the question? 12

What makes a “good” forecast? Quality Timeliness Uncertainty Relevance No ambiguity Consistency the outcome

What makes a “good” forecast? Quality Timeliness Uncertainty Relevance No ambiguity Consistency the outcome should correspond with the forecast must be issued early enough for response the forecast must be about something that was not inevitable the forecast must be about something of interest (includes timeliness) the precise meaning of the forecast should not be subject to interpretation the forecast should indicate what the forecaster believes will happen

Seasonal forecasts Many seasonal forecasts are not as “good” as they could be because:

Seasonal forecasts Many seasonal forecasts are not as “good” as they could be because: • They use outdated, unavailable and difficult to understand climatological periods • They do not forecast things of much interest • The sharpness is poor • Verification information is very difficult to find (and very difficult to understand if you can find it) • The forecasts often do not accurately reflect what the forecaster really thinks, but rather what the forecaster thinks is safest to say 14

Why verify? • Why verify seasonal forecasts? – Scientific Are the forecasts skillful? How

Why verify? • Why verify seasonal forecasts? – Scientific Are the forecasts skillful? How can the forecasts be improved? – Administrative Is the cost of the forecasts justified? Are the forecasts improving? – Economic Do the forecasts have value? 15

“Good” scores If the forecasts are correct 60% of the time, is that good?

“Good” scores If the forecasts are correct 60% of the time, is that good? 1. Is 60% low? 2. Does the score tell you what you want to know in the first place? More generally: 1. Are the forecasts good? 2. Is the choice of score appropriate? 16

What makes a good forecast? Forecast: There is a 40% probability that West Indies

What makes a good forecast? Forecast: There is a 40% probability that West Indies will win the 2012 ICC Twenty 20 World Cup Final Verification: ? 17

How to verify forecasts? Procedures for verifying forecasts depend on: • The format of

How to verify forecasts? Procedures for verifying forecasts depend on: • The format of the observations; • The format of the forecasts; • How many forecast and observations we have. 18

Seasonal forecast formats Most seasonal forecasts are in one of two classes: 1. A

Seasonal forecast formats Most seasonal forecasts are in one of two classes: 1. A (set of) deterministic forecast(s) – outputs from a dynamical or statistical model. These types of forecasts are might be provided by WMO Global Producing Centres (GPCs). Procedures for the verification of GPC model outputs are defined in the WMO Standardized Verification System. (These procedures are built into CPT. ) 19

Seasonal forecast formats 2. Maps showing probabilities of the verification falling within one of

Seasonal forecast formats 2. Maps showing probabilities of the verification falling within one of multiple categories These types of forecasts are typical of Regional Climate Outlook Forums (RCOFs). There are no formal WMO verification procedures, but some guidance on procedures is being published by WMO CCl. A training workshop for RCCs is planned for October 2013. The training workshop this week covers a subset of these procedures. 20

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be 10 mm of rain tomorrow There is a p% There is a 50% chance of more chance of rain than k mm of tomorrow rain tomorrow 21

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be 10 mm of rain tomorrow There is a p% There is a 50% chance of more chance of rain than k mm of tomorrow rain tomorrow 22

Binary Forecasts There are four possible outcomes: 23

Binary Forecasts There are four possible outcomes: 23

Bias Did the event occur as often as forecast? 24

Bias Did the event occur as often as forecast? 24

Hit Rate How many of the events were forecast? “Wall Street indices predicted nine

Hit Rate How many of the events were forecast? “Wall Street indices predicted nine out of the last five recessions!” Newsweek, 19 Sept. , 1966. False alarms are incorrect forecasts. 25

False Alarm Rate How many of the non-events were forecast? If we issue more

False Alarm Rate How many of the non-events were forecast? If we issue more warnings, the hit rate can be increased, but the false-alarm rate will also be increased – there is a trade-off. We want the hit rate to be as high as possible and the false-alarm rate as low as possible, and, at the very least, the hit rate should be higher than the false-alarm rate. 26

Proportion Correct How many times was the forecast correct? 27

Proportion Correct How many times was the forecast correct? 27

Finley’s Tornado Forecasts A set of tornado forecasts for the U. S. Midwest published

Finley’s Tornado Forecasts A set of tornado forecasts for the U. S. Midwest published in 1884. 28

No Tornado Forecasts A better score can be achieved by issuing no forecasts of

No Tornado Forecasts A better score can be achieved by issuing no forecasts of tornadoes! 29

Heidke Skill Score 30

Heidke Skill Score 30

What is “skill”? Is one set of forecasts better than another? 31

What is “skill”? Is one set of forecasts better than another? 31

Skill Scores • Skill scores compare one set of forecasts with another set of

Skill Scores • Skill scores compare one set of forecasts with another set of forecasts. • A skill score identifies which set of forecasts is best and by how much better it is. • It is quite possible for both sets of forecasts to be very good, or both sets to be very bad. • Skill scores are relative measures of forecast quality – they do not tell us whether forecasts are “good” or “bad” only “better” or “worse”. • Usually, with seasonal forecasts, one of the sets of forecasts is a simple forecasting strategy that is not likely to be very useful. 32

Multiple Categories Near-misses Is the scoring table above a good idea? 33

Multiple Categories Near-misses Is the scoring table above a good idea? 33

Equitable Scoring 34

Equitable Scoring 34

CPT’s Categorical measures provides information on the skill of the deterministic forecasts with the

CPT’s Categorical measures provides information on the skill of the deterministic forecasts with the observations and / or the forecasts converted to categories. 35

Multiple Categories For probabilistic forecasts, the categories are ordered by probability, so if the

Multiple Categories For probabilistic forecasts, the categories are ordered by probability, so if the forecaster thinks the normal category is the most likely, (s)he still has to indicate whether above-normal or belownormal is the next most likely. For example: 36

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be

Types of Forecast Deterministic Probabilistic Discrete Continuous It will rain tomorrow There will be 10 mm of rain tomorrow There is a p% There is a 50% chance of more chance of rain than k mm of tomorrow rain tomorrow 37

CPT’s Continuous measures compare the best-guess forecasts with the observed values without regard to

CPT’s Continuous measures compare the best-guess forecasts with the observed values without regard to the categories. They measure the skill of the deterministic forecasts. 38

What makes a “good” deterministic forecast? Correct the outcome matches the forecast Accurate the

What makes a “good” deterministic forecast? Correct the outcome matches the forecast Accurate the outcome is close to the forecast Unbiased the average outcome matches the average forecast Skill the forecasts are better than a second set of forecasts Association the outcomes and the forecasts increase and decrease together Discrimination the forecasts increase and decrease as the observations increase and decrease 39