Ensemble Verification - PowerPoint PPT Presentation

About This Presentation
Title:

Ensemble Verification

Description:

Adding up two outliers subtract the average. ... Ideal forecasts will ... re-label fcst prob by obs frequency associated with fcst. calibrated. Un-calibrated ... – PowerPoint PPT presentation

Number of Views:24
Avg rating:3.0/5.0
Slides: 24
Provided by: NCEP8
Category:

less

Transcript and Presenter's Notes

Title: Ensemble Verification


1
Ensemble Verification
  • Yuejian Zhu
  • Environmental Modeling Center
  • NOAA/NWS/NCEP
  • Acknowledgements
  • Zoltan Toth EMC

2
Outlines
  • Climatological Data
  • RMS and Spread
  • Mean Error and Absolute Error
  • Histogram and Outlier
  • RPS and RPSS
  • CRPS and CRPSS
  • BSS (Resolution and Reliability)
  • ROC (Hit Rate and False Alarm Rate)
  • Economic Value (cost-loss analysis)

3
Climatological Data
  • NCEP/NCAR 40 years (1958-1997) reanalysis
  • Monthly Sampling
  • For example 40301200
  • 10 equally-a-likely, based on sampling
  • Projected to verify date
  • All forecast skills will base on 10
    equally-a-likely climatological bins.

4
One day advantage
Due to model imperfection ?
5
Winter 0607 NAEFS Statistics
BIAS
6
Prob. Evaluation (simple measurement)
  • 1. Talagrand Distribution (histogram
    distribution)
  • Sorting forecast in order, to check where
    the analysis is falling
  • Reliability measurement, system bias
    detected.
  • positive/negative biased for forecasting
    model,
  • example of these forecasts --gt cold bias,
  • assume analysis is bias-free (perfect).
    Common -"U" sharp

avg distribution
7
(No Transcript)
8
Prob. Evaluation (simple measurement)
  • 1. Talagrand distribution (continue).
  • . Outlier evolution by different leading
    time
  • .. Adding up two outliers subtract the
    average.
  • Ideal forecasts will have zero
    outliers.

Due to inability of ensemble to capture model
related errors?
9
Prob. Evaluation (simple measurement)
  • Outlier --gt diagnostic
  • forecasts .vs. next forecasts ( f24hrs
    valid at same time)
  • assume forecasting model is perfect,
    f24.
  • perfect forecast system will expect the
    outliers are zero.

Detecting model initial uncertainty?
10
Prob. Evaluation (multi-categories)
  • Based on climatological equally likely bins ( for
    example. 5 bins )
  • For verifying multi-category probability
    forecasts.
  • measure both reliability and resolution.
  • 1. Ranked (ordered) probability score ( RPS) and
    RPSS
  • RPSS( RPSf - RPSc )/( 1 - RPSc )

11
What is THORPEXs goal for 10 years ?
12
Continuous Rank Probability Score
Xo
100
Obs (truth)
Heaviside Function H
50
0
X
p07
p09
p08
p06
p03
p02
p01
p04
p05
p10
Order of 10 ensemble members (p01, p02,,p10)
13
CRPS for winter 0607
CRPSS for winter 0607
14
Prob. Evaluation (multi-categories)
  • 2. Brier Score(BS, non-ranked), Brier Skill
    Score(BSS).
  • from two categories to multi-categories/probab
    ilistic
  • ----measure both reliability and resolution

Brier Skill Score
Skill line (ref. is climatology)
15
Prob. Evaluation (multi-categories)
  • 3. Decomposition of Brier Score
  • consider sub-sample and overall-sample
  • reliability, resolution and uncertainty.
  • for reliability 0 is perfectly reliable
  • for resolution 0 is no resolution (
    climatology )
  • when resolution reliability ? no skill
  • example of global ensemble

No skill beyond this point
resolution
reliability
16
Prob. Evaluation (multi-categories)
  • 4. Reliability and possible calibration ( remove
    bias )
  • For period precipitation evaluation

Calibrated forecast
Skill line
Raw forecast
Resolution line Climatological prob.
17
Prob. Evaluation (multi-categories)
  • 4. Reliability and possible probabilistic
    calibration
  • re-label fcst prob by obs frequency
    associated with fcst

calibrated
Un-calibrated
18
Prob. Evaluation (cost-loss analysis)
  • Based on hit rate (HR) and false alarm (FA) rate.
  • 1. Relative Operating Characteristics (ROC) area
    - Appl. of signal detection theory for measuring
    discrimination between two alternative outcome.
  • ROCarea Intergrated area 2 ( 0-1
    normality )

h/(hm)
Relative Operating Characteristics
-------------------------- o\f
y(f) n(f) --------------------------
y(o) h m -------------------------
- n(o) f c -------------------
-------
f/(hf)
19
Relative Operating Characteristics area (ROC area)
f(noise)
f(signal)
Near perfect forecast
1
 
           
 

 
Hit rate
No skill forecast
 
 
 
Real forecast
0
1

False alarm rate
Decision threshold
20
(No Transcript)
21
Prob. Evaluation (cost-loss analysis)
  • 2. Economic Value (EV) of forecasts.
  • Given a particular forecast, a user either
    does or does not take action

Highest value (110)
Ensemble forecast
Value line
Deterministic forecast
22
Prob. Evaluation (cost-loss analysis)
  • Based on hit rate (HR) and false alarm (FA)
    analysis
  • .. Economic Value (EV) of forecasts

Ensemble forecast
Average 2-day advantage
Deterministic forecast
23
Decision Theory Example
Forecast?
YES NO
Critical Event sfc winds gt 50kt Cost (of
protecting) 150K Loss (if damage ) 1M
Hit False Alarm
Miss Correct Rejection
YES NO
150K
1000K
Observed?
150K
0K
Write a Comment
User Comments (0)
About PowerShow.com