Title: IN SEARCH FOR THE BEST FORECAST 2
 1IN SEARCH FOR THE BEST FORECAST - 2
1) ESTIMATING THE FIRST MOMENT OF THE FORECAST 
PDF BASED ON INFORMATION FROM HIGH RESOLUTION 
CONTROL  LOWER RESOLUTION ENSEMBLE 
FORECASTS2) ENSEMBLE BASED ESTIMATES OF 
FORECAST UNCERTAINTY AND THEIR DOWNSCALING TO 
HIGHER RESOLUTION GRIDS 
- Zoltan Toth, Jun Du(1),  Yuejian Zhu 
 - Environmental Modeling Center 
 - NOAA/NWS/NCEP 
 - Ackn. Bo Cui, David Unger, Malaquias Pena 
 - (1)  SAIC at EMC/NCEP/NWS/NOAA 
 - http//wwwt.emc.ncep.noaa.gov/gmb/ens/index.html
 
  2OUTLINE / SUMMARY
- ENSEMBLE BASED ESTIMATES OF FORECAST UNCERTAINTY 
AND THEIR DOWNSCALING TO HIGHER RESOLUTION GRIDS  - ESTIMATING THE FORECAST PDF BASED ON 
 - SINGLE FORECASTS 
 - ENSEMBLES 
 - STATISTICAL POSTPROCESSING 
 - BIAS REDUCTION WRT OPERATIONAL ANALYSIS 
 - DOWNSCALING 
 - FORECAST REPRESENTATIVE OF SMALLER SCALES 
 - FORECAST FORMAT 
 - ALL ENSEMBLE MEMBERS 
 - DERIVED PRODUCTS 
 - NDGD PROPOSAL 
 - REPRESENTING FORECAST UNCERTAINTY 
 - 10, 50, 90 PERCENTILE OF FORECAST PDF 
 
  3INTRODUCTION
- GOAL 
 - Best estimate of future state 
 - Verify by standard probabilistic statistics 
 - APPROACH 
 - Use all available information, including 
 - High resolution control forecast 
 - Lower resolution ensemble forecasts 
 - Climatology of observations/analysis and 
forecasts if available  - FORMAT 
 - PDF for single variables 
 - Ensemble traces to carry temporal/spatial/cross-va
riable covariance info 
  4GENERAL FORECAST APPROACH
- BASED ON A SINGLE FORECAST 
 - In nonlinear regime, sub-optimal performance 
 - Does not provide best estimate for expected value 
 - Does not provide case dependent estimates of 
forecast uncertainty  - In near-linear regime, can offer case dependent 
corrections for  - Lower resolution ensemble 
 - BASED ON AN ENSEMBLE 
 - Better estimate of expected value 
 - Case dependent estimates of forecast uncertainty 
 - DATABASE 
 - Collection of all relevant information 
 - High resolution control forecast 
 - Lower resolution ensemble forecasts 
 - Climatology of observations/analysis and 
forecasts if available  - Short range 
 - SREF ensemble  control forecasts 
 - Medium-  extended range 
 
  5FORECASTING IN A CHAOTIC ENVIRONMENT 
 PROBABILISTIC FORECASTING BASED A ON SINGLE 
FORECAST  One integration with an NWP model, 
combined with past verification statistics 
 DETERMINISTIC APPROACH - PROBABILISTIC FORMAT 
- Does not contain all forecast information 
 - Not best estimate for future evolution of system 
 - UNCERTAINTY CAPTURED IN TIME AVERAGE SENSE - 
 - NO ESTIMATE OF CASE DEPENDENT VARIATIONS IN FCST 
UNCERTAINTY  
  6FORECASTING IN A CHAOTIC ENVIRONMENT - 
3DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- MONTE CARLO APPROACH  ENSEMBLE FORECASTING 
 -  IDEA Sample sources of forecast error 
 - Generate initial ensemble perturbations 
 - Represent model related uncertainty 
 -  PRACTICE Run multiple NWP model integrations 
 - Advantage of perfect parallelization 
 - Use lower spatial resolution if short on 
resources  -  USAGE Construct forecast pdf based on finite 
sample  - Ready to be used in real world applications 
 - Verification of forecasts 
 - Statistical post-processing (remove bias in 1st, 
2nd, higher moments)  - CAPTURES FLOW DEPENDENT VARIATIONS 
 -  IN FORECAST UNCERTAINTY 
 
  7CONFIGURATION, OUTPUT CHARACTERISTICS
2005, 2006, 2007, 2008 
 8PROCESS
- Collection of relevant information (model grid) 
 - SREF 
 - NAEFS 
 - Statistical post-processing (model grid) 
 - Bias correction 
 - Statistical 
 - Case dependent (using hires control) 
 - Weights 
 - Verify added value 
 - Manual modification of forecast guidance (model 
grid)  - Verify added value 
 - Downscaling (NDGD grid) 
 - Observation locations 
 - NDGD grid 
 - Manual modification (NDGD grid) 
 - Verify added value
 
  9RAW  BASIC PRODUCT AVAILABILITY
2005, 2006, 2007, 2008 
 10STATISTICAL POST-PROCESSING
- PURPOSE 
 - Make all forecasts look like model analysis 
 - Independent of lead time 
 - Selected set of often used variables 
 - All ensemble members and hi-lower resolution 
controls  - Assign weights to each member, corresponding to 
its performance  - DATA 
 - Medium-  extended range application 
 - 1x1 global grid, 50 variables 
 - NAEFS ensemble  control forecasts 
 - Operational analyses 
 - TWO METHODS 
 - Frequentists approach 
 - Compare recent statistics of forecasts and 
analyses  - In collaboration with David Unger of CPC 
 - Bayesian approach 
 - In collaboration with Roman Krzysztofowicz
 
  11POST-PROCESSING METHODS - 1
- FREQUENTISTS APPROACH 
 - First moment (expected value) June 2006 
 - Estimate bias in first moment 
 - 35 selected variables 
 - Compare weighted mean of recent forecasts with 
that of verifying analysis  - Kalman-filter type adaptive technique works well 
for short lead time  - Remove estimated bias from each forecast member 
 - Improved 1st moment Oct 07 
 - More efficient for longer lead times 
 - Second moment bias correction Oct 07 
 - Weights Oct 07 
 - Consider case dependent reduction of bias in 
lowres ensemble  - Jun Dus hybrid approach 
 - Compare high  lower resolution (ensemble) 
control forecasts  - Difference interpreted as resolution vector 
 - Adjust all members by resolution vector 
 - Potential for short lead times, linear regime 
 
  12(No Transcript) 
 13POST-PROCESSING METHODS - 2
- BAYESIAN APPROACH 
 - Based on single forecast PQPF processor of 
Krzysztofowicz et al  - Modified for ensemble forecast application 
 - Interpret/correct latest ensemble forecast in 
context of prior information  - Apply procedure on each variable to be corrected 
separately  - Use NCEP/NCAR reanalysis climatology as prior 
 - Good estimate, based on 50 years, 2.5x2.5 grid 
 - Adjust reanalysis climate (2.5x2.5) to estimate 
operational analysis climate (1x1)  - See downscaling vector below 
 - Represent climate pdf with parametric 
distribution (2-3 pars)  - Compare forecast distributions with observations, 
conditioned on fcst pdf?  - Represent forecast ensemble distribution with 
parametric distribution (2-3 vars)  - Use shorter time period (90 days) with current 
model/ensemble configuration  - Use Kalman-filter type approach, update prior 
with most recent comparison  - Transform statistics to multinormal space 
 - Form posterior pdf, based on prior and fcst-obs 
comparison  - Estimate approximate weights for each member 
 - Based on Bayesian statistics combined for a set 
of selected variables  - Retain temporal/spatial/cross-variable rank 
correlations present in ensemble  
  14DOWNSCALING METHODS
- PURPOSE 
 - Provide bias-corrected forecasts at small scales 
of interest  - Observation points 
 - High resolution grid 
 - National Digital Forecast (or Guidance) Database 
(NDFD, NDGD)  - DATA 
 - Input 
 - Bias corrected ensemble forecasts 
 - Output 
 - Observation locations 
 - NDGD grid 
 - TWO METHODS 
 - Climate anomalies  June 2006 
 - Downscaling vector - Oct 2007 
 
  15DOWNSCALING  CLIMATE ANOMALIES
- Apply procedure on each ensemble member 
 - 19 selected bias-corrected variables 
 - Adjust bias-corrected forecast to look like 
reanalysis  - Use standard Kalman-filter type bias-correction 
algorithm  - Evaluate systematic difference between 
 - CDAS analysis (2.5x2.5 grid) and 
 - Operational analyses (1x1 grid) 
 - Remove systematic difference from bias-corrected 
forecast (2.5x2.5)  - Compare adjusted bias-corrected forecast to 
reanalysis climate pdf  - Represent climate distribution by parametric pdf 
(2-3 vars)  - Determine climate percentile corresponding to 
forecast value 
  16DOWNSCALING WITH DOWNSCALING VECTOR
- Determine downscaling vector 
 - Systematic difference between 
 - Operational analysis (1x1 grid) 
 - Interpolate to NDGD grid using standard routine 
 - Real Time Meso-scale Analysis (RTMA) 
 - Based on RUC, not cycled analysis 
 - Quality measured against independent data 
 - 4-5 variables (2m temp, dewpoint, 10m winds, 
precip)  - 5x5km NDGD grid 
 - Use standard Kalman-filter type bias-estimation  
correction algorithm  - Independent of lead time 
 - Ship downscaling vector to users 
 - 5x5 km grid  send only for area of interest, 1 
field per variable /day  - Ship bias-corrected forecast information 
 - 1x1 lat/lon grid, all lead times of interest 
 - Interpolate on-site to 5x5km grid, using same 
standard routine as above  - Add downscaling vector to interpolated 
bias-corrected forecast  - Performance measure 
 
  17FORECAST FORMAT
- Different users have different requirements 
 - Multiple formats must be offered 
 - Full format 
 - All bias-corrected ensemble members 
 - In-house, NCEP service centers, interested ftp 
users  - All other formats 
 - Can be derived from full format (all members) 
 - Ensemble functionalities 
 - Different products can be generated and 
distributed as needed  - NDGD proposal 
 - Important special application 
 - In addition to most likely scenario 
 - Provide information on forecast uncertainty
 
  18ENSEMBLE PRODUCTS - FUNCTIONALITIES
List of centrally/locally/interactively generated 
products required by NCEP Service Centers for 
each functionality are provided in attached 
tables (eg., MSLP, Z,T,U,V,RH, etc, at 
925,850,700,500, 400, 300, 250, 100, etc hPa)
Potentially useful functionalities that need 
further development - Mean/Spread/Median/Ranges 
for amplitude of specific features - 
Mean/Spread/Median/Ranges for phase of specific 
features 
Additional basic GUI functionalities - Ability 
to manually select/identify members - Ability to 
weight selected members Sept. 2005 
 19ENSEMBLE PRODUCT REQUEST LIST NCEP SERVICE 
CENTERS, OTHER PROJECTS 
 20NDGD PROPOSAL
- NDFD  current status 
 - 5x5km (or 2.5x2.5km) grid, 15-20 variables, out 
to 7 days  - Most likely value (expected value of forecast 
pdf) provided only  - Format of official NWS forecast 
 - NDGD  new vehicle to provide guidance to be used 
in NDFD process  - Same format as NDFD 
 - Proposal to add forecast uncertainty information 
in NDGD  - Forecast values corresponding to 10, 50,  90 
percentile of forecast pdf  - Bias corrected forecast data, 1x1 grid June 06 
 - Downscaling vector for 1-2 variables, 5x5km 
grid Oct 07  - Quality to be assessed 
 - Can be expanded to include more NDFD variables
 
  21MANUAL MANIPULATION OF FORECASTS
- Manipulation needed for 
 - Bias-corrected forecast on model grid (national 
guidance)  - Downscaled forecast on NDGD grid (local guidance) 
 - Suggestion 
 - Initially modify only forecast value 
corresponding to 50 percentile of fcst pdf  - Use standard IFPS, NAWIPS tools 
 - Verify added value 
 - Move whole forecast distribution (10  90 
percentiles) with same adjustment  - Later attempt to modify forecast values 
corresponding to 10  90 perc.  - Verify added value
 
  22BACKGROUND 
 23PRACTICAL APPROACH
- BIAS CORRECTION 
 - Adjust both hires control  lores ensemble 
 - Ensures maximum utility of all forecasts 
 - Limits hires advantage to 
 - Flow dependent bias correction that statistical 
method cannot achieve  - IDENTIFY CROSS-OVER POINT 
 - Compare measures normalized by ens spread 
 - Measure of nonlinearity 
 - Control minus ens mean over ens spread 
 - Measure of hires advantage 
 - Error reduction in hires vs. lores controls over 
ens spread  - Alternatively, assess ens mean error dropping 
below hires control error  - USE OF HIRES CONTROL 
 - Before turnover point 
 - Quantitatively 
 - Develop methods to objectively combine info from 
hires control  lores ens  - Bayesian approach? 
 
  24INTRODUCTION
- GOAL 
 - Best estimate of future state 
 - Verify by standard probabilistic statistics 
 - FORMAT 
 - PDF for single variables 
 - Ensemble traces to carry temporal/spatial/cross-va
riable covariance info  - DATABASE  
 - Collection of all relevant information 
 - High resolution control forecast 
 - Lower resolution ensemble forecasts 
 - Climatology of observations/analysis and 
forecasts if available  - Short range 
 - SREF ensemble  control forecasts 
 - Medium-  extended range 
 - North American Ensemble Forecast System (NAEFS) 
 - Current Canadian  NCEP global ensembles (80 
members/day)  - Planned FNMOC, JMA added by FY08 (160 
members/day)  
  25OUTLINE / SUMMARY
- ENSEMBLE BASED ESTIMATES OF FORECAST UNCERTAINTY 
AND THEIR DOWNSCALING TO HIGHER RESOLUTION GRIDS  - UNCERTAINTY RELATED TO THE MOST LIKELY SOLUTION 
 - LINK WITH TRADITIONAL FORECAST APPROACH 
 - STATISTICAL METHODS OF ASSESSING FORECAST 
UNCERTAINTY  - CHOICE OF THE PAST 
 - NOT CASE DEPENDENT  NEED ANOTHER APPROACH 
 - MUST USE (POSSIBLY LOWER RESOLUTION) ENSEMBLE 
 - CAPTURE UNCERTAINTY ON RESOLVED SCALES 
 - COMMUNICATING FORECAST UNCERTAINTY 
 - MULTI-TIER APPROACH DEPENDING ON LEVEL OF USER 
SOPHISTICATION  - 10, 50, 90 PERCENTILE FORECAST VALUES FOR GENERAL 
USERS  - MORE DETAILED INFO FOR SOPHISTICATED USERS 
 - MULTIPLE SCENARIOS (MODES) 
 - FULL PDF 
 
  26STATISTICAL ESTIMATION OF FORECAST UNCERTAINTY
- BIAS CORRECTION 
 - Adjust both hires control  lores ensemble 
 - Ensures maximum utility of all forecasts 
 - Limits hires advantage to 
 - Flow dependent bias correction that statistical 
method cannot achieve  - IDENTIFY CROSS-OVER POINT 
 - Compare measures normalized by ens spread 
 - Measure of nonlinearity 
 - Control minus ens mean over ens spread 
 - Measure of hires advantage 
 - Error reduction in hires vs. lores controls over 
ens spread  - Alternatively, assess ens mean error dropping 
below hires control error  - USE OF HIRES CONTROL 
 - Before turnover point 
 - Quantitatively 
 - Develop methods to objectively combine info from 
hires control  lores ens  - Bayesian approach? 
 
  27IN SEARCH FOR THE BEST FORECAST - 1
1) ESTIMATING THE FIRST MOMENT OF THE FORECAST 
PDF BASED ON INFORMATION FROM HIGH RESOLUTION 
CONTROL  LOWER RESOLUTION ENSEMBLE 
FORECASTS2) ENSEMBLE BASED ESTIMATES OF 
FORECAST UNCERTAINTY AND THEIR DOWNSCALING TO 
HIGHER RESOLUTION GRIDS 
- Zoltan Toth, Jun Du(1),  Yuejian Zhu 
 - Environmental Modeling Center 
 - NOAA/NWS/NCEP 
 - Ackn. Bo Cui, David Unger, Malaquias Pena 
 - (1)  SAIC at EMC/NCEP/NWS/NOAA 
 - http//wwwt.emc.ncep.noaa.gov/gmb/ens/index.html
 
  28OUTLINE / SUMMARY
- ESTIMATING THE FIRST MOMENT OF THE FORECAST PDF 
BASED ON INFORMATION FROM HIGH RESOLUTION CONTROL 
 LOWER RESOLUTION ENSEMBLE FORECASTS  - RECONCILING INFO FROM HIRES CONTROL VS. LORES ENS 
 - ADVISE USER COMMUNITY 
 - REDESIGN NWP SUITE IF NEEDED 
 - ATTRIBUTES OF FORECAST SYSTEMS 
 - POSITIVE ATTRIBUTES OF HIRES CONTROL VS. LORES 
ENSEMBLE  - FIRST MOMENT ESTIMATION (MOST LIKELY STATE) 
 - WHAT LIMITES USE OF HIRES CONTROL 
 - LEVEL OF DIFFERENCE IN QUALITY OF HIRES VS. LORES 
CONTROLS  - EMERGENCE OF NONLINEARITIES 
 - PRACTICAL APPROACH 
 - BIAS CORRECT ALL FORECASTS 
 - USE OF HIRES CONTROL 
 - LINEAR PHASE 
 - Objectively combine hires control  lores ens 
 
  29INTRODUCTION
- PROBLEM 
 - Given high resolution single control and lower 
resolution ensemble forecasts  - What is best way to construct forecast 
probability density function (pdf)?  - Estimate first moment of distribution 
 - Estimate second and higher moments 
 - How to modify operational suite of NWP forecasts 
for best performance?  - Future choices to be made based on evaluation 
results  - ATTRIBUTES OF FORECAST SYSTEMS 
 - RELIABILITY 
 - Statistically, forecasts look like reality, 
irrespective of how skilful they are  - With large enough sample of forecast-observation 
pairs, can be improved  - RESOLUTION 
 - Sequence of observed events captured, 
irrespective of realism of forecasts  - Cannot be improved by statistical methods 
 - USER REQUIREMENTS 
 - Both forecast attributes important 
 
  30HIGHRES CONTROL VS. LOWRES ENSEMBLE
- RELIABILITY 
 - Highres expected to be better, especially at low 
levels affected by terrain etc  - Reduced bias 
 - Some phenomena simulated with more fidelity (eg, 
diurnal cycle, frontal structure, etc)  - Quantify advantage of highres model before and 
after bias correction  - Bias correction benefits more the forecasts with 
larger bias gt  - Effect of bias correction expected to reduce 
advantage of highres forecast  - RESOLUTION 
 - At short lead time 
 - Highres may offer advantage? 
 - Never tested, needs to be evaluated 
 - At later lead time 
 - Ensemble may become advantageous 
 - Filtering of nonlinearly growing errors 
 - HOW TO RECONCILE FORECAST INFORMATION 
 - FROM HIRES CONTROL  LOWRES ENS?
 
  31RECONCILING HIGHRES CONTROL VS. LOWRES ENS
- WMO/CBS Ensemble Expert Team meeting (Feb. 2006) 
 - Guidance sought by WMO members as to relevance of 
hires vs. lowres ensemble forecasts for  - First moment estimation 
 - Forecast uncertainty estimation 
 - Some related studies from various viewpoints in 
literature  - Hires control vs. ens mean (Zhu, Atger, Du, etc) 
 - Probabilistic forecasts based on hires control 
vs. lores ens  - Zhu et al, Atger, Du 
 - Combined use of hires control  lores ensemble 
 - Smith et al., J. Du 
 - Following careful evaluation of strength  
weaknesses of both systems  - Advise user community on best practices 
 - Develop new tools if necessary to facilitate best 
use  - Redesign NWP forecast suite if necessary
 
  32ESTIMATING FIRST MOMENT
- EMERGENCE OF NONLINEARITIES 
 - Initially symmetric cloud of ensemble becomes 
distorted with lead time  - Measure level of nonlinearity by deviation of 
lores control and ens mean  - Normalized measure Control minus ens mean over 
ens spread  - What is critical level? 20 or what? (0linear, 
1nonlinear)  - Linearity is lost sooner for smaller scale / more 
unstable phenomena  - Once linearity is critically violated, relying on 
single control (even hires) not appropriate  - All ensemble solutions must be considered 
 - BEFORE NONLINEARITY BECOMES DOMINANT 
 - Hires control can have value beyond lores 
ensemble  - Value beyond ensemble is limited 
 - Resolution is typically only twice that of 
ensemble  - Bias correction of all forecasts further reduces 
value of hires control  - Quantify advantage of hires control statistically 
 - Error reduction (hires vs. lores controls) as 
function of lead time  - Normalize by ensemble spread? 
 - After cross-over time, only very special, 
qualitative use of hires model  
  33PRACTICAL APPROACH
- BIAS CORRECTION 
 - Adjust both hires control  lores ensemble 
 - Ensures maximum utility of all forecasts 
 - Limits hires advantage to 
 - Flow dependent bias correction that statistical 
method cannot achieve  - IDENTIFY CROSS-OVER POINT 
 - Compare measures normalized by ens spread 
 - Measure of nonlinearity 
 - Control minus ens mean over ens spread 
 - Measure of hires advantage 
 - Error reduction in hires vs. lores controls over 
ens spread  - Alternatively, assess ens mean error dropping 
below hires control error  - USE OF HIRES CONTROL 
 - Before turnover point 
 - Quantitatively 
 - Develop methods to objectively combine info from 
hires control  lores ens  - Bayesian approach? 
 
  34SCIENTIFIC BACKGROUND WEATHER FORECASTS ARE 
UNCERTAIN
Buizza 2002 
 35(No Transcript) 
 36BACKGROUND 
 37COMMENTS ON JUNS APPROACH
- Bias correction / post-processing algorithm based 
on  - Difference between hires and lores controls 
 - Useful as long as 
 - Controls evolve similarly gt 
 - Difference dominated by systematic differences 
between 2 model resolutions  - Once 2 controls diverge chaotically, random 
errors introduced  - How to quantify? 
 - Measure difference between 2 controls, normalized 
by ens spread?  - Related to limit of linearity discussed earlier? 
 - Stop using method beyond a point (use correction 
with tapering weights?)  - Characteristics 
 - Positive 
 - Completely flow dependent corrections 
 - Negative 
 - Correction is limited by resolution/performance 
of hires model gt  - Statistical bias correction still needed 
 
  38EVALUATION OF FORECAST SYSTEMS
- Some statistics based on forecast system only 
 - Other statistics based on comparison of forecast 
and observed systems gt  -  FORECAST SYSTEM ATTRIBUTES 
 - Abstract concepts (like length) 
 - Reliability and Resolution 
 - Both can be measured through different statistics 
 - Statistical properties 
 - Interpreted for large set of forecasts (ie, 
describe behavior of forecast system),  -  not for a single forecast 
 - For their definition 
 - Assume that forecasts 
 - Can be of any format 
 - Take a finite number of different classes 
 - Consider empirical frequency distribution of 
 - Verifying observations corresponding to large 
number of forecasts of same class gt  -  Observed Frequency Distribution (ofd)
 
  39STATISTICAL RELIABILITY STATISTICAL CONSISTENCY 
OF FORECASTS WITH OBSERVATIONS 
- BACKGROUND 
 - Consider particular forecast class  Fa 
 - Consider distribution of observations Oa that 
follow forecasts Fa  - DEFINITION 
 - If forecast Fa has the exact same form as Oa, for 
all forecast classes,  -  the forecast system is statistically consistent 
with observations gt  -  The forecast system is perfectly reliable 
 - MEASURES OF RELIABILITY 
 - Based on different ways of comparing Fa and Oa 
 
  40STATISTICAL RESOLUTION ABILITY TO DISTINGUISH, 
AHEAD OF TIME, AMONG DIFFERENT OUTCOMES
- BACKGROUND 
 - Assume observed events are classified into finite 
number of classes  - DEFINITION 
 - If all observed classes are preceded by 
distinctly different forecasts, the forecasts 
resolve the problem gt  -  The forecast system has perfect resolution 
 - MEASURES OF RELIABILITY 
 - Based on degree of separation of distributions of 
observations that follow various forecast classes 
  - Measured by difference between ofds  climate 
distribution  - Measures differ by how differences between 
distributions are quantified 
FORECASTS
OBSERVATIONS
EXAMPLES 
 41CHARACTERISTICS OF FORECAST SYSTEM ATTRIBUTES
- Reliability  resolution are general forecast 
attributes  - Valid for any forecast format (single, 
categorical, probabilistic, etc)  - Reliability 
 - Can be statistically imposed at one time level 
 - If both natural  forecast systems are stationary 
in time, and  - If there is a large enough set of 
observed-forecast pairs  - Replace forecast by corresponding observed 
frequency distribution  - Not related to time sequence of forecast/observed 
systems  - Resolution reflects inherent value of forecast 
system  - Can be improved only through more knowledge about 
time sequence of events  - Statistical consistency at one time level 
(reliability) is irrelevant  - Reliability  resolution are independent 
attributes  - Climate pdf forecast is perfectly reliable, yet 
has no resolution  - Reversed rain / no-rain forecast can have perfect 
resolution and no reliability  - Perfect reliability and perfect resolution  
perfect forecast system  
  42FORECAST METHODS
- Empirically based 
 - Based on record of observations gt 
 - Possibly very good reliability 
 - Will fail in new (not yet observed) situations 
(eg., climate trend, etc)  - Resolution (forecast skill) depends on length of 
observations  - Useful for now-casting, climate applications 
 - Not practical for typical weather forecasting 
 - Theoretically based 
 - Based on general scientific principles 
 - Incomplete/approximate knowledge in NWP models gt 
 - Prone to statistical inconsistency 
 - Run-of-the-mill cases can be statistically 
calibrated to insure reliability  - For forecasting rare/extreme events, statistical 
consistency of model must be improved  - Predictability limited by 
 - Gaps in knowledge about system 
 - Errors in initial state of system
 
  43USER REQUIREMENTSPROBABILISTIC FORECAST 
INFORMATION IS CRITICAL
  44FORECASTING IN A CHAOTIC ENVIRONMENT 
 PROBABILISTIC FORECASTING BASED A ON SINGLE 
FORECAST  One integration with an NWP model, 
combined with past verification statistics 
 DETERMINISTIC APPROACH - PROBABILISTIC FORMAT 
- Does not contain all forecast information 
 - Not best estimate for future evolution of system 
 - UNCERTAINTY CAPTURED IN TIME AVERAGE SENSE - 
 - NO ESTIMATE OF CASE DEPENDENT VARIATIONS IN FCST 
UNCERTAINTY  
  45- FORECASTING IN A CHAOTIC ENVIRONMENT - 2 
 - DETERMINISTIC APPROACH - PROBABILISTIC FORMAT 
 -  
 - PROBABILISTIC FORECASTING - 
 - Based on Liuville Equations 
 -  Continuity equation for probabilities, given 
dynamical eqs. of motion  -  Initialize with probability distribution 
function (pdf) at analysis time  -  Dynamical forecast of pdf based on conservation 
of probability values  -  Prohibitively expensive - 
 -  Very high dimensional problem (state space x 
probability space)  -  Separate integration for each lead time 
 -  Closure problems when simplified solution sought
 
  46FORECASTING IN A CHAOTIC ENVIRONMENT - 
3DETERMINISTIC APPROACH - PROBABILISTIC FORMAT
- MONTE CARLO APPROACH  ENSEMBLE FORECASTING 
 -  IDEA Sample sources of forecast error 
 - Generate initial ensemble perturbations 
 - Represent model related uncertainty 
 -  PRACTICE Run multiple NWP model integrations 
 - Advantage of perfect parallelization 
 - Use lower spatial resolution if short on 
resources  -  USAGE Construct forecast pdf based on finite 
sample  - Ready to be used in real world applications 
 - Verification of forecasts 
 - Statistical post-processing (remove bias in 1st, 
2nd, higher moments)  - CAPTURES FLOW DEPENDENT VARIATIONS 
 -  IN FORECAST UNCERTAINTY 
 
  47NCEP GLOBAL ENSEMBLE FORECAST SYSTEM
MARCH 2004 CONFIGURATION 
 48MOTIVATION FOR ENSEMBLE FORECASTING
- FORECASTS ARE NOT PERFECT - IMPLICATIONS FOR 
 - USERS 
 - Need to know how often / by how much forecasts 
fail  - Economically optimal behavior depends on 
 - Forecast error characteristics 
 - User specific application 
 - Cost of weather related adaptive action 
 - Expected loss if no action taken 
 - EXAMPLE Protect or not your crop against 
possible frost  - Cost  10k, Potential Loss  100k gt Will protect 
if P(frost) gt Cost/Loss0.1  - NEED FOR PROBABILISTIC FORECAST INFORMATION 
 - DEVELOPERS 
 - Need to improve performance - Reduce error in 
estimate of first moment  - Traditional NWP activities (I.e., model, data 
assimilation development)  - Need to account for uncertainty - Estimate higher 
moments  - New aspect  How to do this? 
 -  Forecast is incomplete without information on 
forecast uncertainty  - NEED TO USE PROBABILISTIC FORECAST FORMAT 
 
  49HOW TO DEAL WITH FORECAST UNCERTAINTY?
- No matter what / how sophisticated forecast 
methods we use  - Forecast skill limited 
 - Skill varies from case to case 
 - Forecast uncertainty must be assessed by 
meteorologists 
THE PROBABILISTIC APPROACH 
 50SOCIO-ECONOMIC BENEFITS OFSEAMLESS 
WEATHER/CLIMATE FORECAST SUITE
Commerce Energy
Ecosystem Health
Hydropower Agriculture
Boundary Condition Sensitivity
Reservoir control Recreation
Transportation Fire weather
Initial Condition Sensitivity
Flood mitigation Navigation
Protection of Life/Property
Weeks
Minutes
Days
Hours
Years
Seasons
Months 
 51(No Transcript) 
 52(No Transcript) 
 53144 hr forecast
Poorly predictable large scale wave Eastern 
Pacific  Western US 
Highly predictable small scale wave Eastern US
Verification 
 54(No Transcript) 
 55(No Transcript) 
 56(No Transcript) 
 57FORECAST PERFORMANCE MEASURES
COMMON CHARACTERISTIC Function of both forecast 
and observed values
MEASURES OF RELIABILITY DESCRIPTION Statisticall
y compares any sample of forecasts with sample of 
corresponding observations GOAL To assess 
similarity of samples (e.g., whether 1st and 2nd 
moments match) EXAMPLES Reliability component 
of Brier Score Ranked Probability 
Score Analysis Rank Histogram Spread vs. Ens. 
Mean error Etc. 
MEASURES OF RESOLUTION DESCRIPTION Compares the 
distribution of observations that follows 
different classes of forecasts with the climate 
distribution GOAL To assess how well the 
observations are separated when grouped by 
different classes of preceding fcsts EXAMPLES Res
olution component of Brier Score Ranked 
Probability Score Information content Relative 
Operational Characteristics Relative Economic 
Value Etc.
COMBINED (RELRES) MEASURES Brier, Ranked 
Probab. Scores, rmse, PAC, etc 
 58EXAMPLE  PROBABILISTIC FORECASTS 
 RELIABILITY Forecast probabilities for given 
event match observed frequencies of that event 
(with given prob. fcst) RESOLUTION Many 
forecasts fall into classes corresponding to high 
or low observed frequency of given 
event (Occurrence and non-occurrence of event is 
well resolved by fcst system) 
 59(No Transcript) 
 60PROBABILISTIC FORECAST PERFORMANCE MEASURES
TO ASSESS TWO MAIN ATTRIBUTES OF PROBABILISTIC 
FORECASTS RELIABILITY AND RESOLUTION Univariate 
measures Statistics accumulated point by 
point in space Multivariate measures Spatial 
covariance is considered
BRIER SKILL SCORE (BSS)
EXAMPLE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION 
 61BRIER SKILL SCORE (BSS)
COMBINED MEASURE OF RELIABILITY AND RESOLUTION
-  METHOD 
 - Compares pdf against analysis 
 - Resolution (random error) 
 - Reliability (systematic error) 
 -  EVALUATION 
 - BSS Higher better 
 - Resolution Higher better 
 - Reliability Lower better 
 -  RESULTS 
 - Resolution dominates initially 
 - Reliability becomes important later 
 - ECMWF best throughout 
 - Good analysis/model? 
 - NCEP good days 1-2 
 - Good initial perturbations? 
 - No model perturb. hurts later? 
 - CANADIAN good days 8-10 
 
May-June-July 2002 average Brier skill score for 
the EC-EPS (grey lines with full circles), the 
MSC-EPS (black lines with open circles) and the 
NCEP-EPS (black lines with crosses). Bottom 
resolution (dotted) and reliability(solid) 
contributions to the Brier skill score. Values 
refer to the 500 hPa geopotential height over the 
northern hemisphere latitudinal band 20º-80ºN, 
and have been computed considering 10 
equally-climatologically-likely intervals (from 
Buizza, Houtekamer, Toth et al, 2004) 
 62BRIER SKILL SCORE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION 
 63RANKED PROBABILITY SCORE
COMBINED MEASURE OF RELIABILITY AND RESOLUTION 
 64ANALYSIS RANK HISTOGRAM (TALAGRAND DIAGRAM)
MEASURE OF RELIABILITY 
 65ENSEMBLE MEAN ERROR VS. ENSEMBLE SPREAD
MEASURE OF RELIABILITY
Statistical consistency between the ensemble and 
the verifying analysis means that the verifying 
analysis should be statistically 
indistinguishable from the ensemble members 
gt Ensemble mean error (distance between ens. 
mean and analysis) should be equal to ensemble 
spread (distance between ensemble mean and 
ensemble members)
In case of a statistically consistent ensemble, 
ens. spread  ens. mean error, and they are both 
a MEASURE OF RESOLUTION. In the presence of bias, 
both rms error and PAC will be a combined measure 
of reliability and resolution 
 66INFORMATION CONTENT
MEASURE OF RESOLUTION 
 67RELATIVE OPERATING CHARACTERISTICS
MEASURE OF RESOLUTION 
 68ECONOMIC VALUE OF FORECASTS
MEASURE OF RESOLUTION 
 69PERTURBATION VS. ERROR CORRELATION ANALYSIS (PECA)
MULTIVATIATE COMBINED MEASURE OF RELIABILITY  
RESOLUTION
- METHOD Compute correlation between ens 
perturbtns and error in control fcst for  - Individual members 
 - Optimal combination of members 
 - Each ensemble 
 - Various areas, all lead time 
 - EVALUATION Large correlation indicates ens 
captures error in control forecast  - Caveat  errors defined by analysis 
 - RESULTS 
 - Canadian best on large scales 
 - Benefit of model diversity? 
 - ECMWF gains most from combinations 
 - Benefit of orthogonalization? 
 - NCEP best on small scale, short term 
 - Benefit of breeding (best estimate initial 
error)?  - PECA increases with lead time 
 - Lyapunov convergence 
 - Nonlilnear saturation 
 - Higher values on small scales
 
  70WHAT WE NEED FOR POSTPROCESSING TO WORK?
- LARGE SET OF FCST  OBS PAIRS 
 - Consistency defined over large sample  need same 
for post-processing  - Larger the sample, more detailed corrections can 
be made  - BOTH FCST AND REAL SYSTEMS MUST BE STATIONARY IN 
TIME  - Otherwise can make things worse 
 - Subjective forecasts difficult to calibrate
 
HOW WE MEASURE STATISTICAL INCONSISTENCY?
- MEASURES OF STATIST. RELIABILITY 
 - Time mean error 
 - Analysis rank histogram (Talagrand diagram) 
 - Reliability component of Brier etc scores 
 - Reliability diagram 
 
  71SOURCES OF STATISTICAL INCONSISTENCY
- TOO FEW FORECAST MEMBERS 
 - Single forecast  inconsistent by definition, 
unless perfect  - MOS fcst hedged toward climatology as fcst skill 
is lost  - Small ensemble  sampling error due to limited 
ensemble size  -  (Houtekamer 1994?) 
 - MODEL ERROR (BIAS) 
 - Deficiencies due to various problems in NWP 
models  - Effect is exacerbated with increasing lead time 
 - SYSTEMATIC ERRORS (BIAS) IN ANALYSIS 
 - Induced by observations 
 - Effect dies out with increasing lead time 
 - Model related 
 - Bias manifests itself even in initial conditions 
 - ENSEMBLE FORMATION (INPROPER SPREAD) 
 - Not appropriate initial spread 
 - Lack of representation of model related 
uncertainty in ensemble  - I. E., use of simplified model that is not able 
to account for model related uncertainty 
  72HOW TO IMPROVE STATISTICAL CONSISTENCY? 
- MITIGATE SOURCES OF INCONSISTENCY 
 - TOO FEW MEMBERS 
 - Run large ensemble 
 - MODEL ERRORS 
 - Make models more realistic 
 - INSUFFICIENT ENSEMBLE SPREAD 
 - Enhance models so they can represent model 
related forecast uncertainty  - OTHERWISE gt 
 - STATISTICALLY ADJUST FCST TO REDUCE INCONSISTENCY 
 - Unpreferred way of doing it 
 - What we learn can feed back into development to 
mitigate problem at sources  - Can have LARGE impact on (inexperienced) users
 
  73(No Transcript) 
 74SUMMARY
- WHY DO WE NEED PROBABILISTIC FORECASTS? 
 - Isnt the atmosphere deterministic? YES, but 
its also CHAOTIC  - FORECASTERS PERSPECTIVE USERS PERSPECTIVE 
 - Ensemble techniques Probabilistic description 
 - WHAT ARE THE MAIN ATTRIBUTES OF FORECAST SYSTEMS? 
 - RELIABILITY Stat. consistency with distribution 
of corresponding observations  - RESOLUTION Different events are preceded by 
different forecasts  - WHAT ARE THE MAIN TYPES OF FORECAST METHODS? 
 - EMPIRICAL Good reliability, limited resolution 
(problems in new situations)  - THEORETICAL Potentially high resolution, prone to 
inconsistency  - ENSEMBLE METHODS 
 - Only practical way of capturing fluctuations in 
forecast uncertainty due to  - Case dependent dynamics acting on errors in 
 - Initial conditions 
 - Forecast methods 
 - HOW CAN PROBABILSTIC FORECAST PERFORMANCE BE 
MEASURED?  
  75Toth, Z., O. Talagrand, and Y. Zhu, 2005 The 
Attributes of Forecast Systems A Framework for 
the Evaluation and Calibration of Weather 
Forecasts. In Predictability Seminars, 9-13 
September 2002, Ed. T. Palmer, ECMWF, in press. 
 Toth, Z., O. Talagrand, G. Candille, and Y. Zhu, 
2003 Probability and ensemble forecasts. In 
Environmental Forecast Verification A 
practitioner's guide in atmospheric science. Ed. 
I. T. Jolliffe and D. B. Stephenson. Wiley, p. 
137-164.