250 likes | 261 Views
This article provides an outline and summary of the Winter Storm Reconnaissance Program (WSR) implemented by the US National Weather Service. It covers the history, contributions, and operational procedures of the program, including case selection, sensitivity calculations, and observed data impact. The article also highlights open questions and explores the potential economic benefits and limitations of current techniques.
E N D
Zoltan Toth GSD/ESRL/OAR/NOAA Formerly at EMC/NCEP/NWS/NOAA Acknowledgements: Yucheng Song – Plurality at EMC Sharan Majumdar – U. Miami Istvan Szunyogh – Texas AMU Craig Bishop - NRL Rolf Langland - NRL THORPEX Symposium, Sept 14-18 2009, Monterey, CA THE WINTER STORM RECONNESSAINCE PROGRAM OF THEUS NATIONAL WEATHER SERVICE
OUTLINE / SUMMARY • History • Outgrowth of FASTEX & NORPEX research • Operationally implemented at NWS in 2001 • Contributions / documentation • Community effort • Refereed and other publications, rich info on web • Highlights • Operational procedures for case selection, ETKF sensitivity calculations • Positive results consistent from year to year • Open questions • Does operational targeting have economic benefits? • Can similar or better results be achieved with cheaper obs. systems? • What are the limitations of current techniques?
HISTORY OF WSR • Sensitivity calculation method • Ensemble Transform (ET) method developed around 1996 • Field tests • FASTEX – 1997, Atlantic • Impact from sensitive areas compared with that from non-sensitive areas (“null” cases) • NORPEX – 1998, Pacific • Comparison with adjoint methods • CALJET, PACJET, WC-TOST, ATReC, AMMA, T-PARC • WSR • 1999 - First test in research environment • 2000 - Pre-implementation test • 2001 - Full operational implementation
CONTRIBUTIONS • Craig Bishop (NASA, PSU, NRL) • ET & ETKF method development • Sharan Majumdar (PSU, U. Miami) • ETKF method development and implementation • Rolf Langland (NRL), Kerry Emanuel (MIT) • Field testing and comparisons in FASTEX, NORPEX, TPARC • Istvan Szunyogh (UCAR Scientist at NCEP, U. MD, Texas AMU) • Operational implementation, impact analysis, dynamics of data impact • Yucheng Song (Plurality at EMC/NCEP/NWS/NOAA) • Updates, maintenance, coordination • Observations • NOAA Aircraft Operations Center (G-lV) • US Air Force Reserve (C130s) • Operations • Case selection by NWS forecasters (NCEP/HPC, Regions) • Decision making by Senior Duty Meteorologists (SDM)
DOCUMENTATION • Papers (refereed / not reviewed) • Methods • ET Bishop & Toth • ETKF Bishop et al, Majumdar et al • Field tests • Langland et al FASTEX • Langland et al NORPEX • Szunyogh et al FASTEX • Szunyogh et al NORPEX • Song et al TPARC (under preparation) • Operational implementation • Toth et al 2 papers • WSR results • Szunyogh et al • Toth et al (under preparation) • Web • Details on procedures • Detailed documentation for each case in WSR99-09 (11 years, ~200+ cases) • Identification of threatening high impact forecast events • Sensitivity calculation results • Flight request • Data impact analysis
OPERATIONAL PROCEDURES • Case selection • Forecaster input – time and location of high impact event • Based on perceived threat and forecast uncertainty • SDM compiles daily prioritized list of cases for which targeted data may be collected • Ensemble-based sensitivity calculations • Forward assessment • Predict impact of targeted data from predesigned flight tracks • Backward sensitivity • Statistical analysis of forward results for selected verification cases • Decision process • SDM evaluates sensitivity results • Consider predicted impact, priority of cases, available resources • Predesigned flight track # or no flight decision for next day • Outlook for flight / no flight for day after next • Observations • Drop-sondes from manned aircraft flying over predesigned tracks • Aircraft based in Alaska (Anchorage) and/or Hawaii (Honolulu) • Real time QC & transmission to NWP centers via GTS • NWP • Assimilate all adaptively taken data along with regular data • Operational forecasts benefit from targeted data
HIGHLIGHTS • Case selection • No systematic evaluation available • Some errors in position / timing of threatening events in 4-6 day forecast range • Affects stringent verification results • Need for objective case selection guidance based on ensembles • Sensitivity calculations • Predicted and observed impact from targeted data compared in statistical sense • Sensitivity related to dynamics of flow • Variations on daily and longer time scales (regime dependency) • Decision process • Subjective due to limitations in sensitivity methods • Spurious correlations due to small sample size • Observations • Aircraft dedicated to operational observing program used • Are there lower cost alternatives? • Thorough processing of satellite data • UAVs? • NWP forecast improvement • Compare data assimilation / forecast results with / without use of targeted data • Cycled comparison for cumulative impact • One at a time comparison for better tracking of impact dynamics in individual cases
Observed data impact Predicted data impact Forecastimprovement / degradation
WHY TARGETING MAY WORK Impact of data removal over Pacific - Kelly et al. 2007 Figure 1. Winter Pacific forecasts: Verification of mean 500 hPa geopotential rmse up to day 10 for SEAOUT in grey dotted and SEAIN in black: Both experiments are verified using ECMWF operational analysis. Verification regions: (a) North Pacific, (b) North America, (c) North Atlantic and (d) the European region.
FORECAST EVALUATION RESULTS Based on 10 years of experience (1999-2008) • Error reduced in ~70% of targeted forecasts • Verified against observations at preselected time / region • Wind & temperature profiles, surface pressure • 10-20% rms error reduction in preselected regions • Verified against analysis fields • 12-hour gain in predictability • 48-hr forecast with targeted data as skillful as 36-hr forecast without
WSR Summary statistics for 2004-07 Wind vector error, 2007 25+22+19+26= 92 POSITIVE CASES 0+1+0+0 = 1 NEUTRAL CASE 10+7+8+11 = 36 NEGATIVE CASES 71.3% improved 27.9% degraded OVERALL EFFECT: Without targeted data With targeted data
Valentine’s day Storm2007 • Weather event with a large societal impact • Each GFS run verified against its own analysis – 60 hr forecast • Impact on surface pressure verification • RMS error improvement: 19.7% • (2.48mb vs. 2.97mb) • Targeted in high impact weather area marked by the circle Surface pressure from analysis (hPa; solid contours) Forecast Improvement (hPa; shown in red) Forecast Degradation (hPa; blue)
Average surface pressure forecast error reduction from WSR 2000 The average surface pressure forecast error reduction for Alaska (55°–70°N, 165°–140°W), the west coast (25°–50°N, 125°–100°W), the east coast (100°–75°W), and the lower 48 states of the United States (125°–75°W). Positive values show forecast improvement, while negative values show forecast degradation (From Szunyogh et al 2002)
Forecast Verification for Wind (2007) 10-20% rms error reduction in winds Close to 12-hour gain in predictability RMS error reduction vs. forecast lead time
Forecast Verification for Temperature (2007) 10-20 % rms error reduction Close to 12-hour gain in predictability RMS error reduction vs. forecast lead time
CONCLUSIONS • High impact cases can be identified in advance using ensemble methods • Data impact can be predicted in statistical sense using ET / ETKF methods • Optimal observing locations / times for high impact cases can be identified • It is possible to operationally conduct a targeted observational program • Open questions remain
OPEN QUESTIONS • Does operational targeting have economic benefits? • Cost-benefit analysis needs to be done for different regions – SERA research • Are there differences between Pacific (NA) & Atlantic (Europe)? • Can similar or better results be achieved with cheaper observing systems? • Observing systems of opportunity • Targeted processing of satellite data • AMDAR • UAVs? • Sensitivity to data assimilation techniques • Advanced DA methods extracts more info from any data • Better analysis without targeted data • Larger impact from targeted data (relative to improved analysis with standard data)? • What are the limitations of current techniques? • What can be said beyond linear regime? • Need larger ensemble for that? • Can we quantify expected forecast improvement (not only impact)? • Distinction between predicting impact vs. predicting positive impact • Effect of sub-grid scales ignored so far • Ensemble displays more orderly dynamics than reality? • Overly confident signal propagation predictions?
DISCUSSION POINTS How to explain large apparent differences between various studies regarding effectiveness of targeted observations? • Case selection important • Only every ~3rd day there is a “good” case • Targeting is not cure for all diseases • If all cases averaged, signal washed out at factor of 1/3 • Measure impact over target area • Effect expected in specific area • If measured over much larger area, signal washes out by factor of 1/3 • 2 factors above may explain 10-fold difference in quantitative assessment of utility in targeting observations • Not all cases expected to yield positive results • Artifact of statistical nature of DA methods • Should expect some negative impact • Current DA methods lead to forecast improvements in 70-75% of cases • Geographical differences • Potentially larger impact over larger Pacific vs smaller Atlantic basins?
Example: Impact of WSRP targeted dropsondes Binned Impact 1 Jan – 28 Feb 2006 00UTC Analysis Beneficial (-0.01 to -0.1 J kg-1) Non-beneficial (0.01 to 0.1 J kg-1) NOAA-WSRP 191 Profiles Small impact (-0.01 to 0.01 J kg-1) Average dropsonde ob impact is beneficial and ~2-3x greater than average radiosonde impact
Composite summary maps 139.6W 59.8N 36hrs (7 cases) - 1422km 92W 38.6N 60hrs (5 cases)- 4064km 80W 38.6N 63.5hrs (8 cases) - 5143km 122W 37.5N 49.5hrs (8 cases) - 2034km Verification Region Verification Region
North Pacific observation impact sum - NAVDAS Change in 24h moist total energy error norm (J kg-1) 1-31 Jan 2007 (00UTC analyses) Error Reduction
North Pacific forecast error reductionper-observation Change in 24h moist total energy error norm (J kg-1) 1-31 Jan 2007 (00UTC analyses) Error Reduction (x 1.0e5) Ship Obs Targeted dropsondes = high-impact per- ob, low total impact
ETS 5mm 10mm 16.35 18.56 CTL 16.50 20.44 OPR Positive vs. negative cases 4:1 3:1 Precipitation verification • Precipitation verification is still in a testing stage due to the lack of station observation data in some regions.