scholarly journals Objective Classification of Tornadic and Nontornadic Severe Weather Outbreaks

2009 ◽  
Vol 137 (12) ◽  
pp. 4355-4368 ◽  
Author(s):  
Andrew E. Mercer ◽  
Chad M. Shafer ◽  
Charles A. Doswell ◽  
Lance M. Leslie ◽  
Michael B. Richman

Abstract Tornadoes often strike as isolated events, but many occur as part of a major outbreak of tornadoes. Nontornadic outbreaks of severe convective storms are more common across the United States but pose different threats than do those associated with a tornado outbreak. The main goal of this work is to distinguish between significant instances of these outbreak types objectively by using statistical modeling techniques on numerical weather prediction output initialized with synoptic-scale data. The synoptic-scale structure contains information that can be utilized to discriminate between the two types of severe weather outbreaks through statistical methods. The Weather Research and Forecast model (WRF) is initialized with synoptic-scale input data (the NCEP–NCAR reanalysis dataset) on a set of 50 significant tornado outbreaks and 50 nontornadic severe weather outbreaks. Output from the WRF at 18-km grid spacing is used in the objective classification. Individual severe weather parameters forecast by the model near the time of the outbreak are analyzed from simulations initialized at 24, 48, and 72 h prior to the outbreak. An initial candidate set of 15 variables expected to be related to severe storms is reduced to a set of 6 or 7, depending on lead time, that possess the greatest classification capability through permutation testing. These variables serve as inputs into two statistical methods, support vector machines and logistic regression, to classify outbreak type. Each technique is assessed based on bootstrap confidence limits of contingency statistics. An additional backward selection of the reduced variable set is conducted to determine which variable combination provides the optimal contingency statistics. Results for the contingency statistics regarding the verification of discrimination capability are best at 24 h; at 48 h, modest degradation is present. By 72 h, the contingency statistics decline by up to 15%. Overall, results are encouraging, with probability of detection values often exceeding 0.8 and Heidke skill scores in excess of 0.7 at 24-h lead time.

2017 ◽  
Vol 32 (1) ◽  
pp. 117-139 ◽  
Author(s):  
Sanjib Sharma ◽  
Ridwan Siddique ◽  
Nicholas Balderas ◽  
Jose D. Fuentes ◽  
Seann Reed ◽  
...  

Abstract The quality of ensemble precipitation forecasts across the eastern United States is investigated, specifically, version 2 of the National Centers for Environmental Prediction (NCEP) Global Ensemble Forecast System Reforecast (GEFSRv2) and Short Range Ensemble Forecast (SREF) system, as well as NCEP’s Weather Prediction Center probabilistic quantitative precipitation forecast (WPC-PQPF) guidance. The forecasts are verified using multisensor precipitation estimates and various metrics conditioned upon seasonality, precipitation threshold, lead time, and spatial aggregation scale. The forecasts are verified, over the geographic domain of each of the four eastern River Forecasts Centers (RFCs) in the United States, by considering first 1) the three systems or guidance, using a common period of analysis (2012–13) for lead times from 1 to 3 days, and then 2) GEFSRv2 alone, using a longer period (2004–13) and lead times from 1 to 16 days. The verification results indicate that, across the eastern United States, precipitation forecast bias decreases and the skill and reliability improve as the spatial aggregation scale increases; however, all the forecasts exhibit some underforecasting bias. The skill of the forecasts is appreciably better in the cool season than in the warm one. The WPC-PQPFs tend to be superior, in terms of the correlation coefficient, relative mean error, reliability, and forecast skill scores, than both GEFSRv2 and SREF, but the performance varies with the RFC and lead time. Based on GEFSRv2, medium-range precipitation forecasts tend to have skill up to approximately day 7 relative to sampled climatology.


2010 ◽  
Vol 138 (11) ◽  
pp. 4098-4119 ◽  
Author(s):  
Chad M. Shafer ◽  
Andrew E. Mercer ◽  
Lance M. Leslie ◽  
Michael B. Richman ◽  
Charles A. Doswell

Abstract Recent studies, investigating the ability to use the Weather Research and Forecasting (WRF) model to distinguish tornado outbreaks from primarily nontornadic outbreaks when initialized with synoptic-scale data, have suggested that accurate discrimination of outbreak type is possible up to three days in advance of the outbreaks. However, these studies have focused on the most meteorologically significant events without regard to the season in which the outbreaks occurred. Because tornado outbreaks usually occur during the spring and fall seasons, whereas the primarily nontornadic outbreaks develop predominantly during the summer, the results of these studies may have been influenced by climatological conditions (e.g., reduced shear, in the mean, in the summer months), in addition to synoptic-scale processes. This study focuses on the impacts of choosing outbreaks of severe weather during the same time of year. Specifically, primarily nontornadic outbreaks that occurred during the summer have been replaced with outbreaks that do not occur in the summer. Subjective and objective analyses of the outbreak simulations indicate that the WRF’s capability of distinguishing outbreak type correctly is reduced when the seasonal constraints are included. However, accuracy scores exceeding 0.7 and skill scores exceeding 0.5 using 1-day simulation fields of individual meteorological parameters, show that precursor synoptic-scale processes play an important role in the occurrence or absence of tornadoes in severe weather outbreaks. Low-level storm-relative helicity parameters and synoptic parameters, such as geopotential heights and mean sea level pressure, appear to be most helpful in distinguishing outbreak type, whereas thermodynamic instability parameters are noticeably both less accurate and less skillful.


2018 ◽  
Author(s):  
Rochelle P. Worsnop ◽  
Michael Scheuerer ◽  
Thomas M. Hamill ◽  
Julie K. Lundquist

Abstract. Wind power forecasting is gaining international significance as more regions promote policies to increase the use of renewable energy. Wind ramps, large variations in wind power production during a period of minutes to hours, challenge utilities and electrical balancing authorities. A sudden decrease in wind energy production must be balanced by other power generators to meet energy demands, while a sharp increase in unexpected production results in excess power that may not be used in the power grid, leading to a loss of potential profits. In this study, we compare different methods to generate probabilistic ramp forecasts from the High Resolution Rapid Refresh (HRRR) numerical weather prediction model with up to twelve hours of lead time at two tall-tower locations in the United States. We validate model performance using 21 months of 80-m wind speed observations from towers in Boulder, Colorado and near the Columbia River Gorge in eastern Oregon. We employ four statistical post-processing methods, three of which are not currently used in the literature for wind forecasting. These procedures correct biases in the model and generate short-term wind speed scenarios which are then converted to power scenarios. This probabilistic enhancement of HRRR point forecasts provides valuable uncertainty information of ramp events and improves the skill of predicting ramp events over the raw forecasts. We compute Brier skill scores for each method at predicting up- and down-ramps to determine which method provides the best prediction. We find that the Standard Schaake Shuffle method yields the highest skill at predicting ramp events for these data sets, especially for up-ramp events at the Oregon site. Increased skill for ramp prediction is limited at the Boulder, CO site using any of the multivariate methods, because of the poor initial forecasts in this area of complex terrain. These statistical methods can be implemented by wind farm operators to generate a range of possible wind speed and power scenarios to aid and optimize decisions before ramp events occur.


2019 ◽  
Vol 34 (4) ◽  
pp. 1137-1160 ◽  
Author(s):  
Ryan Lagerquist ◽  
Amy McGovern ◽  
David John Gagne II

AbstractThis paper describes the use of convolutional neural nets (CNN), a type of deep learning, to identify fronts in gridded data, followed by a novel postprocessing method that converts probability grids to objects. Synoptic-scale fronts are often associated with extreme weather in the midlatitudes. Predictors are 1000-mb (1 mb = 1 hPa) grids of wind velocity, temperature, specific humidity, wet-bulb potential temperature, and/or geopotential height from the North American Regional Reanalysis. Labels are human-drawn fronts from Weather Prediction Center bulletins. We present two experiments to optimize parameters of the CNN and object conversion. To evaluate our system, we compare the objects (predicted warm and cold fronts) with human-analyzed warm and cold fronts, matching fronts of the same type within a 100- or 250-km neighborhood distance. At 250 km our system obtains a probability of detection of 0.73, success ratio of 0.65 (or false-alarm rate of 0.35), and critical success index of 0.52. These values drastically outperform the baseline, which is a traditional method from numerical frontal analysis. Our system is not intended to replace human meteorologists, but to provide an objective method that can be applied consistently and easily to a large number of cases. Our system could be used, for example, to create climatologies and quantify the spread in forecast frontal properties across members of a numerical weather prediction ensemble.


2016 ◽  
Vol 144 (10) ◽  
pp. 3799-3823 ◽  
Author(s):  
Glen S. Romine ◽  
Craig S. Schwartz ◽  
Ryan D. Torn ◽  
Morris L. Weisman

Over the central Great Plains, mid- to upper-tropospheric weather disturbances often modulate severe storm development. These disturbances frequently pass over the Intermountain West region of the United States during the early morning hours preceding severe weather events. This region has fewer in situ observations of the atmospheric state compared with most other areas of the United States, contributing toward greater uncertainty in forecast initial conditions. Assimilation of supplemental observations is hypothesized to reduce initial condition uncertainty and improve forecasts of high-impact weather. During the spring of 2013, the Mesoscale Predictability Experiment (MPEX) leveraged ensemble-based targeting methods to key in on regions where enhanced observations might reduce mesoscale forecast uncertainty. Observations were obtained with dropsondes released from the NSF/NCAR Gulfstream-V aircraft during the early morning hours preceding 15 severe weather events over areas upstream from anticipated convection. Retrospective data-denial experiments are conducted to evaluate the value of dropsonde observations in improving convection-permitting ensemble forecasts. Results show considerable variation in forecast performance from assimilating dropsonde observations, with a modest but statistically significant improvement, akin to prior targeted observation studies that focused on synoptic-scale prediction. The change in forecast skill with dropsonde information was not sensitive to the skill of the control forecast. Events with large positive impact sampled both the disturbance and adjacent flow, akin to results from past synoptic-scale targeting studies, suggesting that sampling both the disturbance and adjacent flow is necessary regardless of the horizontal scale of the feature of interest.


2018 ◽  
Vol 3 (1) ◽  
pp. 371-393 ◽  
Author(s):  
Rochelle P. Worsnop ◽  
Michael Scheuerer ◽  
Thomas M. Hamill ◽  
Julie K. Lundquist

Abstract. Wind power forecasting is gaining international significance as more regions promote policies to increase the use of renewable energy. Wind ramps, large variations in wind power production during a period of minutes to hours, challenge utilities and electrical balancing authorities. A sudden decrease in wind-energy production must be balanced by other power generators to meet energy demands, while a sharp increase in unexpected production results in excess power that may not be used in the power grid, leading to a loss of potential profits. In this study, we compare different methods to generate probabilistic ramp forecasts from the High Resolution Rapid Refresh (HRRR) numerical weather prediction model with up to 12 h of lead time at two tall-tower locations in the United States. We validate model performance using 21 months of 80 m wind speed observations from towers in Boulder, Colorado, and near the Columbia River gorge in eastern Oregon. We employ four statistical post-processing methods, three of which are not currently used in the literature for wind forecasting. These procedures correct biases in the model and generate short-term wind speed scenarios which are then converted to power scenarios. This probabilistic enhancement of HRRR point forecasts provides valuable uncertainty information of ramp events and improves the skill of predicting ramp events over the raw forecasts. We compute Brier skill scores for each method with regard to predicting up- and down-ramps to determine which method provides the best prediction. We find that the Standard Schaake shuffle method yields the highest skill at predicting ramp events for these datasets, especially for up-ramp events at the Oregon site. Increased skill for ramp prediction is limited at the Boulder, CO, site using any of the multivariate methods because of the poor initial forecasts in this area of complex terrain. These statistical methods can be implemented by wind farm operators to generate a range of possible wind speed and power scenarios to aid and optimize decisions before ramp events occur.


2018 ◽  
Vol 33 (1) ◽  
pp. 331-345 ◽  
Author(s):  
John L. Cintineo ◽  
Michael J. Pavolonis ◽  
Justin M. Sieglaff ◽  
Daniel T. Lindsey ◽  
Lee Cronce ◽  
...  

Abstract The empirical Probability of Severe (ProbSevere) model, developed by the National Oceanic and Atmospheric Administration (NOAA) and the Cooperative Institute for Meteorological Satellite Studies (CIMSS), automatically extracts information related to thunderstorm development from several data sources to produce timely, short-term, statistical forecasts of thunderstorm intensity. More specifically, ProbSevere utilizes short-term numerical weather prediction guidance (NWP), geostationary satellite, ground-based radar, and ground-based lightning data to determine the probability that convective storm cells will produce severe weather up to 90 min in the future. ProbSevere guidance, which updates approximately every 2 min, is available to National Weather Service (NWS) Weather Forecast Offices with very short latency. This paper focuses on the integration of ground-based lightning detection data into ProbSevere. In addition, a thorough validation analysis is presented. The validation analysis demonstrates that ProbSevere has slightly less skill compared to NWS severe weather warnings, but can offer greater lead time to initial hazards. Feedback from NWS users has been highly favorable, with most forecasters responding that ProbSevere increases confidence and lead time in numerous warning situations.


Author(s):  
Makenzie J. Krocak ◽  
Harold E. Brooks

AbstractWhile many studies have looked at the quality of forecast products, few have attempted to understand the relationship between them. We begin to consider whether or not such an influence exists by analyzing storm-based tornado warning product metrics with respect to whether they occurred within a severe weather watch and, if so, what type of watch they occurred within.The probability of detection, false alarm ratio, and lead time all show a general improvement with increasing watch severity. In fact, the probability of detection increased more as a function of watch-type severity than the change in probability of detection during the time period of analysis. False alarm ratio decreased as watch type increased in severity, but with a much smaller magnitude than the difference in probability of detection. Lead time also improved with an increase in watch-type severity. Warnings outside of any watch had a mean lead time of 5.5 minutes, while those inside of a particularly dangerous situation tornado watch had a mean lead time of 15.1 minutes. These results indicate that the existence and type of severe weather watch may have an influence on the quality of tornado warnings. However, it is impossible to separate the influence of weather watches from possible differences in warning strategy or differences in environmental characteristics that make it more or less challenging to warn for tornadoes. Future studies should attempt to disentangle these numerous influences to assess how much influence intermediate products have on downstream products.


2017 ◽  
Vol 98 (2) ◽  
pp. 347-359 ◽  
Author(s):  
Steven M. Martinaitis ◽  
Jonathan J. Gourley ◽  
Zachary L. Flamig ◽  
Elizabeth M. Argyle ◽  
Robert A. Clark ◽  
...  

Abstract There are numerous challenges with the forecasting and detection of flash floods, one of the deadliest weather phenomena in the United States. Statistical metrics of flash flood warnings over recent years depict a generally stagnant warning performance, while regional flash flood guidance utilized in warning operations was shown to have low skill scores. The Hydrometeorological Testbed—Hydrology (HMT-Hydro) experiment was created to allow operational forecasters to assess emerging products and techniques designed to improve the prediction and warning of flash flooding. Scientific goals of the HMT-Hydro experiment included the evaluation of gridded products from the Multi-Radar Multi-Sensor (MRMS) and Flooded Locations and Simulated Hydrographs (FLASH) product suites, including the experimental Coupled Routing and Excess Storage (CREST) model, the application of user-defined probabilistic forecasts in experimental flash flood watches and warnings, and the utility of the Hazard Services software interface with flash flood recommenders in real-time experimental warning operations. The HMT-Hydro experiment ran in collaboration with the Flash Flood and Intense Rainfall (FFaIR) experiment at the Weather Prediction Center to simulate the real-time workflow between a national center and a local forecast office, as well as to facilitate discussions on the challenges of short-term flash flood forecasting. Results from the HMT-Hydro experiment highlighted the utility of MRMS and FLASH products in identifying the spatial coverage and magnitude of flash flooding, while evaluating the perception and reliability of probabilistic forecasts in flash flood watches and warnings. NSSL scientists and NWS forecasters evaluate new tools and techniques through real-time test bed operations for the improvement of flash flood detection and warning operations.


2015 ◽  
Vol 8 (10) ◽  
pp. 11255-11284
Author(s):  
J. Brunner ◽  
R. B. Pierce ◽  
A. Lenzen

Abstract. A satellite based surface visibility retrieval has been developed using Moderate Resolution Imaging Spectroradiometer (MODIS) measurements as a proxy for Advanced Baseline Imager (ABI) data from the next generation of Geostationary Operational Environmental Satellites (GOES-R). The retrieval uses a multiple linear regression approach to relate satellite aerosol optical depth, fog/low cloud probability and thickness retrievals, and meteorological variables from numerical weather prediction forecasts to National Weather Service Automated Surface Observing System (ASOS) surface visibility measurements. Validation using independent ASOS measurements shows that the GOES-R ABI surface visibility retrieval (V) has an overall success rate of 64.5% for classifying Clear (V ≥ 30 km), Moderate (10 km ≤ V < 30 km), Low (2 km ≤ V < 10 km) and Poor (V < 2 km) visibilities and shows the most skill during June through September, when Heidke skill scores are between 0.2 and 0.4. We demonstrate that the aerosol (clear sky) component of the GOES-R ABI visibility retrieval can be used to augment measurements from the United States Environmental Protection Agency (EPA) and National Park Service (NPS) Interagency Monitoring of Protected Visual Environments (IMPROVE) network, and provide useful information to the regional planning offices responsible for developing mitigation strategies required under the EPA's Regional Haze Rule, particularly during regional haze events associated with smoke from wildfires.


Sign in / Sign up

Export Citation Format

Share Document