Skill Scores and modified Lorenz domination in default forecasts

2019 ◽  
Vol 181 ◽  
pp. 61-64
Author(s):  
Walter Krämer ◽  
Simon Neumärker
2021 ◽  
Vol 13 (5) ◽  
pp. 1042
Author(s):  
Jung-Hyun Yang ◽  
Jung-Moon Yoo ◽  
Yong-Sang Choi

The detection of low stratus and fog (LSF) at dawn remains limited because of their optical features and weak solar radiation. LSF could be better identified by simultaneous observations of two geostationary satellites from different viewing angles. The present study developed an advanced dual-satellite method (DSM) using FY-4A and Himawari-8 for LSF detection at dawn in terms of probability indices. Optimal thresholds for identifying the LSF from the spectral tests in DSM were determined by the comparison with ground observations of fog and clear sky in/around Japan between April to November of 2018. Then the validation of these thresholds was carried out for the same months of 2019. The DSM essentially used two traditional single-satellite tests for daytime such as the 0.65-μm reflectance (R0.65), and the brightness temperature difference between 3.7 μm and 11 μm (BTD3.7-11); in addition to four more tests such as Himawari-8 R0.65 and BTD13.5-8.5, the dual-satellite stereoscopic difference in BTD3.7-11 (ΔBTD3.7-11), and that in the Normalized Difference Snow Index (ΔNDSI). The four were found to show very high skill scores (POD: 0.82 ± 0.04; FAR, 0.10 ± 0.04). The radiative transfer simulation supported optical characteristics of LSF in observations. The LSF probability indices (average POD: 0.83, FAR: 0.10) were constructed by a statistical combination of the four to derive the five-class probability values of LSF occurrence in a grid. The indices provided more details and useful results in LSF spatial distribution, compared to the single satellite observations (i.e., R0.65 and/or BTD3.7-11) of either LSF or no LSF. The present DSM could apply for remote sensing of environmental phenomena if the stereoscopic viewing angle between two satellites is appropriate.


2014 ◽  
Vol 18 (11) ◽  
pp. 4467-4484 ◽  
Author(s):  
B. Revilla-Romero ◽  
J. Thielen ◽  
P. Salamon ◽  
T. De Groeve ◽  
G. R. Brakenridge

Abstract. One of the main challenges for global hydrological modelling is the limited availability of observational data for calibration and model verification. This is particularly the case for real-time applications. This problem could potentially be overcome if discharge measurements based on satellite data were sufficiently accurate to substitute for ground-based measurements. The aim of this study is to test the potentials and constraints of the remote sensing signal of the Global Flood Detection System for converting the flood detection signal into river discharge values. The study uses data for 322 river measurement locations in Africa, Asia, Europe, North America and South America. Satellite discharge measurements were calibrated for these sites and a validation analysis with in situ discharge was performed. The locations with very good performance will be used in a future project where satellite discharge measurements are obtained on a daily basis to fill the gaps where real-time ground observations are not available. These include several international river locations in Africa: the Niger, Volta and Zambezi rivers. Analysis of the potential factors affecting the satellite signal was based on a classification decision tree (random forest) and showed that mean discharge, climatic region, land cover and upstream catchment area are the dominant variables which determine good or poor performance of the measure\\-ment sites. In general terms, higher skill scores were obtained for locations with one or more of the following characteristics: a river width higher than 1km; a large floodplain area and in flooded forest, a potential flooded area greater than 40%; sparse vegetation, croplands or grasslands and closed to open and open forest; leaf area index > 2; tropical climatic area; and without hydraulic infrastructures. Also, locations where river ice cover is seasonally present obtained higher skill scores. This work provides guidance on the best locations and limitations for estimating discharge values from these daily satellite signals.


2010 ◽  
Vol 27 (3) ◽  
pp. 409-427 ◽  
Author(s):  
Kun Tao ◽  
Ana P. Barros

Abstract The objective of spatial downscaling strategies is to increase the information content of coarse datasets at smaller scales. In the case of quantitative precipitation estimation (QPE) for hydrological applications, the goal is to close the scale gap between the spatial resolution of coarse datasets (e.g., gridded satellite precipitation products at resolution L × L) and the high resolution (l × l; L ≫ l) necessary to capture the spatial features that determine spatial variability of water flows and water stores in the landscape. In essence, the downscaling process consists of weaving subgrid-scale heterogeneity over a desired range of wavelengths in the original field. The defining question is, which properties, statistical and otherwise, of the target field (the known observable at the desired spatial resolution) should be matched, with the caveat that downscaling methods be as a general as possible and therefore ideally without case-specific constraints and/or calibration requirements? Here, the attention is focused on two simple fractal downscaling methods using iterated functions systems (IFS) and fractal Brownian surfaces (FBS) that meet this requirement. The two methods were applied to disaggregate spatially 27 summertime convective storms in the central United States during 2007 at three consecutive times (1800, 2100, and 0000 UTC, thus 81 fields overall) from the Tropical Rainfall Measuring Mission (TRMM) version 6 (V6) 3B42 precipitation product (∼25-km grid spacing) to the same resolution as the NCEP stage IV products (∼4-km grid spacing). Results from bilinear interpolation are used as the control. A fundamental distinction between IFS and FBS is that the latter implies a distribution of downscaled fields and thus an ensemble solution, whereas the former provides a single solution. The downscaling effectiveness is assessed using fractal measures (the spectral exponent β, fractal dimension D, Hurst coefficient H, and roughness amplitude R) and traditional operational scores statistics scores [false alarm rate (FR), probability of detection (PD), threat score (TS), and Heidke skill score (HSS)], as well as bias and the root-mean-square error (RMSE). The results show that both IFS and FBS fractal interpolation perform well with regard to operational skill scores, and they meet the additional requirement of generating structurally consistent fields. Furthermore, confidence intervals can be directly generated from the FBS ensemble. The results were used to diagnose errors relevant for hydrometeorological applications, in particular a spatial displacement with characteristic length of at least 50 km (2500 km2) in the location of peak rainfall intensities for the cases studied.


2013 ◽  
Vol 6 (1) ◽  
pp. 1269-1310 ◽  
Author(s):  
T. Zinner ◽  
C. Forster ◽  
E. de Coning ◽  
H.-D. Betz

Abstract. In this manuscript, recent changes to the DLR METEOSAT thunderstorm TRacking And Monitoring algorithm (Cb-TRAM) are presented as well as a validation of Cb-TRAM against the European ground-based LIghtning NETwork data (LINET) of Nowcast GmbH and Lightning Detection Network (LDN) data of the South African Weather Service (SAWS). The validation is conducted along the well known skill scores probability of detection (POD) and false alarm ratio (FAR) on the basis of METEOSAT/SEVIRI pixels as well as on the basis of thunderstorm objects. The values obtained demonstrate the limits of Cb-TRAM in specific as well as the limits of satellite methods in general which are based on thermal emission and solar reflectivity information from thunderstorm tops. Although the climatic conditions and the occurence of thunderstorms is quite different for Europe and South Africa, the quality score values are similar. Our conclusion is that Cb-TRAM provides robust results of well-defined quality for very different climatic regimes. The POD for a thunderstorm with intense lightning is about 80% during the day. The FAR for a Cb-TRAM detected thunderstorm which is not at least close to intense lightning activity is about 50%; if the proximity to any lightning activity is evaluated the FAR is even much lower at about 15%. Pixel-based analysis shows that the detected thunderstorm object size is not indiscriminately large, but well within the physical limitations of the method. Nighttime POD and FAR are somewhat worse as the detection scheme can not use high resolution visible information. Nowcasting scores show useful values up to approximatelly 30 min.


2012 ◽  
Vol 8 (1) ◽  
pp. 89-115 ◽  
Author(s):  
V. K. C. Venema ◽  
O. Mestre ◽  
E. Aguilar ◽  
I. Auer ◽  
J. A. Guijarro ◽  
...  

Abstract. The COST (European Cooperation in Science and Technology) Action ES0601: advances in homogenization methods of climate series: an integrated approach (HOME) has executed a blind intercomparison and validation study for monthly homogenization algorithms. Time series of monthly temperature and precipitation were evaluated because of their importance for climate studies and because they represent two important types of statistics (additive and multiplicative). The algorithms were validated against a realistic benchmark dataset. The benchmark contains real inhomogeneous data as well as simulated data with inserted inhomogeneities. Random independent break-type inhomogeneities with normally distributed breakpoint sizes were added to the simulated datasets. To approximate real world conditions, breaks were introduced that occur simultaneously in multiple station series within a simulated network of station data. The simulated time series also contained outliers, missing data periods and local station trends. Further, a stochastic nonlinear global (network-wide) trend was added. Participants provided 25 separate homogenized contributions as part of the blind study. After the deadline at which details of the imposed inhomogeneities were revealed, 22 additional solutions were submitted. These homogenized datasets were assessed by a number of performance metrics including (i) the centered root mean square error relative to the true homogeneous value at various averaging scales, (ii) the error in linear trend estimates and (iii) traditional contingency skill scores. The metrics were computed both using the individual station series as well as the network average regional series. The performance of the contributions depends significantly on the error metric considered. Contingency scores by themselves are not very informative. Although relative homogenization algorithms typically improve the homogeneity of temperature data, only the best ones improve precipitation data. Training the users on homogenization software was found to be very important. Moreover, state-of-the-art relative homogenization algorithms developed to work with an inhomogeneous reference are shown to perform best. The study showed that automatic algorithms can perform as well as manual ones.


2010 ◽  
Vol 138 (11) ◽  
pp. 4098-4119 ◽  
Author(s):  
Chad M. Shafer ◽  
Andrew E. Mercer ◽  
Lance M. Leslie ◽  
Michael B. Richman ◽  
Charles A. Doswell

Abstract Recent studies, investigating the ability to use the Weather Research and Forecasting (WRF) model to distinguish tornado outbreaks from primarily nontornadic outbreaks when initialized with synoptic-scale data, have suggested that accurate discrimination of outbreak type is possible up to three days in advance of the outbreaks. However, these studies have focused on the most meteorologically significant events without regard to the season in which the outbreaks occurred. Because tornado outbreaks usually occur during the spring and fall seasons, whereas the primarily nontornadic outbreaks develop predominantly during the summer, the results of these studies may have been influenced by climatological conditions (e.g., reduced shear, in the mean, in the summer months), in addition to synoptic-scale processes. This study focuses on the impacts of choosing outbreaks of severe weather during the same time of year. Specifically, primarily nontornadic outbreaks that occurred during the summer have been replaced with outbreaks that do not occur in the summer. Subjective and objective analyses of the outbreak simulations indicate that the WRF’s capability of distinguishing outbreak type correctly is reduced when the seasonal constraints are included. However, accuracy scores exceeding 0.7 and skill scores exceeding 0.5 using 1-day simulation fields of individual meteorological parameters, show that precursor synoptic-scale processes play an important role in the occurrence or absence of tornadoes in severe weather outbreaks. Low-level storm-relative helicity parameters and synoptic parameters, such as geopotential heights and mean sea level pressure, appear to be most helpful in distinguishing outbreak type, whereas thermodynamic instability parameters are noticeably both less accurate and less skillful.


2021 ◽  
Author(s):  
Andrea Ficchì ◽  
Hannah Cloke ◽  
Linda Speight ◽  
Douglas Mulangwa ◽  
Irene Amuron ◽  
...  

<p>Global flood forecasting systems are helpful in complementing local resources and in-country data to support humanitarians and trigger early action before an impactful flood occurs. Freely available global flood forecast information from the European Commission’s Global Flood Awareness System (GloFAS, a Copernicus EMS service) is being used by the Uganda Red Cross Society (URCS) alongside in-country knowledge to develop appropriate triggers for early actions for flood preparedness, within the Forecast-based Financing (FbF) initiative. To scale up the first FbF pilot to a national level, in 2020 URCS collaborated with several partners including the Red Cross Red Crescent Climate Centre (RCCC), the Uganda’s Ministry of Water and Environment, through the Directorate of Water Resources Management (DWRM), the Uganda National Meteorological Authority (UNMA), the 510 Global team and the University of Reading, through the UK-supported project Forecasts for Anticipatory Humanitarian Action (FATHUM). The new Early Action Protocol (EAP) for floods, submitted to the IFRC’s validation committee in September 2020, is now under review.</p><p>One of the aims of an EAP is to set the triggers for early action, based on forecast skill information, alongside providing a local risk analysis, and describing the early actions, operational procedures, and responsibilities. Working alongside our partners and practitioners in Uganda, we developed a methodology to tailor flood forecast skill analysis to EAP development, that could be potentially useful for humanitarians in other Countries and forecasters engaging with them. The key aim of the analysis is to identify skilful lead times and appropriate triggers for early action based on available operational forecasts, considering action parameters, such as an Action Lifetime of 30 days, and focusing on relevant flood thresholds and skill scores. We analysed the skill of probabilistic flood forecasts from the operational GloFAS (v2.1) system across Uganda against river flow observations and reanalysis data. One of the challenges was to combine operational needs with statistical robustness requirements, using relevant flood thresholds for action. Here we present the results from the analysis carried out for Uganda and the verification workflow, that we plan to make openly available to all practitioners and scientists working on the implementation of forecast-based actions.</p>


2021 ◽  
pp. 1
Author(s):  
Jacob Coburn ◽  
S.C. Pryor

AbstractThis work quantitatively evaluates the fidelity with which the Northern Annular Mode (NAM), Southern Annular Mode (SAM), Pacific-North American pattern (PNA), El Niño-Southern Oscillation (ENSO), Pacific Decadal Oscillation (PDO) and Atlantic Multidecadal Oscillation (AMO) and the first-order mode interactions are represented in Earth System Model (ESM) output from the CMIP6 archive. Several skill metrics are used as part of a differential credibility assessment (DCA) of both spatial and temporal characteristics of the modes across ESMs, ESM families and specific ESM realizations relative to ERA5. The spatial patterns and probability distributions are generally well represented but skill scores that measure the degree to which the frequencies of maximum variance are captured are consistently lower for most ESMs and climate modes. Substantial variability in skill scores manifests across realizations from individual ESMs for the PNA and oceanic modes. Further, the ESMs consistently overestimate the strength of the NAM-PNA first-order interaction and underestimate the NAM-AMO connection. These results suggest that the choice of ESM and ESM realizations will continue to play a critical role in determining climate projections at the global and regional scale at least in the near-term.


2021 ◽  
Author(s):  
Samuele Segoni ◽  
Minu Treesa Abraham ◽  
Neelima Satyam ◽  
Ascanio Rosi ◽  
Biswajeet Pradhan

<p>SIGMA (Sistema Integrato Gestione Monitoraggio Allerta – integrated system for management, monitoring and alerting) is a landslide forecasting model at regional scale which is operational in Emilia Romagna (Italy) for more than 20 years. It was conceived to be operated with a sparse rain gauge network with coarse (daily) temporal resolution and to account for both shallow landslides (typically triggered by short and intense rainstorms) and deep seated landslides (typically triggered by long and less intense rainfalls). SIGMA model is based on the statistical distribution of cumulative rainfall values (calculated over varying time windows), and rainfall thresholds are defined as the multiples of standard deviation of the same, to identify anomalous rainfalls with the potential of triggering landslides.</p><p>In this study, SIGMA model is applied for the first time in a geographical location outside of Italy, i.e. Kalimpong town in India. The SIGMA algorithm is customized using the historical rainfall and landslide data of Kalimpong from 2010 to 2015 and has been validated using the data from 2016 to 2017. The model was validated by building a confusion matrix and calculating statistical skill scores, which were compared with those of the state-of-the-art intensity-duration rainfall thresholds derived for the region.</p><p>Results of the comparison clearly show that SIGMA performs much better than the other models in forecasting landslides: all instances of the validation confusion matrix are improved, and all skill scores are higher than I-D thresholds, with an efficiency of 92% and a likelihood ratio of 11.28. We explain this outcome mainly with technical characteristics of the site: when only daily rainfall measurements from a spare gauge network are available, SIGMA outperforms other approaches based on peak measurements, like intensity – duration thresholds, which cannot be captured adequately by daily measurements. SIGMA model thus showed a good potential to be used as a part of the local Landslide Early Warning System (LEWS).</p>


Sign in / Sign up

Export Citation Format

Share Document