scholarly journals A multi-collocation method for coastal zone observations with applications to Sentinel-3A altimeter wave height data

Ocean Science ◽  
2019 ◽  
Vol 15 (2) ◽  
pp. 249-268 ◽  
Author(s):  
Johannes Schulz-Stellenfleth ◽  
Joanna Staneva

Abstract. In many coastal areas there is an increasing number and variety of observation data available, which are often very heterogeneous in their temporal and spatial sampling characteristics. With the advent of new systems, like the radar altimeter on board the Sentinel-3A satellite, a lot of questions arise concerning the accuracy and added value of different instruments and numerical models. Quantification of errors is a key factor for applications, like data assimilation and forecast improvement. In the past, the triple collocation method to estimate systematic and stochastic errors of measurements and numerical models was successfully applied to different data sets. This method relies on the assumption that three independent data sets provide estimates of the same quantity. In coastal areas with strong gradients even small distances between measurements can lead to larger differences and this assumption can become critical. In this study the triple collocation method is extended in different ways with the specific problems of the coast in mind. In addition to nearest-neighbour approximations considered so far, the presented method allows for use of a large variety of interpolation approaches to take spatial variations in the observed area into account. Observation and numerical model errors can therefore be estimated, even if the distance between the different data sources is too large to assume that they measure the same quantity. If the number of observations is sufficient, the method can also be used to estimate error correlations between certain data source components. As a second novelty, an estimator for the uncertainty in the derived observation errors is derived as a function of the covariance matrices of the input data and the number of available samples. In the first step, the method is assessed using synthetic observations and Monte Carlo simulations. The technique is then applied to a data set of Sentinel-3A altimeter measurements, in situ wave observations, and numerical wave model data with a focus on the North Sea. Stochastic observation errors for the significant wave height, as well as bias and calibration errors, are derived for the model and the altimeter. The analysis indicates a slight overestimation of altimeter wave heights, which become more pronounced at higher sea states. The smallest stochastic errors are found for the in situ measurements. Different observation geometries of in situ data and altimeter tracks are furthermore analysed, considering 1-D and 2-D interpolation approaches. For example, the geometry of an altimeter track passing between two in situ wave instruments is considered with model data being available at the in situ locations. It is shown that for a sufficiently large sample, the errors of all data sources, as well as the error correlations of the model, can be estimated with the new method.

2018 ◽  
Author(s):  
Johannes Schulz-Stellenfleth ◽  
Joanna Staneva

Abstract. In many coastal areas there is an increasing number and variety of observation data available, which are often very heterogeneous in their temporal and spatial sampling characteristics. With the advent of new systems, like the radar altimeter onboard the SENTINEL-3a satellite, a lot of questions arise concerning the accuracy and added value of different instruments and numerical models. Quantification of errors is a key factor for applications, like data assimilation and forecast improvement. In the past, the triple collocation method to estimate systematic and stochastic errors of measurements and numerical models was successfully applied to different data sets. This method relies on the assumption, that three independent data sets provide estimates of the same quantity. In coastal areas with strong gradients even small distances between measurements can lead to larger differences and this assumption can become critical. In this study the triple collocation method is extended in different ways with the specific problems of the coast in mind. In addition to nearest neighbor approximations considered so far, the presented method allows to use a large variety of interpolation approaches to take spatial variations in the observed area into account. Observation and numerical model errors can therefore be estimated, even if the distance between the different data sources is too big to assume, that they measure the same quantity. If the number of observations is sufficient, the method can also be used to estimate error correlations between certain data source components. As a second novelty, an estimator for the uncertainty of the derived observation errors is derived as a function of the covariance matrices of the input data and the number of available samples. In the first step, the method is assessed using synthetic observations and Monte Carlo simulations. The technique is then applied to a data set of SENTINEL-3a altimeter measurements, insitu wave observation, and numerical wave model data with a focus on the North Sea. Stochastic observation errors for the significant wave height, as well as bias and calibration errors are derived for the model and the altimeter. The analysis indicates a slight overestimation of altimeter wave heights, which becomes more pronounced at higher sea states. The smallest stochastic errors are found for the insitu measurements. Different observation geometries of insitu data and altimeter tracks are furthermore analysed, considering 1D and 2D interpolation approaches. For example, the geometry of an altimeter track passing between two insitu wave instruments is considered with model data being available at the insitu locations. It is shown, that for a sufficiently large sample, the errors of all data sources, as well as the error correlations of the model, can be estimated with the new method.


Author(s):  
Gus Jeans ◽  
Dave Quantrell ◽  
Andrew Watson ◽  
Laure Grignon ◽  
Gil Lizcano

Engineering design codes specify a variety of different relationships to quantify vertical variations in wind speed, gust factor and turbulence intensity. These are required to support applications including assessment of wind resource, operability and engineering design. Differences between the available relationships lead to undesirable uncertainty in all stages of an offshore wind project. Reducing these uncertainties will become increasingly important as wind energy is harnessed in deeper waters and at lower costs. Installation of a traditional met mast is not an option in deep water. Reliable measurement of the local wind, gust and turbulence profiles from floating LiDAR can be challenging. Fortunately, alternative data sources can provide improved characterisation of winds at offshore locations. Numerical modelling of wind in the lower few hundred metres of the atmosphere is generally much simpler at remote deepwater locations than over complex onshore terrain. The sophistication, resolution and reliability of such models is advancing rapidly. Mesoscale models can now allow nesting of large scale conditions to horizontal scales less than one kilometre. Models can also provide many decades of wind data, a major advantage over the site specific measurements gathered to support a wind energy development. Model data are also immediately available at the start of a project at relatively low cost. At offshore locations these models can be validated and calibrated, just above the sea surface, using well established satellite wind products. Reliable long term statistics of near surface wind can be used to quantify winds at the higher elevations applicable to wind turbines using the wide range of existing standard profile relationships. Reduced uncertainty in these profile relationships will be of considerable benefit to the wider use of satellite and model data sources in the wind energy industry. This paper describes a new assessment of various industry standard wind profile relationships, using a range of available met mast datasets and numerical models.


2013 ◽  
Vol 6 (2) ◽  
pp. 779-809 ◽  
Author(s):  
B. Geyer

Abstract. The coastDat data sets were produced to give a consistent and homogeneous database mainly for assessing weather statistics and long-term changes for Europe, especially in data sparse regions. A sequence of numerical models was employed to reconstruct all aspects of marine climate (such as storms, waves, surges etc.) over many decades. Here, we describe the atmospheric part of coastDat2 (Geyer and Rockel, 2013, doi:10.1594/WDCC/coastDat-2_COSMO-CLM). It consists of a regional climate reconstruction for entire Europe, including Baltic and North Sea and parts of the Atlantic. The simulation was done for 1948 to 2012 with a regional climate model and a horizontal grid size of 0.22° in rotated coordinates. Global reanalysis data were used as forcing and spectral nudging was applied. To meet the demands on the coastDat data set about 70 variables are stored hourly.


2010 ◽  
Vol 651 ◽  
pp. 37-64 ◽  
Author(s):  
Ian C. Madsen ◽  
Ian E. Grey ◽  
Stuart J. Mills

A study of the thermal decomposition sequence of a sample of natural arsenian plumbojarosite has been undertaken using in situ X-ray diffraction. The sample was heated to 900°C using an Anton-Paar heating stage fitted to an INEL CPS120 diffractometer. The data were analysed using a whole-pattern, Rietveld based approach for the extraction of quantitative phase abundances. The instrument configuration used required the development and application of algorithms to correct for aberrations in the (i) peak intensities due to differing path lengths of incident and diffracted beams in the sample and (ii) peak positions due to sample displacement. Details of the structural models used were refined at selected steps in the pattern and then fixed for subsequent analysis. The data sequence consists of some 110 individual data sets which were analysed sequentially with the output of each run forming the input for analysis of the next data set. The results of the analysis show a complex breakdown and recrystallisation sequence including the formation of a major amount of amorphous material after initial breakdown of the plumbojarosite.


2018 ◽  
Vol 22 (1) ◽  
pp. 241-263 ◽  
Author(s):  
Yu Zhang ◽  
Ming Pan ◽  
Justin Sheffield ◽  
Amanda L. Siemann ◽  
Colby K. Fisher ◽  
...  

Abstract. Closing the terrestrial water budget is necessary to provide consistent estimates of budget components for understanding water resources and changes over time. Given the lack of in situ observations of budget components at anything but local scale, merging information from multiple data sources (e.g., in situ observation, satellite remote sensing, land surface model, and reanalysis) through data assimilation techniques that optimize the estimation of fluxes is a promising approach. Conditioned on the current limited data availability, a systematic method is developed to optimally combine multiple available data sources for precipitation (P), evapotranspiration (ET), runoff (R), and the total water storage change (TWSC) at 0.5∘ spatial resolution globally and to obtain water budget closure (i.e., to enforce P-ET-R-TWSC= 0) through a constrained Kalman filter (CKF) data assimilation technique under the assumption that the deviation from the ensemble mean of all data sources for the same budget variable is used as a proxy of the uncertainty in individual water budget variables. The resulting long-term (1984–2010), monthly 0.5∘ resolution global terrestrial water cycle Climate Data Record (CDR) data set is developed under the auspices of the National Aeronautics and Space Administration (NASA) Earth System Data Records (ESDRs) program. This data set serves to bridge the gap between sparsely gauged regions and the regions with sufficient in situ observations in investigating the temporal and spatial variability in the terrestrial hydrology at multiple scales. The CDR created in this study is validated against in situ measurements like river discharge from the Global Runoff Data Centre (GRDC) and the United States Geological Survey (USGS), and ET from FLUXNET. The data set is shown to be reliable and can serve the scientific community in understanding historical climate variability in water cycle fluxes and stores, benchmarking the current climate, and validating models.


2019 ◽  
Vol 35 (1) ◽  
pp. 137-165
Author(s):  
Jack Lothian ◽  
Anders Holmberg ◽  
Allyson Seyb

Abstract The linking of disparate data sets across time, space and sources is probably the foremost current issue facing Central Statistical Agencies (CSA). If one reviews the current literature looking for the prevalent challenges facing CSAs, three issues stand out: 1) using administrative data effectively; 2) big data and what it means for CSAs; and 3) integrating disparate data set (such as health, education and wealth) to provide measurable facts that can guide policy makers. CSAs are being challenged to explore the same kind of challenges faced by Google, Facebook, and Yahoo, which are using graphical/semantic web models for organizing, searching and analysing data. Additionally, time and space (geography) are becoming more important dimensions (domains) for CSAs as they start to explore new data sources and ways to integrate those to study relationships. Central agency methodologists are being pushed to include these new perspectives into their standard theories, practises and policies. Like most methodologists, the authors see surveys and the publications of their results as a process where estimation is the key tool to achieve the final goal of an accurate statistical output. Randomness and sampling exists to support this goal, and early on it was clear to us that the incoming “it-is-what-it-is” data sources were not randomly selected. These sources were obviously biased and thus would produce biased estimates. So, we set out to design a strategy to deal with this issue. This article presents a schema for integrating and linking traditional and non-traditional datasets. Like all survey methodologies, this schema addresses the fundamental issues of representativeness, estimation and total survey error measurement.


2021 ◽  
Author(s):  
Wouter Dorigo ◽  
Irene Himmelbauer ◽  
Daniel Aberer ◽  
Lukas Schremmer ◽  
Ivana Petrakovic ◽  
...  

Abstract. In 2009, the International Soil Moisture Network (ISMN) was initiated as a community effort, funded by the European Space Agency, to serve as a centralised data hosting facility for globally available in situ soil moisture measurements (Dorigo et al., 2011a, b). The ISMN brings together in situ soil moisture measurements collected and freely shared by a multitude of organisations, harmonizes them in terms of units and sampling rates, applies advanced quality control, and stores them in a database. Users can freely retrieve the data from this database through an online web portal (https://ismn.earth). Meanwhile, the ISMN has evolved into the primary in situ soil moisture reference database worldwide, as evidenced by more than 3000 active users and over 1000 scientific publications referencing the data sets provided by the network. As of December 2020, the ISMN now contains data of 65 networks and 2678 stations located all over the globe, with a time period spanning from 1952 to present.The number of networks and stations covered by the ISMN is still growing and many of the data sets contained in the database continue to be updated. The main scope of this paper is to inform readers about the evolution of the ISMN over the past decade,including a description of network and data set updates and quality control procedures. A comprehensive review of existing literature making use of ISMN data is also provided in order to identify current limitations in functionality and data usage, and to shape priorities for the next decade of operations of this unique community-based data repository.


Healthcare ◽  
2018 ◽  
Vol 6 (4) ◽  
pp. 136 ◽  
Author(s):  
Stephanie Partridge ◽  
Eloise Howse ◽  
Gwynnyth Llewellyn ◽  
Margaret Allman-Farinelli

Young adulthood is a period of transition, which for many includes higher education. Higher education is associated with specific risks to wellbeing. Understanding the available data on wellbeing in this group may help inform the future collection of data to inform policy and practice in the sector. This scoping review aimed to identify the availability of data sources on the wellbeing of the Australian young adult population who are attending tertiary education. Using the methods of Arksey and O’Malley, data from three primary sources, i.e., Australian Bureau of Statistics, Australian Institute of Health and Welfare and relevant longitudinal studies, were identified. Data sources were screened and coded, and relevant information was extracted. Key data for eight areas related to wellbeing, namely, family and community, health, education and training, work, economic wellbeing, housing, crime and justice, and culture and leisure sources were identified. Forty individual data sets from 16 surveys and six active longitudinal studies were identified. Two data sets contained seven of the areas of wellbeing, of which one was specific to young adults in tertiary education, while the other survey was not limited to young adults. Both data sets lacked information concerning crime and justice variables, which have recently been identified as being of major concern among Australian university students. We recommend that government policy address the collection of a comprehensive data set encompassing each of the eight areas of wellbeing to inform future policy and practice.


2006 ◽  
Vol 8 (2) ◽  
pp. 141-148 ◽  
Author(s):  
A. K. M. Saiful Islam ◽  
Michael Piasecki

Sharing of data sets between numerical models is considered an important and pressing issue in the modeling community, because of (i) the time consumed to convert data sets and (ii) the need to connect different types of numerical codes to better map inter-connectedness of aquatic domains. One of the reasons for the data sharing problem arises from the lack of sufficient description of the data, or lack of metadata, which is due to the absence of a standardized framework for these metadata sets. This paper describes the development of a metadata framework for hydrodynamic data descriptions using the Geographic Information Metadata, 19115:2003 standard published by the International Standards Organization (ISO). This standard has been chosen not only because of its extent and adequacy to describe geospatial data, but also because of its widespread use and flexibility to extend the coverage. The latter is particularly important, as further extensions of the metadata standard are needed to provide a comprehensive metadata representation of hydrodynamics and their I/O data. In order to enable the community to share and reuse numerical code data sets, however, they need to be published in both human and machine understandable format. One such format is the Web Ontology language (OWL), whose syntax is compliant with the Extensible Markup Language (XML). In this paper, we present an extensive metadata profile using the available elements of ISO 19115:2003 as well as its extension rules. Based on the metadata profile, an explicit specification or ontology for the model data domain has been created using OWL. The use of OWL not only permits flexibility when extending the coverage but also to share data sets as resources across the internet as part of the Semantic Web. We demonstrate the use of the framework using a two-dimensional finite element code and its associated data sets.


2016 ◽  
Vol 72 (9) ◽  
pp. 1026-1035 ◽  
Author(s):  
Ulrich Zander ◽  
Michele Cianci ◽  
Nicolas Foos ◽  
Catarina S. Silva ◽  
Luca Mazzei ◽  
...  

Recent advances in macromolecular crystallography have made it practical to rapidly collect hundreds of sub-data sets consisting of small oscillations of incomplete data. This approach, generally referred to as serial crystallography, has many uses, including an increased effective dose per data set, the collection of data from crystals without harvesting (in situdata collection) and studies of dynamic events such as catalytic reactions. However, selecting which data sets from this type of experiment should be merged can be challenging and new methods are required. Here, it is shown that a genetic algorithm can be used for this purpose, and five case studies are presented in which the merging statistics are significantly improved compared with conventional merging of all data.


Sign in / Sign up

Export Citation Format

Share Document