EIDA: The European Integrated Data Archive and Service Infrastructure within ORFEUS

Author(s):  
Angelo Strollo ◽  
Didem Cambaz ◽  
John Clinton ◽  
Peter Danecek ◽  
Christos P. Evangelidis ◽  
...  

Abstract The European Integrated Data Archive (EIDA) is the infrastructure that provides access to the seismic-waveform archives collected by European agencies. This distributed system is managed by Observatories and Research Facilities for European Seismology. EIDA provides seamless access to seismic data from 12 data archives across Europe by means of standard services, exposing data on behalf of hundreds of network operators and research organizations. More than 12,000 stations from permanent and temporary networks equipped with seismometers, accelerometers, pressure sensors, and other sensors are accessible through the EIDA federated services. A growing user base currently counting around 3000 unique users per year has been requesting data and using EIDA services. The EIDA system is designed to scale up to support additional new services, data types, and nodes. Data holdings, services, and user numbers have grown substantially since the establishment of EIDA in 2013. EIDA is currently active in developing suitable data management approaches for new emerging technologies (e.g., distributed acoustic sensing) and challenges related to big datasets. This article reviews the evolution of EIDA, the current data holdings, and service portfolio, and gives an outlook on the current developments and the future envisaged challenges.

Author(s):  
Christos P. Evangelidis ◽  
Nikolaos Triantafyllis ◽  
Michalis Samios ◽  
Kostas Boukouras ◽  
Kyriakos Kontakos ◽  
...  

Abstract The National Observatory of Athens data center for the European Integrated Data Archive (EIDA@NOA) is the national and regional node that supports International Federation of Digital Seismograph Networks and related webservices for seismic waveform data coming from the southeastern Mediterranean and the Balkans. At present, it serves data from eight permanent broadband and strong-motion networks from Greece and Cyprus, individual stations from the Balkans, temporary networks and aftershock deployments, and earthquake engineering experimental facilities. EIDA@NOA provides open and unlimited access from redundant node end points, intended mainly for research purposes (see Data and Resources). Analysis and quality control of the complete seismic data archive is performed initially by calculating waveform metrics and data availability. Seismic ambient noise metrics are estimated based on power spectral densities, and an assessment of each station’s statistical mode is achieved within each network and across networks. Moreover, the minimum ambient noise level expected for strong-motion installations is defined. Sensor orientation is estimated using surface-wave polarization methods to detect stations with misalignment on particular epochs. A single data center that hosts the complete seismic data archives with their respective metadata from networks covering similar geographical areas allows coordination between network operators and facilitates the adhesion to widely used best practices regarding station installation, data curation, and metadata definition. The overall achievement is harmonization among all contributing networks and a wider usage of all data archives, ultimately strengthening seismological research efforts in the region.


1989 ◽  
Vol 110 ◽  
pp. 139-140
Author(s):  
Chris Benn

The Observatorio del Roque de los Muchachos is perched atop a volcanic caldera on the island of La Palma in the Canary Islands, 400 km off the coast of North Africa. Three of the telescopes at the observatory are products of a collaboration between the UK, the Netherlands, Spain and the Republic of Ireland. They are the 1.0-m Jacobus Kapteyn Telescope, the 2.5-m Isaac Newton Telescope and the 4.2-m William Herschel Telescope (which saw first light in July 1987)1. The telescopes are computer controlled (running under ADAM software), and the observations are recorded primarily in electronic form. Recognising the success of astronomical-satellite data archives, such as that generated by the International Ultraviolet Explorer, a La Palma Data Archive has been established at the Royal Greenwich Observatory. The archive will be used by astronomers wishing to exploit data obtained by other observers, by engineers interested in the performance of telescope and instruments under varying conditions, and for monitoring the way in which the telescope is used.


2010 ◽  
Vol 133-134 ◽  
pp. 485-489
Author(s):  
Elisa Adorni ◽  
Carlo Blasi

Diagnostic analysis, required to characterize materials and mechanical parameters of ancient masonry, needs a systematic collection and a careful reading of data results, as well as the cross-check of results of the different samples collected. This work assumes that diagnostic is part of the knowledge phase of existing buildings: knowledge level is fundamental to evaluate the state of conservation, as prescribed also in italian Technical Standards for Construction (2008). This paper describe some characteristics of a database system, M.A.I.D.A - Masonry Analysis Interpretation and Data Archive, performed to contain diagnostic analysis results of traditional materials and constituted by interactive analysis cards. Cards nomenclature and method follow UNI Normal rules for the construction materials: stones, bricks, mortars. Some tests on M.A.I.D.A system were carried out in a cooperation project between University of Parma and University of Rome ‘La Sapienza’. The project comprehends analysis of the Carolingian masonry and mortars from some buildings in Rome: SS. Quattro Coronati ecclesiastic complex, S. Francesca Romana Church and a Carolingian part of the Aurelian Walls in Rome. In order to enforce the data archives and to share the results M.A.I.D.A system will be put in a internet server.


2006 ◽  
Vol 134 (5) ◽  
pp. 952-960 ◽  
Author(s):  
R. KOSMIDER ◽  
L. KELLY ◽  
S. EVANS ◽  
G. GETTINBY

Worldwide, early detection systems have been used in public health to aid the timely detection of increases in disease reporting that may be indicative of an outbreak. To date, their application to animal surveillance has been limited and statistical methods to analyse human health data have not been viewed as being applicable for animal health surveillance data. This issue was investigated by developing an early detection system for Salmonella disease in British livestock. We conclude that an early detection system, as for public health surveillance, can be an effective tool for enhanced surveillance. In order to implement this system in the future and extend it for other data types, we provide recommendations for improving the current data collection process. These recommendations will ensure that quality surveillance data are collected and used effectively to monitor disease in livestock populations.


2016 ◽  
Vol 18 ◽  
pp. 3-10
Author(s):  
Eva-Maria Gerstner ◽  
Yvonne Bachmann ◽  
Karen Hahn ◽  
Anne Mette Lykke ◽  
Marco Schmidt

Although there is an increasing need for data in ecological studies, many datasets are still lost or not sufficiently visible due to a lack of appropriate data archives. With the West African Data and Metadata Repository, we present a secure long-term archive for a data-poor region allowing detailed documentation by metadata following the EML standard and giving data holders the opportunity to define levels of data access and conditions of use. This article gives an overview of structure, functions and content. The repository is online at the URL http://westafricandata.senckenberg.de.


2021 ◽  
Vol 1 ◽  
pp. 9-10
Author(s):  
Ernst Niederleithinger ◽  
Vera Lay ◽  
Christian Köpp ◽  
Erika Holt ◽  
Maria Oksa

Abstract. The EURATOM PREDIS project (http://www.predis-h2020.eu, last access: 25 October 2021) targets the development and implementation of activities for predisposal treatment of radioactive waste streams other than nuclear fuel and high-level radioactive waste. It started on 1 September 2020 with a 4 year duration. The consortium includes 47 partners from 17 member states. The overall budget of the project is EUR 23.7 million, with EC contribution of EUR 14 million. The PREDIS project develops and increases the technological readiness level (TRL) of treatment and conditioning methodologies for wastes for which no adequate or industrially mature solutions are currently available, including metallic materials, liquid organic waste and solid organic waste. The PREDIS project also develops innovations in cemented waste handling and predisposal storage by testing and evaluating. The technical work packages align with priorities formulated within the Roadmap Theme 2 of EURAD (https://www.ejp-eurad.eu/sites/default/files/2021-09/2_Predisposal_Theme_Overview.pdf, last access: 15 October 2021), Nugenia Global Vision (https://snetp.eu/wp-content/uploads/2020/10/Global-vision-document-ves-1-april-2015-aa.pdf, last access: 15 October 2021) and with those identified by the project's industrial end users group (EUG). The PREDIS will produce tools guiding decision making on the added value of the developed technologies and their impact on the design, safety and economics of waste management and disposal. Four technical work packages are focusing on specific waste types: metallic, liquid organic, solid organic, and cemented wastes. For the first three, the main aim lies in processing, stabilizing, and packaging the different waste streams, e.g. by using novel geopolymers, to deliver items which are in line with national and international waste acceptance criteria. In contrast, the fourth technical work package has a different focus. To provide better ways for a safe and effective monitoring of cemented waste packages including prediction tools to assess the future integrity development during predisposal activities, several digital tools are evaluated and improved. Safety enhancement (e.g. less exposure of testing personnel) and cost-effectiveness are part of the intended impact. The work includes but is not limited to inspection methods, such as muon imaging, wireless sensors integrated into waste packages as well as external package and facility monitoring, such as remote fiber optic sensors. The sensors applied will go beyond radiation monitoring and include proxy parameters important for long-term integrity assessment (e.g. internal pressure). Sensors will also be made cost-effective to allow the installation of many more sensors compared to current practice. The measured data will be used in digital twins of the waste packages for specific simulations (geochemical, integrity) providing a prediction of future behavior. Machine learning techniques trained by the characterization of older waste packages will help to connect the models to the current data. All data (measured and simulated) will be collected in a joint database and connected to a decision framework to be used at actual facilities. The presentation includes detailed information about the various tools under consideration in the monitoring of cemented waste packages, their connection and first results of the research.


Author(s):  
Timothy G. Barraclough

This final chapter summarizes conclusions from the book and highlights a few general areas for future work. The species model for the structure of diversity is found to be useful and largely supported by current data, but is open to future tests against explicit alternative models. It is also a vital component for understanding and predicting contemporary evolution in the diverse systems that all organisms live in. The common evolutionary framework for microbial and multicellular life is highlighted, while drawing attention to current gaps in understanding for each type of organism. Future work needs to scale up to develop model systems of diverse assemblages and clades, including time-series data ranging from contemporary to geological scales. The imminent avalanche of genome data for thousands of individuals sampled within and between species is identified as a key challenge and opportunity. Finally, this chapter repeats the challenge that evolutionary biologists should embrace diversity and need to attempt to predict evolution in diverse systems, in order to deliver solutions of benefit to society.


2021 ◽  
Author(s):  
Philipp Kaestli ◽  
Daniel Armbruster ◽  
The EIDA Technical Committee

<p>With the setup of EIDA (the European Integrated Data Archive https://www.orfeus-eu.org/data/eida/) in the framework of ORFEUS, and the implementation of FDSN-standardized web services, seismic waveform data and instrumentation metadata of most seismic networks and data centers in Europe became accessible in a homogeneous way. EIDA has augmented this with the WFcatalog service for waveform quality metadata, and a routing service to find out which data center offers data of which network, region, and type. However, while a distributed data archive has clear advantages for maintenance and quality control of the holdings, it complicates the life of researchers who wish to collect data archived across different data centers. To tackle this, EIDA has implemented the “federator” as a one-stop transparent gateway service to access the entire data holdings of EIDA.</p><p>To its users the federator acts just like a standard FDSN dataselect, station, or EIDA WFcatalog service, except for the fact that it can (due to a fully qualified internal routing cache) directly answer data requests on virtual networks.</p><p>Technically, the federator fulfills a user request by decomposing it into single stream epoch requests targeted at a single data center, collecting them, and re-assemble them to a single result.</p><p>This implementation has several technical advantages:</p><ul><li>It avoids response size limitations of EIDA member services, reducing limitations to those imposed by assembling cache space of the federator instance itself.</li> <li>It allows easy merging of partial responses using request sorting and concatenation, and reducing needs to interpret them. This reduces computational needs of the federator and allows high throughput of parallel user requests.</li> <li>It reduces the variability of requests to end member services. Thus, the federator can implement a reverse loopback cache and protect end node services from delivering redundant information and reducing their load.</li> <li>As partial results are quick, and delivered in small subunits, they can be streamed to the user more or less continuously, avoiding both service timeouts and throughput bottlenecks.</li> </ul><p>The advantage of having a one-stop data access for entire EIDA still comes with some limitations and shortcomings. Having requests which ultimately map to a single data center performed by the federator can be slower by that data center directly. FDSN-defined standard error codes sent by end member services have limited utility as they refer to a part of the request only. Finally, the federator currently does not provide access to restricted data.</p><p>Nevertheless, we believe that the one-stop data access compensates these shortcomings in many use cases.</p><p>Further documentation of the service is available with ORFEUS at http://www.orfeus-eu.org/data/eida/nodes/FEDERATOR/</p>


2020 ◽  
Vol 15 (1) ◽  
pp. 5
Author(s):  
Laurence Horton ◽  
Anja Perry

In this paper we outline the process of revising data access categories for research data sets in GESIS – a large European social science data archive based in Germany. The challenge is to create a minimal set of workable access conditions that cope with a) facilitating as “open as possible, closed as necessary” expectations for data reuse; b) map on to existing legacy access categories and conditions in a data archive. The paper covers the work done in gathering data on data access categories used by data archives in their existing data catalogues, the choices offered to depositors of data in their user agreements, and work done by other data reuse platforms in categorising access to their data. Finally, we talk through the process of refining a minimal set of data access conditions for the GESIS data archive.  


2018 ◽  
Vol 13 (1) ◽  
pp. 114-124 ◽  
Author(s):  
Thu-Mai Lewis Christian ◽  
Sophia Lafferty-Hess ◽  
William G Jacoby ◽  
Thomas Carsey

In response to widespread concerns about the integrity of research published in scholarly journals, several initiatives have emerged that are promoting research transparency through access to data underlying published scientific findings. Journal editors, in particular, have made a commitment to research transparency by issuing data policies that require authors to submit their data, code, and documentation to data repositories to allow for public access to the data. In the case of the American Journal of Political Science (AJPS) Data Replication Policy, the data also must undergo an independent verification process in which materials are reviewed for quality as a condition of final manuscript publication and acceptance. Aware of the specialized expertise of the data archives, AJPS called upon the Odum Institute Data Archive to provide a data review service that performs data curation and verification of replication datasets. This article presents a case study of the collaboration between AJPS and the Odum Institute Data Archive to develop a workflow that bridges manuscript publication and data review processes. The case study describes the challenges and the successes of the workflow integration, and offers lessons learned that may be applied by other data archives that are considering expanding their services to include data curation and verification services to support reproducible research.


Sign in / Sign up

Export Citation Format

Share Document