scholarly journals Data Migration from One Database to Another: Nervous breakdown of a database manager!

Author(s):  
Henry Engledow

Migrating from one database to another is always accompanied by challenges. Recently Meise Botanic Garden has migrated its Living Collections data from LivCol, a bespoke database, to BG-BASE, a commercial product. Differences in database structure, degree of atomisation and field definition increase the complexity of such a transfer of data. The greater the number of fields used in the original data source, the greater the number of problems there are to resolve. Living collections are often centered around the 'accession information' of the living material, but the way one does this and the philosophy behind this may differ. The different approach to accessioning material in LivCol and BG-BASE affected the structure of the data model in each program. The LivCol approach was not as strictly defined as the BG-BASE approach e.g. new generations derived from existing accessions in LivCol retained the same accession number despite being not genetically identical (of seed origin), whereas in BG-BASE a new accession number would be generated with reference to the parent accession. In the data transfer LivCol accession number where grouped by accession number and garden location, and the inter-generation information combined in a single record in BG-BASE (this is not ideally in accordance with the BG-BASE concept, the alternative was to create 'artificial' new accession numbers but this would have complicated matters more both from a data and management point of view). The use of standards would greatly improve data transfer, and indeed many standards have been adopted by both the above mentioned databases. However, it soon became evident that there are multiple standards for a single topic e.g. for information concerning conservation status: NatureServe Global Conservation Status Ranks; Fish & Wildlife conservation category; International Union for Conservation of Nature (IUCN) - old and new codes (plus version); Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES); etc. Also, some values used are region specific and do not translate well to ‘global’ standards e.g. in Belgium there are three principal regions controlling conservation status (Brussels, Flanders, Wallonia) and they differ in their approach and definitions, all these need to be taken into account as there are legal implications - this was done by finding close matches in IUCN (New) codes and combining them with 'non-standard' World Geographical Scheme for Recording Plant Distributions (WGSRPD). The latter TDWG standard is out of date and in many circumstances not sufficiently atomised to be of practical use. There were also certain fields that would benefit from having standards, but are at present absent e.g. invasiveness - BG-BASE uses Cronk and Fuller (1995) whereas LivCol uses AlterIAS (http://www.alterias.be/), Belgian Forum on Invasive Species (http://ias.biodiversity.be) and Lambinon et al. (1992). The above problems will be discussed and their impact on mapping the data. Decisions had to be made with respect to ‘best fit’ solutions. The latter lead to the loss of information or a slight variation in its interpretation, examples will be given to highlight these aspects. As the structures of the databases differed, sometimes assumptions had to be made, this too will be illustrated. These changes were reasonable, but represent an interpretation of the original data and therefore not strictly the same. Databases are crucial to the management of Living Collections and the research done on them. This talk will look at the lessons learned during the data transfer and the problems associated with mapping (decisions, assumptions and standards). Databases are 'living' entities that need to grow, adapt, be maintained and regularly updated to new developments in technology. Databases are not seen as new or innovative by funding bodies and are often left to struggle along in suboptimal conditions. If we want data quality to improve and increase interoperability between systems, maybe we should start at the point where data is entered.

Author(s):  
M. B. Sergeev ◽  
V. A. Nenashev ◽  
A. M. Sergeev

Introduction: The problem of noise-free encoding for an open radio channel is of great importance for data transfer. The results presented in this paper are aimed at stimulating scientific interest in new codes and bases derived from quasi-orthogonal matrices, as a basis for the revision of signal processing algorithms.Purpose: Search for new code sequences as combinations of codes formed from the rows of Mersenne and Raghavarao quasi-orthogonal matrices, as well as complex and more efficient Barker — Mersenne — Raghavarao codes.Results: We studied nested code sequences derived from the rows of quasi-orthogonal cyclic matrices of Mersenne, Raghavarao and Hadamard, providing estimates for the characteristics of the autocorrelation function of nested Barker, Mersenne and Raghavarao codes, and their combinations: in particular, the ratio between the main peak and the maximum positive and negative “side lobes”. We have synthesized new codes, including nested ones, formed on the basis of quasi-orthogonal matrices with better characteristics than the known Barker codes and their nested constructions. The results are significant, as this research influences the establishment and development of methods for isolation, detection and processing of useful information. The results of the work have a long aftermath because new original code synthesis methods need to be studied, modified, generalized and expanded for new application fields.Practical relevance: The practical application of the obtained results guarantees an increase in accuracy of location systems, and detection of a useful signal in noisy background. In particular, these results can be used in radar systems with high distance resolution, when detecting physical objects, including hidden ones.


<i>Abstract</i>.—Fish population recoveries can result from ecosystem change in the absence of targeted restoration actions. In Lake Ontario, native Deepwater Sculpin <i>Myoxocephalus thompsonii</i> were common in the late 1800s, but by the mid-1900s the species was possibly extirpated. During this period, mineral nutrient inputs increased and piscivore abundance declined, which increased the abundance of the nonnative planktivores Alewife <i>Alosa pseudoharengus</i> and Rainbow Smelt <i>Osmerus mordax</i>. Deepwater Sculpin larvae are pelagic and vulnerable to predation by planktivores. Annual bottom trawl surveys did not capture Deepwater Sculpin from 1978 to 1995 (<i>n</i> = 6,666 tows) despite sampling appropriate habitat (trawl depths: 7–170 m). The absence of observations during this time resulted in an elevated conservation status for the species, but no restoration actions were initiated. In 1996, three individuals were caught in bottom trawls, the first observed since 1972. Since then, their abundance has increased, and in 2017, they were the second most abundant Lake Ontario prey fish. The food-web changes that occurred from 1970 through the 1990s contributed to this recovery. Alewife and Rainbow Smelt abundance declined during this period due to predation by stocked salmonids and legislation that reduced nutrient inputs and food web productivity. In the 1990s, proliferation of nonnative, filter-feeding dreissenid mussels dramatically increased water clarity. As light penetration increased, the early-spring depth distribution of Alewife and Rainbow Smelt shifted deeper, away from larval Deepwater Sculpin habitat. The intentional and unintentional changes that occurred in Lake Ontario were not targeted at Deepwater Sculpin restoration but resulted in conditions that favored the species’ recovery. While standard surveys documented the recovery, more diverse information (e.g., observations in deep habitats and early-life stages) would have improved our understanding of why the species recovered when it did. Annual Lake Ontario trawl surveys have collaboratively expanded their spatial extent and diversified habitat sampled, based on lessons learned from the Deepwater Sculpin recovery.


2011 ◽  
pp. 2094-2113
Author(s):  
Amparo C. Villablanca ◽  
Hassan Baxi ◽  
Kent Anderson

This chapter discusses critical success factors in the design, implementation, and utility of a new construct and interface for data transfer with broad applicability to clinical data set management. In the context of a data coordinating center for evaluating cardiovascular outcomes in high-risk women, we detail and provide a framework for bridging the gap between extensible markup language (XML) and XML schema definition file (XSD) in order to provide greater accessibility using visual basic for applications (VBA) and Excel. Applications and lessons learned are discussed in light of current challenges to healthcare information technology management and clinical data administration. The authors hope that this approach, as well as the logic utilized and implementation examples, will provide a user-friendly model for data management and relational database design that is replicable, flexible, understandable, and has broad utility to research professionals in healthcare.


Smart Cities ◽  
2019 ◽  
Vol 2 (1) ◽  
pp. 106-117
Author(s):  
Chengxi Siew ◽  
Pankaj Kumar

Spatial Data Infrastructures (SDIs) are frequently used to exchange 2D & 3D data, in areas such as city planning, disaster management, urban navigation and many more. City Geography Mark-up Language (CityGML), an Open Geospatial Consortium (OGC) standard has been developed for the storage and exchange of 3D city models. Due to its encoding in XML based format, the data transfer efficiency is reduced which leads to data storage issues. The use of CityGML for analysis purposes is limited due to its inefficiency in terms of file size and bandwidth consumption. This paper introduces XML based compression technique and elaborates how data efficiency can be achieved with the use of schema-aware encoder. We particularly present CityGML Schema Aware Compressor (CitySAC), which is a compression approach for CityGML data transaction within SDI framework. Our test results show that the encoding system produces smaller file size in comparison with existing state-of-the-art compression methods. The encoding process significantly reduces the file size up to 7–10% of the original data.


2007 ◽  
Vol 20 (3) ◽  
pp. 252 ◽  
Author(s):  
Stephen D. Hopper ◽  
Andrew P. Brown

Drakaea Lindley, 1840 is a genus of 10 species of geophytic orchids endemic to the South-west Australian Floristic Region. The genus is renowned for its morphological and chemical adaptations, achieving pollination by sexual deception of male thynnid wasps. The history of taxa in Drakaea has been one of dispute and confusion right to the present day. Here we provide a revision of the genus, the first made by using modern collections and field data, formalising names for undescribed taxa featured by Hoffman and Brown (1992, 1998), several of which are threatened with extinction. We describe six new species: D. andrewsiae, D. concolor, D. confluens, D. gracilis, D. isolata and D. micrantha. Experimental baiting of male wasps has helped show the specific status of some of these new taxa. Molecular phylogenetic research is needed to clarify relationships and patterns of speciation in the genus. Five of the 10 Drakaea species are legally protected under the Western Australian Wildlife Conservation Act and the Commonwealth Environment Protection and Biodiversity Conservation Act, signalling the ongoing need for research and management to ensure the conservation of this unique part of Australia’s orchid heritage. D. andrewsiae has been recorded only three times from the Gnowangerup–Tunney district. Urgent surveys are needed to establish its conservation status.


2018 ◽  
Vol 25 (2) ◽  
pp. 92-104 ◽  
Author(s):  
Ward Priestman ◽  
Shankar Sridharan ◽  
Helen Vigne ◽  
Richard Collins ◽  
Loretta Seamer ◽  
...  

BackgroundNumerous studies have examined factors related to success, failure and implications of electronic patient record (EPR) system implementations, but usually limited to specific aspects.ObjectiveTo review the published peer-reviewed literature and present findings regarding factors important in relation to successful EPR implementations and likely impact on subsequent clinical activity.MethodLiterature review.ResultsThree hundred and twelve potential articles were identified on initial search, of which 117 were relevant and included in the review. Several factors were related to implementation success, such as good leadership and management, infrastructure support, staff training and focus on workflows and usability. In general, EPR implementation is associated with improvements in documentation and screening performance and reduced prescribing errors, whereas there are minimal available data in other areas such as effects on clinical patient outcomes. The peer-reviewed literature appears to under-represent a range of technical factors important for EPR implementations, such as data migration from existing systems and impact of organisational readiness.ConclusionThe findings presented here represent the synthesis of data from peer-reviewed literature in the field and should be of value to provide the evidence-base for organisations considering how best to implement an EPR system.


Author(s):  
Naresh P ◽  
Rajyalakshmi P ◽  
Krishna Vempati ◽  
Saidulu D

Cloud acts as a data storage and also used for data transfer from one cloud to other. Here data exchange takes place among cloud centers of organizations. At each cloud center huge amount of data was stored, which interns hard to store and retrieve information from it. While migrating the data there are some issues like low data transfer rate, end to end latency issues and data storage issues will occur. As data was distributed among so many cloud centers from single source, will reduces the speed of migration. In distributed cloud computing it is very difficult to transfer the data fast and securely. This paper explores MapReduce within the distributed cloud architecture where MapReduce assists at each cloud. It strengthens the data migration process with the help of HDFS. Compared to existing cloud migration approach the proposed approach gives accurate results interns of speed, time and efficiency.


2020 ◽  
Vol 26 (1) ◽  
pp. 78-83
Author(s):  
Demet Cidem Dogan ◽  
Huseyin Altindis

With introduction of smart things into our lives, cloud computing is used in many different areas and changes the communication method. However, cloud computing should guarantee the complete security assurance in terms of privacy protection, confidentiality, and integrity. In this paper, a Homomorphic Encryption Scheme based on Elliptic Curve Cryptography (HES-ECC) is proposed for secure data transfer and storage. The scheme stores the data in the cloud after encrypting them. While calculations, such as addition or multiplication, are applied to encrypted data on cloud, these calculations are transmitted to the original data without any decryption process. Thus, the cloud server has only ability of accessing the encrypted data for performing the required computations and for fulfilling requested actions by the user. Hence, storage and transmission security of data are ensured. The proposed public key HES-ECC is designed using modified Weil-pairing for encryption and additional homomorphic property. HES-ECC also uses bilinear pairing for multiplicative homomorphic property. Security of encryption scheme and its homomorphic aspects are based on the hardness of Elliptic Curve Discrete Logarithm Problem (ECDLP), Weil Diffie-Hellman Problem (WDHP), and Bilinear Diffie-Helman Problem (BDHP).


2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Ma Haifeng ◽  
Yu HaiTao ◽  
Zhang Ji ◽  
Wang Junhua ◽  
Xue Qingshui ◽  
...  

Now, many users have stored files on multiple clouds, and sometime, a large number of files are migrated between clouds. Because cloud providers and cloud servers are not entirely trusted, the corruption of user’s files event occur from time to time during the processes of storage and migration. Therefore, integrity verification must be performed, and the time verification overhead should be as low as possible. The existing provable data migrate methods still have the issue of high time overhead when a large number of files are migrated. Aiming at this problem, this paper proposed a hierarchical provable data migration method, which can provide the efficiency of data transfer integrity verification when moving large number of continuous files between clouds. In this paper, the proposed method is described in detail as well as the security analysis performance evaluation. The results proved that the proposed method can significantly decrease the detection latency of files transfer between clouds.


Sign in / Sign up

Export Citation Format

Share Document