A Holistic Approach to Financial Data Science: Data, Technology, and Analytics

2020 ◽  
Vol 2 (2) ◽  
pp. 64-84
Author(s):  
Tamer Khraisha
PLoS ONE ◽  
2021 ◽  
Vol 16 (3) ◽  
pp. e0248128
Author(s):  
Mark Stewart ◽  
Carla Rodriguez-Watson ◽  
Adem Albayrak ◽  
Julius Asubonteng ◽  
Andrew Belli ◽  
...  

Background The COVID-19 pandemic remains a significant global threat. However, despite urgent need, there remains uncertainty surrounding best practices for pharmaceutical interventions to treat COVID-19. In particular, conflicting evidence has emerged surrounding the use of hydroxychloroquine and azithromycin, alone or in combination, for COVID-19. The COVID-19 Evidence Accelerator convened by the Reagan-Udall Foundation for the FDA, in collaboration with Friends of Cancer Research, assembled experts from the health systems research, regulatory science, data science, and epidemiology to participate in a large parallel analysis of different data sets to further explore the effectiveness of these treatments. Methods Electronic health record (EHR) and claims data were extracted from seven separate databases. Parallel analyses were undertaken on data extracted from each source. Each analysis examined time to mortality in hospitalized patients treated with hydroxychloroquine, azithromycin, and the two in combination as compared to patients not treated with either drug. Cox proportional hazards models were used, and propensity score methods were undertaken to adjust for confounding. Frequencies of adverse events in each treatment group were also examined. Results Neither hydroxychloroquine nor azithromycin, alone or in combination, were significantly associated with time to mortality among hospitalized COVID-19 patients. No treatment groups appeared to have an elevated risk of adverse events. Conclusion Administration of hydroxychloroquine, azithromycin, and their combination appeared to have no effect on time to mortality in hospitalized COVID-19 patients. Continued research is needed to clarify best practices surrounding treatment of COVID-19.


Author(s):  
Sabitha Rajagopal

Data Science employs techniques and theories to create data products. Data product is merely a data application that acquires its value from the data itself, and creates more data as a result; it's not just an application with data. Data science involves the methodical study of digital data employing techniques of observation, development, analysis, testing and validation. It tackles the real time challenges by adopting a holistic approach. It ‘creates' knowledge about large and dynamic bases, ‘develops' methods to manage data and ‘optimizes' processes to improve its performance. The goal includes vital investigation and innovation in conjunction with functional exploration intended to notify decision-making for individuals, businesses, and governments. This paper discusses the emergence of Data Science and its subsequent developments in the fields of Data Mining and Data Warehousing. The research focuses on need, challenges, impact, ethics and progress of Data Science. Finally the insights of the subsequent phases in research and development of Data Science is provided.


2019 ◽  
Vol 9 (15) ◽  
pp. 3065 ◽  
Author(s):  
Dresp-Langley ◽  
Ekseth ◽  
Fesl ◽  
Gohshi ◽  
Kurz ◽  
...  

Detecting quality in large unstructured datasets requires capacities far beyond the limits of human perception and communicability and, as a result, there is an emerging trend towards increasingly complex analytic solutions in data science to cope with this problem. This new trend towards analytic complexity represents a severe challenge for the principle of parsimony (Occam’s razor) in science. This review article combines insight from various domains such as physics, computational science, data engineering, and cognitive science to review the specific properties of big data. Problems for detecting data quality without losing the principle of parsimony are then highlighted on the basis of specific examples. Computational building block approaches for data clustering can help to deal with large unstructured datasets in minimized computation time, and meaning can be extracted rapidly from large sets of unstructured image or video data parsimoniously through relatively simple unsupervised machine learning algorithms. Why we still massively lack in expertise for exploiting big data wisely to extract relevant information for specific tasks, recognize patterns and generate new information, or simply store and further process large amounts of sensor data is then reviewed, and examples illustrating why we need subjective views and pragmatic methods to analyze big data contents are brought forward. The review concludes on how cultural differences between East and West are likely to affect the course of big data analytics, and the development of increasingly autonomous artificial intelligence (AI) aimed at coping with the big data deluge in the near future.


2019 ◽  
Vol 214 ◽  
pp. 08026 ◽  
Author(s):  
Raul H. C. Lopes ◽  
Virginia N. L. Franqueira ◽  
Duncan Rand

Two recent and promising additions to the internet protocols are TCP-BBR and QUIC. BBR defines a congestion policy that promises a better control in TCP bottlenecks on long haul transfers and can also be used in the QUIC protocol. TCP-BBR is implemented in the Linux kernels above 4.9. It has been shown, however, to demand careful fine tuning in the interaction, for example, with the Linux Fair Queue. QUIC, on the other hand, replaces HTTP and TLS with a protocol on the top of UDP and thin layer to serve HTTP. It has been reported to account today for 7% of Google’s traffic. It has not been used in server-to-server transfers even if its creators see that as a real possibility. Our work evaluates the applicability and tuning of TCP-BBR and QUIC for data science transfers. We describe the deployment and performance evaluation of TCP-BBR and comparison with CUBIC and H-TCP in transfers through the TEIN link to Singaren (Singapore). Also described is the deployment and initial evaluation of a QUIC server. We argue that QUIC might be a perfect match in security and connectivity to base services that are today performed by the Xroot redirectors.


2019 ◽  
Author(s):  
Jennifer Goldsack ◽  
Andrea Coravos ◽  
Jessie Bakker ◽  
Brinnae Bent ◽  
Ariel V. Dowling ◽  
...  

UNSTRUCTURED Digital medicine is an interdisciplinary field, drawing together stakeholders with expertise in engineering, manufacturing, clinical science, data science, biostatistics, regulatory considerations, ethics, patient advocacy, and healthcare policy, to name a few. While this diversity is undoubtedly valuable, it can lead to confusion regarding terminology and best practices. There are many instances, as we detail in this paper, where a single term is used by different groups to mean different things, as well as cases where multiple terms are used to describe essentially the same concept. Our intent is to clarify core terminology and best practices for the evaluation of Biometric Monitoring Technologies (BioMeTs), without unnecessarily introducing new terms. We propose and describe a three-component framework intended to provide a foundational evaluation framework for BioMeTs. This framework includes 1) verification, 2) analytical validation, and 3) clinical validation. We aim for this common vocabulary to enable more effective communication and collaboration, generate a common and meaningful evidence base for BioMeTs, and improve the accessibility of the digital medicine field.


Author(s):  
Oleksandr Burov

Keywords: human capital, remote work, cybersecurity, workforce, digital economics The article considers the role of human capital in the transitionto the remote work. The analysis of world changes in the field of safe and effectiveuse of digital business environment and qualification of workforce in the conditions ofgrowth of remote work is carried out. The analysis was conducted in the following areas:general features of the digitalizing in crisis and innovation, a new paradigm of business«Data is the new gold», the organization of the workforce in the transition to teleworking,the priorities of today's professions, the problems of cybersecurity in teleworking. It has been articulated that the main requirements for the today’s workforce are intellectualand creative abilities, competence in the field of creation and use of ICT, bigdata (data science, data mining, data analytics) and artificial intelligence, the role ofwhich has grown even more due to the COVID-19 pandemic. The human component ofintellectual capital (in the form of knowledge, skills and competencies, as well as intellectualand creative abilities) is gaining new importance in the digital economy.The analysis of relationship of the crisis and innovation made on the basis of the ClarivateDerwent report has demonstrated the impact of the pandemic on the global lifecycle of research and innovation projects in the first half of 2020, namely that COVID-19violated innovation strategy of the innovative leaders worldwide. The analysis hasdemonstrated: in the new conditions of accelerated digitalization, ingenuity and speed ofdecision-making and innovation are needed more than ever. These priorities will affectthe world economy in the coming year.Special attention in analysis has been paid to the new business paradigm related touse and role of data. It was highlighted that digitization generates vast amounts of datathat offer many opportunities for business, human well-being, and the environment. As aresult, new capabilities and opportunities arise for business with the ecosystem of cooperationand partnership, as well as collaboration of stakeholders.The core of changes in digitalization is reskilling and upskilling of the workforce accountingnew workplaces and new requirements for them. It is recognized that talentmanagement and creative people selection can be the main engine in future transformationof economics, and workforce becomes an effective pole for investments. At the sametime, it is argued that remote worker is outside the scope of corporate protection, and virtuallyany production information, like human capital, becomes much more vulnerablein such conditions and requires appropriate cybersecurity methods.As a conclusion, it is articulated that the ability of companies to use big data is beginningto play a significant role in the economy, which in turn requires the involvementand training of data processing and analysis specialists. The direction of professions thatis being actively formed recently — data science — is one of the most priority in the labormarket. At the same time, the labor market needs skills and abilities in the field of interpersonalcommunication (soft skills), which are able to ensure the effective operation ofpeople and systems of hybrid intelligence «human-artificial intelligence».For the further research it has been recommended a comprehensive study of protectionof objects and subjects of intellectual property in open networks.


2016 ◽  
Author(s):  
Cristina Ponte Lira ◽  
Ana Nobre Silva ◽  
Rui Taborda ◽  
Cesar Freire de Andrade

Abstract. Regional/global-scale information on coastline rates of change and trends is extremely valuable, but national-scale studies are scarce. A widely accepted standardized methodology for analysing long-term coastline change has been difficult to achieve, but is essential to conduct an integrated and holistic approach to coastline evolution and hence support coastal management actions. Additionally, databases providing knowledge on coastline evolution are of key importance to support both coastal management experts and users. The main objective of this work is to present the first systematic, global and consistent long-term coastline evolution data of Portuguese mainland low-lying sandy. The methodology used quantifies coastline evolution using an unique and robust coastline indicator (the foredune toe), which is independent of short-term changes. The dataset presented comprises: 1) two polyline sets, mapping the 1958 and 2010 sandy beach-dune systems coastline, both optimized for working at 1:50 000 scale or smaller, and 2) one polyline set representing long-term change rates between 1958 and 2010, estimated at each 250 m. The science data produced here are in Open Access at doi:10.1594/PANGAEA.853654 and can be used in other studies. Results show beach erosion as the dominant trend, with a mean change rate of −0.24 ± 0.01 m/year for all mainland Portuguese beach-dune systems. Although erosion is dominant, this evolution is variable in signal and magnitude in different coastal sediment cell and also within each cell. The most relevant beach erosion issues were found in the coastal stretches of Espinho – Torreira and Costa Nova – Praia da Mira, both at sub-cell 1b; Cova Gala – Leirosa, at sub-cell 1c and Cova do Vapor – Costa da Caparica, at cell 4. Cells 1 and 4 exhibit a history of major human interventions interfering with the coastal system, many of which originated and maintained a sediment deficit. In contrast, cells 5 and 6 have been less intervened and show stable or moderate accretion behaviour.


2021 ◽  
Vol 8 (2) ◽  
pp. 205395172110401
Author(s):  
Anna Sapienza ◽  
Sune Lehmann

For better and worse, our world has been transformed by Big Data. To understand digital traces generated by individuals, we need to design multidisciplinary approaches that combine social and data science. Data and social scientists face the challenge of effectively building upon each other’s approaches to overcome the limitations inherent in each side. Here, we offer a “data science perspective” on the challenges that arise when working to establish this interdisciplinary environment. We discuss how we perceive the differences and commonalities of the questions we ask to understand digital behaviors (including how we answer them), and how our methods may complement each other. Finally, we describe what a path toward common ground between these fields looks like when viewed from data science.


Sign in / Sign up

Export Citation Format

Share Document