scholarly journals Pangeo Forge: Crowdsourcing Analysis-Ready, Cloud Optimized Data Production

2021 ◽  
Author(s):  
Charles Stern ◽  
Ryan Abernathey ◽  
Joseph Hamman ◽  
Rachel Wegener ◽  
Chiara Lepore ◽  
...  

Pangeo Forge is a new community-driven platform that accelerates science by providing high-level recipe frameworks alongside cloud compute infrastructure for extracting data from provider archives, transforming it into analysis-ready, cloud-optimized (ARCO) data stores, and providing a human- and machine-readable catalog for browsing and loading. In abstracting the scientific domain logic of data recipes from cloud infrastructure concerns, Pangeo Forge aims to open a door for a broader community of scientists to participate in ARCO data production. A wholly open-source platform composed of multiple modular components, Pangeo Forge presents a foundation for the practice of reproducible, cloud-native, big-data ocean, weather, and climate science without relying on proprietary or cloud-vendor-specific tooling.

Author(s):  
Abhishek Dubey

The term 'Big Data' portrays inventive methods and advances to catch, store, disseminate, oversee and break down petabyte-or bigger estimated sets of data with high-speed & diverted structures. Enormous information can be organized, non-structured or half-organized, bringing about inadequacy of routine information administration techniques. Information is produced from different distinctive sources and can touch base in the framework at different rates. With a specific end goal to handle this lot of information in an economical and proficient way, parallelism is utilized. Big Data is information whose scale, differences, and unpredictability require new engineering, methods, calculations, and investigation to oversee it and concentrate esteem and concealed learning from it. Hadoop is the center stage for organizing Big Data, and takes care of the issue of making it valuable for examination purposes. Hadoop is an open source programming venture that empowers the dispersed handling of huge information sets crosswise over bunches of ware servers. It is intended to scale up from a solitary server to a huge number of machines, with a high level of adaptation to non-critical failure.


Author(s):  
Michael Goul ◽  
T. S. Raghu ◽  
Ziru Li

As procurement organizations increasingly move from a cost-and-efficiency emphasis to a profit-and-growth emphasis, flexible data architecture will become an integral part of a procurement analytics strategy. It is therefore imperative for procurement leaders to understand and address digitization trends in supply chains and to develop strategies to create robust data architecture and analytics strategies for the future. This chapter assesses and examines the ways companies can organize their procurement data architectures in the big data space to mitigate current limitations and to lay foundations for the discovery of new insights. It sets out to understand and define the levels of maturity in procurement organizations as they pertain to the capture, curation, exploitation, and management of procurement data. The chapter then develops a framework for articulating the value proposition of moving between maturity levels and examines what the future entails for companies with mature data architectures. In addition to surveying the practitioner and academic research literature on procurement data analytics, the chapter presents detailed and structured interviews with over fifteen procurement experts from companies around the globe. The chapter finds several important and useful strategies that have helped procurement organizations design strategic roadmaps for the development of robust data architectures. It then further identifies four archetype procurement area data architecture contexts. In addition, this chapter details exemplary high-level mature data architecture for each archetype and examines the critical assumptions underlying each one. Data architectures built for the future need a design approach that supports both descriptive and real-time, prescriptive analytics.


Sensors ◽  
2021 ◽  
Vol 21 (11) ◽  
pp. 3850
Author(s):  
Bastien Vincke ◽  
Sergio Rodriguez Rodriguez Florez ◽  
Pascal Aubert

Emerging technologies in the context of Autonomous Vehicles (AV) have drastically evolved the industry’s qualification requirements. AVs incorporate complex perception and control systems. Teaching the associated skills that are necessary for the analysis of such systems becomes a very difficult process and existing solutions do not facilitate learning. In this study, our efforts are devoted to proposingan open-source scale model vehicle platform that is designed for teaching the fundamental concepts of autonomous vehicles technologies that are adapted to undergraduate and technical students. The proposed platform is as realistic as possible in order to present and address all of the fundamental concepts that are associated with AV. It includes all on-board components of a stand-alone system, including low and high level functions. Such functionalities are detailed and a proof of concept prototype is presented. A set of experiments is carried out, and the results obtained using this prototype validate the usability of the model for the analysis of time- and energy-constrained systems, as well as distributed embedded perception systems.


2017 ◽  

As machine-readable data comes to play an increasingly important role in everyday life, researchers find themselves with rich resources for studying society. The novel methods and tools needed to work with such data require not only new knowledge and skills, but also a new way of thinking about best research practices. This book critically reflects on the role and usefulness of big data, challenging overly optimistic expectations about what such information can reveal, introducing practices and methods for its analysis and visualisation, and raising important political and ethical questions regarding its collection, handling, and presentation.


Infolib ◽  
2021 ◽  
Vol 26 (2) ◽  
pp. 34-37
Author(s):  
Anna Chulyan ◽  

The article touches upon the importance of long-term digital preservation of Armenian cultural heritage through creation of digital repositories using Open-Source Software in Armenian libraries. The research highlights the advantages of Open-Source Software in context of providing free access to digital materials, as well as its high level of functionality in order to empower libraries with new technologies for more efficient organization and dissemination of information.


2021 ◽  
Vol 125 (3) ◽  
pp. 972-976
Author(s):  
Myles W. O’Brien ◽  
Jennifer L. Petterson ◽  
Derek S. Kimmerly

The pressor responses to spontaneous bursts of muscle sympathetic nerve activity provide important information regarding sympathetic regulation of the circulation. Many laboratories worldwide quantify sympathetic neurohemodynamic transduction using in-house, customized software requiring high-level programming skills and/or costly computer programs. To overcome these barriers, this study presents a simple, open-source, Microsoft Excel-based analysis program along with video instructions to assist researchers without the necessary resources to quantify sympathetic neurohemodynamic transduction.


2012 ◽  
Vol 2 (1) ◽  
pp. 44-57 ◽  
Author(s):  
Yannis Siahos ◽  
Iasonas Papanagiotou ◽  
Alkis Georgopoulos ◽  
Fotis Tsamis ◽  
Ioannis Papaioannou

The authors present their experience and practices of introducing cloud services, as a means to simplify the adoption of ICT (Information Communication and Technology) in education, using Free/Open Source Software. The solution creates a hybrid cloud infrastructure, in order to provide a pre-installed (Ubuntu and Linux Terminal Server Project) virtual machine, acting as a server inside the school, providing desktop environment based on the Software as a Service cloud model, where legacy PCs act as stateless devices. Classroom management is accomplished using the application “Epoptes.” To minimize administration tasks, educational software is provided accordingly, either on-line or through repositories to automate software installation (including patches and updates). The advantages of the hybrid cloud implementation, include services that are not completely dependent on broadband connections’ state, minimal cost, reusability of obsolete equipment, ease of administration, centralized management, patches and educational software provisioning and, above all, facilitation of the educational procedure.


2017 ◽  
Vol 10 (2) ◽  
pp. 174-183 ◽  
Author(s):  
Sidney D’Mello ◽  
Arvid Kappas ◽  
Jonathan Gratch

Affective computing (AC) adopts a computational approach to study affect. We highlight the AC approach towards automated affect measures that jointly model machine-readable physiological/behavioral signals with affect estimates as reported by humans or experimentally elicited. We describe the conceptual and computational foundations of the approach followed by two case studies: one on discrimination between genuine and faked expressions of pain in the lab, and the second on measuring nonbasic affect in the wild. We discuss applications of the measures, analyze measurement accuracy and generalizability, and highlight advances afforded by computational tipping points, such as big data, wearable sensing, crowdsourcing, and deep learning. We conclude by advocating for increasing synergies between AC and affective science and offer suggestions toward that direction.


Sign in / Sign up

Export Citation Format

Share Document