data types
Recently Published Documents


TOTAL DOCUMENTS

3451
(FIVE YEARS 1312)

H-INDEX

63
(FIVE YEARS 16)

2022 ◽  
Vol 9 ◽  
Author(s):  
Ning Wei ◽  
Xiaochun Li ◽  
Zhunsheng Jiao ◽  
Philip H. Stauffer ◽  
Shengnan Liu ◽  
...  

Carbon dioxide (CO2) storage in deep saline aquifers is a vital option for CO2 mitigation at a large scale. Determining storage capacity is one of the crucial steps toward large-scale deployment of CO2 storage. Results of capacity assessments tend toward a consensus that sufficient resources are available in saline aquifers in many parts of the world. However, current CO2 capacity assessments involve significant inconsistencies and uncertainties caused by various technical assumptions, storage mechanisms considered, algorithms, and data types and resolutions. Furthermore, other constraint factors (such as techno-economic features, site suitability, risk, regulation, social-economic situation, and policies) significantly affect the storage capacity assessment results. Consequently, a consensus capacity classification system and assessment method should be capable of classifying the capacity type or even more related uncertainties. We present a hierarchical framework of CO2 capacity to define the capacity types based on the various factors, algorithms, and datasets. Finally, a review of onshore CO2 aquifer storage capacity assessments in China is presented as examples to illustrate the feasibility of the proposed hierarchical framework.


2022 ◽  
Vol 6 (POPL) ◽  
pp. 1-29
Author(s):  
Qianchuan Ye ◽  
Benjamin Delaware

Secure computation allows multiple parties to compute joint functions over private data without leaking any sensitive data, typically using powerful cryptographic techniques. Writing secure applications using these techniques directly can be challenging, resulting in the development of several programming languages and compilers that aim to make secure computation accessible. Unfortunately, many of these languages either lack or have limited support for rich recursive data structures, like trees. In this paper, we propose a novel representation of structured data types, which we call oblivious algebraic data types, and a language for writing secure computations using them. This language combines dependent types with constructs for oblivious computation, and provides a security-type system which ensures that adversaries can learn nothing more than the result of a computation. Using this language, authors can write a single function over private data, and then easily build an equivalent secure computation according to a desired public view of their data.


2022 ◽  
Vol 6 (POPL) ◽  
pp. 1-29
Author(s):  
Hari Govind V K ◽  
Sharon Shoham ◽  
Arie Gurfinkel

This work addresses the problem of verifying imperative programs that manipulate data structures, e.g., Rust programs. Data structures are usually modeled by Algebraic Data Types (ADTs) in verification conditions. Inductive invariants of such programs often require recursively defined functions (RDFs) to represent abstractions of data structures. From the logic perspective, this reduces to solving Constrained Horn Clauses (CHCs) modulo both ADT and RDF. The underlying logic with RDFs is undecidable. Thus, even verifying a candidate inductive invariant is undecidable. Similarly, IC3-based algorithms for solving CHCs lose their progress guarantee: they may not find counterexamples when the program is unsafe. We propose a novel IC3-inspired algorithm Racer for solving CHCs modulo ADT and RDF (i.e., automatically synthesizing inductive invariants, as opposed to only verifying them as is done in deductive verification). Racer ensures progress despite the undecidability of the underlying theory, and is guaranteed to terminate with a counterexample for unsafe programs. It works with a general class of RDFs over ADTs called catamorphisms. The key idea is to represent catamorphisms as both CHCs, via relationification , and RDFs, using novel abstractions . Encoding catamorphisms as CHCs allows learning inductive properties of catamorphisms, as well as preserving unsatisfiabilty of the original CHCs despite the use of RDF abstractions, whereas encoding catamorphisms as RDFs allows unfolding the recursive definition, and relying on it in solutions. Abstractions ensure that the underlying theory remains decidable. We implement our approach in Z3 and show that it works well in practice.


2022 ◽  
Vol 6 (POPL) ◽  
pp. 1-32
Author(s):  
Vikraman Choudhury ◽  
Jacek Karwowski ◽  
Amr Sabry

The Pi family of reversible programming languages for boolean circuits is presented as a syntax of combinators witnessing type isomorphisms of algebraic data types. In this paper, we give a denotational semantics for this language, using weak groupoids à la Homotopy Type Theory, and show how to derive an equational theory for it, presented by 2-combinators witnessing equivalences of type isomorphisms. We establish a correspondence between the syntactic groupoid of the language and a formally presented univalent subuniverse of finite types. The correspondence relates 1-combinators to 1-paths, and 2-combinators to 2-paths in the universe, which is shown to be sound and complete for both levels, forming an equivalence of groupoids. We use this to establish a Curry-Howard-Lambek correspondence between Reversible Logic, Reversible Programming Languages, and Symmetric Rig Groupoids, by showing that the syntax of Pi is presented by the free symmetric rig groupoid, given by finite sets and bijections. Using the formalisation of our results, we perform normalisation-by-evaluation, verification and synthesis of reversible logic gates, motivated by examples from quantum computing. We also show how to reason about and transfer theorems between different representations of reversible circuits.


2022 ◽  
Author(s):  
Nadin Rohland ◽  
Swapan Mallick ◽  
Matthew Mah ◽  
Robert M Maier ◽  
Nick J Patterson ◽  
...  

In-solution enrichment for hundreds of thousands of single nucleotide polymorphisms (SNPs) has been the source of >70% of all genome-scale ancient human DNA data published to date. This approach has made it possible to generate data for one to two orders of magnitude lower cost than random shotgun sequencing, making it economical to study ancient samples with low proportions of human DNA, and increasing the rate of conversion of sampled remains into working data thereby facilitating ethical stewardship of human remains. So far, nearly all ancient DNA data obtained using in-solution enrichment has been generated using a set of bait sequences targeting about 1.24 million SNPs (the 1240k reagent). These sequences were published in 2015, but synthesis of the reagent has been cost-effective for only a few laboratories. In 2021, two companies made available reagents that target the same core set of SNPs along with supplementary content. Here, we test the properties of the three reagents on a common set of 27 ancient DNA libraries across a range of richness of DNA content and percentages of human molecules. All three reagents are highly effective at enriching many hundreds of thousands of SNPs. For all three reagents and a wide range of conditions, one round of enrichment produces data that is as useful as two rounds when tens of millions of sequences are read out as is typical for such experiments. In our testing, the Twist Ancient DNA reagent produces the highest coverages, greatest uniformity on targeted positions, and almost no bias toward enriching one allele more than another relative to shotgun sequencing. Allelic bias in 1240k enrichment has made it challenging to carry out joint analysis of these data with shotgun data, creating a situation where the ancient DNA community has been publishing two important bodes of data that cannot easily be co-analyzed by population genetic methods. To address this challenge, we introduce a subset of hundreds of thousands of SNPs for which 1240k data can be effectively co-analyzed with all other major data types.


2022 ◽  
Vol 4 (1) ◽  
Author(s):  
Pavel P Kuksa ◽  
Yuk Yee Leung ◽  
Prabhakaran Gangadharan ◽  
Zivadin Katanic ◽  
Lauren Kleidermacher ◽  
...  

ABSTRACT Querying massive functional genomic and annotation data collections, linking and summarizing the query results across data sources/data types are important steps in high-throughput genomic and genetic analytical workflows. However, these steps are made difficult by the heterogeneity and breadth of data sources, experimental assays, biological conditions/tissues/cell types and file formats. FILER (FunctIonaL gEnomics Repository) is a framework for querying large-scale genomics knowledge with a large, curated integrated catalog of harmonized functional genomic and annotation data coupled with a scalable genomic search and querying interface. FILER uniquely provides: (i) streamlined access to >50 000 harmonized, annotated genomic datasets across >20 integrated data sources, >1100 tissues/cell types and >20 experimental assays; (ii) a scalable genomic querying interface; and (iii) ability to analyze and annotate user’s experimental data. This rich resource spans >17 billion GRCh37/hg19 and GRCh38/hg38 genomic records. Our benchmark querying 7 × 109 hg19 FILER records shows FILER is highly scalable, with a sub-linear 32-fold increase in querying time when increasing the number of queries 1000-fold from 1000 to 1 000 000 intervals. Together, these features facilitate reproducible research and streamline integrating/querying large-scale genomic data within analyses/workflows. FILER can be deployed on cloud or local servers (https://bitbucket.org/wanglab-upenn/FILER) for integration with custom pipelines and is freely available (https://lisanwanglab.org/FILER).


2022 ◽  
Author(s):  
Zhengyi Zhu ◽  
Glen A Satten ◽  
Yi-Juan Hu

We previously developed LDM for testing hypotheses about the microbiome that performs the test at both the community level and the individual taxon level. LDM can be applied to relative abundance data and presence-absence data separately, which work well when associated taxa are abundant and rare, respectively. Here we propose an omnibus test based on LDM that allows simultaneous consideration of data at different scales, thus offering optimal power across scenarios with different association mechanisms. The omnibus test is available for the wide range of data types and analyses that are supported by LDM. The omnibus test has been added to the R package LDM, which is available on GitHub at https://github.com/yijuanhu/LDM .


2022 ◽  
Author(s):  
Gustave Ronteix ◽  
Valentin Bonnet ◽  
Sebastien Sart ◽  
Jeremie Sobel ◽  
Elric Esposito ◽  
...  

Microscopy techniques and image segmentation algorithms have improved dramatically this decade, leading to an ever increasing amount of biological images and a greater reliance on imaging to investigate biological questions. This has created a need for methods to extract the relevant information on the behaviors of cells and their interactions, while reducing the amount of computing power required to organize this information. This task can be performed by using a network representation in which the cells and their properties are encoded in the nodes, while the neighborhood interactions are encoded by the links. Here we introduce Griottes, an open-source tool to build the "network twin" of 2D and 3D tissues from segmented microscopy images. We show how the library can provide a wide range of biologically relevant metrics on individual cells and their neighborhoods, with the objective of providing multi-scale biological insights. The library's capacities are demonstrated on different image and data types. This library is provided as an open-source tool that can be integrated into common image analysis workflows to increase their capacities.


2022 ◽  
Vol 9 (1) ◽  
pp. 33
Author(s):  
Sam McDevitt ◽  
Haley Hernandez ◽  
Jamison Hicks ◽  
Russell Lowell ◽  
Hamza Bentahaikt ◽  
...  

Wearable technologies are emerging as a useful tool with many different applications. While these devices are worn on the human body and can capture numerous data types, this literature review focuses specifically on wearable use for performance enhancement and risk assessment in industrial- and sports-related biomechanical applications. Wearable devices such as exoskeletons, inertial measurement units (IMUs), force sensors, and surface electromyography (EMG) were identified as key technologies that can be used to aid health and safety professionals, ergonomists, and human factors practitioners improve user performance and monitor risk. IMU-based solutions were the most used wearable types in both sectors. Industry largely used biomechanical wearables to assess tasks and risks wholistically, which sports often considered the individual components of movement and performance. Availability, cost, and adoption remain common limitation issues across both sports and industrial applications.


Sign in / Sign up

Export Citation Format

Share Document