Differential Privacy Preserving Genomic Data Releasing via Factor Graph

Author(s):  
Zaobo He ◽  
Yingshu Li ◽  
Jinbao Wang
2019 ◽  
Author(s):  
Nour Almadhoun ◽  
Erman Ayday ◽  
Özgür Ulusoy

Abstract Motivation The rapid progress in genome sequencing has led to high availability of genomic data. However, due to growing privacy concerns about the participant’s sensitive information, accessing results and data of genomic studies is restricted to only trusted individuals. On the other hand, paving the way to biomedical discoveries requires granting open access to genomic databases. Privacy-preserving mechanisms can be a solution for granting wider access to such data while protecting their owners. In particular, there has been growing interest in applying the concept of differential privacy (DP) while sharing summary statistics about genomic data. DP provides a mathematically rigorous approach but it does not consider the dependence between tuples in a database, which may degrade the privacy guarantees offered by the DP. Results In this work, focusing on genomic databases, we show this drawback of DP and we propose techniques to mitigate it. First, using a real-world genomic dataset, we demonstrate the feasibility of an inference attack on differentially private query results by utilizing the correlations between the tuples in the dataset. The results show that the adversary can infer sensitive genomic data about a user from the differentially private query results by exploiting correlations between genomes of family members. Second, we propose a mechanism for privacy-preserving sharing of statistics from genomic datasets to attain privacy guarantees while taking into consideration the dependence between tuples. By evaluating our mechanism on different genomic datasets, we empirically demonstrate that our proposed mechanism can achieve up to 50% better privacy than traditional DP-based solutions. Availability https://github.com/nourmadhoun/Differential-privacy-genomic-inference-attack. Supplementary information Supplementary data are available at Bioinformatics online.


Author(s):  
Dan Wang ◽  
Ju Ren ◽  
Zhibo Wang ◽  
Xiaoyi Pang ◽  
Yaoxue Zhang ◽  
...  

2021 ◽  
Vol 18 (11) ◽  
pp. 42-60
Author(s):  
Ting Bao ◽  
Lei Xu ◽  
Liehuang Zhu ◽  
Lihong Wang ◽  
Ruiguang Li ◽  
...  

Author(s):  
Shushu Liu ◽  
An Liu ◽  
Zhixu Li ◽  
Guanfeng Liu ◽  
Jiajie Xu ◽  
...  

2017 ◽  
Vol 20 (3) ◽  
pp. 887-895 ◽  
Author(s):  
Md Momin Al Aziz ◽  
Md Nazmus Sadat ◽  
Dima Alhadidi ◽  
Shuang Wang ◽  
Xiaoqian Jiang ◽  
...  

Author(s):  
Cheng Huang ◽  
Rongxing Lu ◽  
Hui Zhu ◽  
Jun Shao ◽  
Abdulrahman Alamer ◽  
...  

2021 ◽  
Author(s):  
Jude TCHAYE-KONDI ◽  
Yanlong Zhai ◽  
Liehuang Zhu

<div>We address privacy and latency issues in the edge/cloud computing environment while training a centralized AI model. In our particular case, the edge devices are the only data source for the model to train on the central server. Current privacy-preserving and reducing network latency solutions rely on a pre-trained feature extractor deployed on the devices to help extract only important features from the sensitive dataset. However, finding a pre-trained model or pubic dataset to build a feature extractor for certain tasks may turn out to be very challenging. With the large amount of data generated by edge devices, the edge environment does not really lack data, but its improper access may lead to privacy concerns. In this paper, we present DeepGuess , a new privacy-preserving, and latency aware deeplearning framework. DeepGuess uses a new learning mechanism enabled by the AutoEncoder(AE) architecture called Inductive Learning, which makes it possible to train a central neural network using the data produced by end-devices while preserving their privacy. With inductive learning, sensitive data remains on devices and is not explicitly involved in any backpropagation process. The AE’s Encoder is deployed on devices to extracts and transfers important features to the server. To enhance privacy, we propose a new local deferentially private algorithm that allows the Edge devices to apply random noise to features extracted from their sensitive data before transferred to an untrusted server. The experimental evaluation of DeepGuess demonstrates its effectiveness and ability to converge on a series of experiments.</div>


2019 ◽  
Vol 90 ◽  
pp. 158-174 ◽  
Author(s):  
Chunhui Piao ◽  
Yajuan Shi ◽  
Jiaqi Yan ◽  
Changyou Zhang ◽  
Liping Liu

Sign in / Sign up

Export Citation Format

Share Document