scholarly journals Product Graph Learning from Multi-domain Data with Sparsity and Rank Constraints

Author(s):  
Sai Kiran Kadambari ◽  
Sundeep Prabhakar Chepuri
Author(s):  
Mei Sun ◽  
Jinxu Tao ◽  
Zhongfu Ye ◽  
Bensheng Qiu ◽  
Jinzhang Xu ◽  
...  

Background: In order to overcome the limitation of long scanning time, compressive sensing (CS) technology exploits the sparsity of image in some transform domain to reduce the amount of acquired data. Therefore, CS has been widely used in magnetic resonance imaging (MRI) reconstruction. </P><P> Discussion: Blind compressed sensing enables to recover the image successfully from highly under- sampled measurements, because of the data-driven adaption of the unknown transform basis priori. Moreover, analysis-based blind compressed sensing often leads to more efficient signal reconstruction with less time than synthesis-based blind compressed sensing. Recently, some experiments have shown that nonlocal low-rank property has the ability to preserve the details of the image for MRI reconstruction. Methods: Here, we focus on analysis-based blind compressed sensing, and combine it with additional nonlocal low-rank constraint to achieve better MR images from fewer measurements. Instead of nuclear norm, we exploit non-convex Schatten p-functionals for the rank approximation. </P><P> Results & Conclusion: Simulation results indicate that the proposed approach performs better than the previous state-of-the-art algorithms.


2020 ◽  
Vol 387 ◽  
pp. 110-122 ◽  
Author(s):  
Xiangpin Bai ◽  
Lei Zhu ◽  
Cheng Liang ◽  
Jingjing Li ◽  
Xiushan Nie ◽  
...  

2021 ◽  
pp. 108101
Author(s):  
Seyed Saman Saboksayr ◽  
Gonzalo Mateos ◽  
Mujdat Cetin
Keyword(s):  

Author(s):  
Yun Peng ◽  
Byron Choi ◽  
Jianliang Xu

AbstractGraphs have been widely used to represent complex data in many applications, such as e-commerce, social networks, and bioinformatics. Efficient and effective analysis of graph data is important for graph-based applications. However, most graph analysis tasks are combinatorial optimization (CO) problems, which are NP-hard. Recent studies have focused a lot on the potential of using machine learning (ML) to solve graph-based CO problems. Most recent methods follow the two-stage framework. The first stage is graph representation learning, which embeds the graphs into low-dimension vectors. The second stage uses machine learning to solve the CO problems using the embeddings of the graphs learned in the first stage. The works for the first stage can be classified into two categories, graph embedding methods and end-to-end learning methods. For graph embedding methods, the learning of the the embeddings of the graphs has its own objective, which may not rely on the CO problems to be solved. The CO problems are solved by independent downstream tasks. For end-to-end learning methods, the learning of the embeddings of the graphs does not have its own objective and is an intermediate step of the learning procedure of solving the CO problems. The works for the second stage can also be classified into two categories, non-autoregressive methods and autoregressive methods. Non-autoregressive methods predict a solution for a CO problem in one shot. A non-autoregressive method predicts a matrix that denotes the probability of each node/edge being a part of a solution of the CO problem. The solution can be computed from the matrix using search heuristics such as beam search. Autoregressive methods iteratively extend a partial solution step by step. At each step, an autoregressive method predicts a node/edge conditioned to current partial solution, which is used to its extension. In this survey, we provide a thorough overview of recent studies of the graph learning-based CO methods. The survey ends with several remarks on future research directions.


Author(s):  
Yongyong Chen ◽  
Xiaolin Xiao ◽  
Chong Peng ◽  
Guangming Lu ◽  
Yicong Zhou

2021 ◽  
Vol 1757 (1) ◽  
pp. 012001
Author(s):  
Ni Li ◽  
Manman Peng ◽  
Buwen Cao ◽  
Kenli Li ◽  
Keqin Li

Author(s):  
Guang-Yu Zhang ◽  
Xiao-Wei Chen ◽  
Yu-Ren Zhou ◽  
Chang-Dong Wang ◽  
Dong Huang ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document