Multimodal Medical Image Fusion Framework Based on Simplified PCNN in Nonsubsampled Contourlet Transform Domain

2013 ◽  
Vol 8 (3) ◽  
Author(s):  
Nianyi Wang ◽  
Yide Ma ◽  
Kun Zhan ◽  
Min Yuan
2013 ◽  
Vol 12 (4) ◽  
pp. 749-755 ◽  
Author(s):  
Shen Yu ◽  
Ren Enen ◽  
Dang Jian-Wu ◽  
Wang Guo-Hua ◽  
Feng Xin

2016 ◽  
Vol 07 (08) ◽  
pp. 1598-1610 ◽  
Author(s):  
Periyavattam Shanmugam Gomathi ◽  
Bhuvanesh Kalaavathi

2015 ◽  
Vol 2015 ◽  
pp. 1-13 ◽  
Author(s):  
Guocheng Yang ◽  
Meiling Li ◽  
Leiting Chen ◽  
Jie Yu

We propose a novel medical image fusion scheme based on the statistical dependencies between coefficients in the nonsubsampled contourlet transform (NSCT) domain, in which the probability density function of the NSCT coefficients is concisely fitted using generalized Gaussian density (GGD), as well as the similarity measurement of two subbands is accurately computed by Jensen-Shannon divergence of two GGDs. To preserve more useful information from source images, the new fusion rules are developed to combine the subbands with the varied frequencies. That is, the low frequency subbands are fused by utilizing two activity measures based on the regional standard deviation and Shannon entropy and the high frequency subbands are merged together via weight maps which are determined by the saliency values of pixels. The experimental results demonstrate that the proposed method significantly outperforms the conventional NSCT based medical image fusion approaches in both visual perception and evaluation indices.


Sign in / Sign up

Export Citation Format

Share Document