scholarly journals MOON: Multi-Hash Codes Joint Learning for Cross-Media Retrieval

Author(s):  
Donglin Zhang ◽  
Xiao-Jun Wu ◽  
He-Feng Yin ◽  
Josef Kittler
2021 ◽  
Author(s):  
Mingrui Chen ◽  
Weiyu Li ◽  
weizhi lu

Recently, it has been observed that $\{0,\pm1\}$-ternary codes which are simply generated from deep features by hard thresholding, tend to outperform $\{-1, 1\}$-binary codes in image retrieval. To obtain better ternary codes, we for the first time propose to jointly learn the features with the codes by appending a smoothed function to the networks. During training, the function could evolve into a non-smoothed ternary function by a continuation method, and then generate ternary codes. The method circumvents the difficulty of directly training discrete functions and reduces the quantization errors of ternary codes. Experiments show that the proposed joint learning indeed could produce better ternary codes.


2021 ◽  
Author(s):  
Mingrui Chen ◽  
Weiyu Li ◽  
weizhi lu

Recently, it has been observed that $\{0,\pm1\}$-ternary codes which are simply generated from deep features by hard thresholding, tend to outperform $\{-1, 1\}$-binary codes in image retrieval. To obtain better ternary codes, we for the first time propose to jointly learn the features with the codes by appending a smoothed function to the networks. During training, the function could evolve into a non-smoothed ternary function by a continuation method, and then generate ternary codes. The method circumvents the difficulty of directly training discrete functions and reduces the quantization errors of ternary codes. Experiments show that the proposed joint learning indeed could produce better ternary codes.


Author(s):  
Donglin Zhang ◽  
Xiao-Jun Wu ◽  
Jun Yu

Hashing methods have sparked a great revolution on large-scale cross-media search due to its effectiveness and efficiency. Most existing approaches learn unified hash representation in a common Hamming space to represent all multimodal data. However, the unified hash codes may not characterize the cross-modal data discriminatively, because the data may vary greatly due to its different dimensionalities, physical properties, and statistical information. In addition, most existing supervised cross-modal algorithms preserve the similarity relationship by constructing an n × n pairwise similarity matrix, which requires a large amount of calculation and loses the category information. To mitigate these issues, a novel cross-media hashing approach is proposed in this article, dubbed label flexible matrix factorization hashing (LFMH). Specifically, LFMH jointly learns the modality-specific latent subspace with similar semantic by the flexible matrix factorization. In addition, LFMH guides the hash learning by utilizing the semantic labels directly instead of the large n × n pairwise similarity matrix. LFMH transforms the heterogeneous data into modality-specific latent semantic representation. Therefore, we can obtain the hash codes by quantifying the representations, and the learned hash codes are consistent with the supervised labels of multimodal data. Then, we can obtain the similar binary codes of the corresponding modality, and the binary codes can characterize such samples flexibly. Accordingly, the derived hash codes have more discriminative power for single-modal and cross-modal retrieval tasks. Extensive experiments on eight different databases demonstrate that our model outperforms some competitive approaches.


Neuróptica ◽  
2020 ◽  
pp. 249-252
Author(s):  
Julia Rigual Mur
Keyword(s):  

Reseña del libro: HERNÁNDEZ PÉREZ, M., Manga, anime y videojuegos. Narrativa cross-media japonesa, Zaragoza, Prensas Universitarias de Zaragoza, 2017.


2009 ◽  
Vol 31 (5) ◽  
pp. 820-826 ◽  
Author(s):  
Hong ZHANG ◽  
Fei WU ◽  
Yue-Ting ZHUANG ◽  
Jian-Xun CHEN
Keyword(s):  

Author(s):  
Atsushi Ando ◽  
Ryo Masumura ◽  
Hosana Kamiyama ◽  
Satoshi Kobashikawa ◽  
Yushi Aono

Sign in / Sign up

Export Citation Format

Share Document