Probability Ordinal-Preserving Semantic Hashing for Large-Scale Image Retrieval

2021 ◽  
Vol 15 (3) ◽  
pp. 1-22
Author(s):  
Zheng Zhang ◽  
Xiaofeng Zhu ◽  
Guangming Lu ◽  
Yudong Zhang

Semantic hashing enables computation and memory-efficient image retrieval through learning similarity-preserving binary representations. Most existing hashing methods mainly focus on preserving the piecewise class information or pairwise correlations of samples into the learned binary codes while failing to capture the mutual triplet-level ordinal structure in similarity preservation. In this article, we propose a novel Probability Ordinal-preserving Semantic Hashing (POSH) framework, which for the first time defines the ordinal-preserving hashing concept under a non-parametric Bayesian theory. Specifically, we derive the whole learning framework of the ordinal similarity-preserving hashing based on the maximum posteriori estimation, where the probabilistic ordinal similarity preservation, probabilistic quantization function, and probabilistic semantic-preserving function are jointly considered into one unified learning framework. In particular, the proposed triplet-ordering correlation preservation scheme can effectively improve the interpretation of the learned hash codes under an economical anchor-induced asymmetric graph learning model. Moreover, the sparsity-guided selective quantization function is designed to minimize the loss of space transformation, and the regressive semantic function is explored to promote the flexibility of the formulated semantics in hash code learning. The final joint learning objective is formulated to concurrently preserve the ordinal locality of original data and explore potentials of semantics for producing discriminative hash codes. Importantly, an efficient alternating optimization algorithm with the strictly proof convergence guarantee is developed to solve the resulting objective problem. Extensive experiments on several large-scale datasets validate the superiority of the proposed method against state-of-the-art hashing-based retrieval methods.

2014 ◽  
Vol 651-653 ◽  
pp. 2197-2200
Author(s):  
Qin Zhen Guo ◽  
Zhi Zeng ◽  
Shu Wu Zhang ◽  
Xiao Feng ◽  
Hu Guan

Due to its fast query speed and reduced storage cost, hashing, which tries to learn binary code representation for data with the expectation of preserving the neighborhood structure in the original data space, has been widely used in a large variety of applications like image retrieval. For most existing image retrieval methods with hashing, there are two main steps: describe images with feature vectors, and then use hashing methods to encode the feature vectors. In this paper, we make two research contributions. First, we creatively propose to use simhash which can be intrinsically combined with the popular image representation method, Bag-of-visual-words (BoW) for image retrieval. Second, we novelly incorporate “locality-sensitive” hashing into simhash to take the correlation of the visual words of BoW into consideration to make similar visual words have similar fingerprint. Extensive experiments have verified the superiority of our method over some state-of-the-art methods for image retrieval task.


2021 ◽  
Author(s):  
Mingrui Chen ◽  
Weiyu Li ◽  
weizhi lu

Recently, it has been observed that $\{0,\pm1\}$-ternary codes which are simply generated from deep features by hard thresholding, tend to outperform $\{-1, 1\}$-binary codes in image retrieval. To obtain better ternary codes, we for the first time propose to jointly learn the features with the codes by appending a smoothed function to the networks. During training, the function could evolve into a non-smoothed ternary function by a continuation method, and then generate ternary codes. The method circumvents the difficulty of directly training discrete functions and reduces the quantization errors of ternary codes. Experiments show that the proposed joint learning indeed could produce better ternary codes.


2021 ◽  
Author(s):  
Mingrui Chen ◽  
Weiyu Li ◽  
weizhi lu

Recently, it has been observed that $\{0,\pm1\}$-ternary codes which are simply generated from deep features by hard thresholding, tend to outperform $\{-1, 1\}$-binary codes in image retrieval. To obtain better ternary codes, we for the first time propose to jointly learn the features with the codes by appending a smoothed function to the networks. During training, the function could evolve into a non-smoothed ternary function by a continuation method, and then generate ternary codes. The method circumvents the difficulty of directly training discrete functions and reduces the quantization errors of ternary codes. Experiments show that the proposed joint learning indeed could produce better ternary codes.


Sign in / Sign up

Export Citation Format

Share Document