ann search
Recently Published Documents


TOTAL DOCUMENTS

10
(FIVE YEARS 3)

H-INDEX

3
(FIVE YEARS 1)

2021 ◽  
Vol 161 ◽  
pp. S1423-S1425
Author(s):  
S. Thulasi Seetha ◽  
K. Driessens ◽  
H. Woodruff ◽  
T. Rancati ◽  
E. Bertocchi ◽  
...  
Keyword(s):  

Information ◽  
2019 ◽  
Vol 10 (6) ◽  
pp. 190
Author(s):  
Xinpan Yuan ◽  
Qunfeng Liu ◽  
Jun Long ◽  
Lei Hu ◽  
Songlin Wang

Image retrieval or content-based image retrieval (CBIR) can be transformed into the calculation of the distance between image feature vectors. The closer the vectors are, the higher the image similarity will be. In the image retrieval system for large-scale dataset, the approximate nearest-neighbor (ANN) search can quickly obtain the top k images closest to the query image, which is the Top-k problem in the field of information retrieval. With the traditional ANN algorithms, such as KD-Tree, R-Tree, and M-Tree, when the dimension of the image feature vector increases, the computing time will increase exponentially due to the curse of dimensionality. In order to reduce the calculation time and improve the efficiency of image retrieval, we propose an ANN search algorithm based on the Product Quantization Table (PQTable). After quantizing and compressing the image feature vectors by the product quantization algorithm, we can construct the image index structure of the PQTable, which speeds up image retrieval. We also propose a multi-PQTable query strategy for ANN search. Besides, we generate several nearest-neighbor vectors for each sub-compressed vector of the query vector to reduce the failure rate and improve the recall in image retrieval. Through theoretical analysis and experimental verification, it is proved that the multi-PQTable query strategy and the generation of several nearest-neighbor vectors are greatly correct and efficient.


Author(s):  
Lifang Zhang ◽  
Qi Shen ◽  
Defang Li ◽  
Guocan Feng ◽  
Xin Tang ◽  
...  

Approximate Nearest Neighbor (ANN) search is a challenging problem with the explosive high-dimensional large-scale data in recent years. The promising technique for ANN search include hashing methods which generate compact binary codes by designing effective hash functions. However, lack of an optimal regularization is the key limitation of most of the existing hash functions. To this end, a new method called Adaptive Hashing with Sparse Modification (AHSM) is proposed. In AHSM, codes consist of vertices on the hypercube and the projection matrix is divided into two separate matrices. Data is rotated through a orthogonal matrix first and modified by a sparse matrix. Here the sparse matrix needs to be learned as a regularization item of hash function which is used to avoid overfitting and reduce quantization distortion. Totally, AHSM has two advantages: improvement of the accuracy without any time cost increasement. Furthermore, we extend AHSM to a supervised version, called Supervised Adaptive Hashing with Sparse Modification (SAHSM), by introducing Canonical Correlation Analysis (CCA) to the original data. Experiments show that the AHSM method stably surpasses several state-of-the-art hashing methods on four data sets. And at the same time, we compare three unsupervised hashing methods with their corresponding supervised version (including SAHSM) on three data sets with labels known. Similarly, SAHSM outperforms other methods on most of the hash bits.


2016 ◽  
Vol 171 ◽  
pp. 283-292 ◽  
Author(s):  
Jian Wang ◽  
Xin-Shun Xu ◽  
Shanqing Guo ◽  
Lizhen Cui ◽  
Xiao-Lin Wang
Keyword(s):  

2014 ◽  
Vol 651-653 ◽  
pp. 2224-2227
Author(s):  
Qin Zhen Guo ◽  
Zhi Zeng ◽  
Shu Wu Zhang

Product quantization (PQ) is an efficient and effective vector quantization approach to fast approximate nearest neighbor (ANN) search especially for high-dimensional data. The basic idea of PQ is to decompose the original data space into the Cartesian product of some low-dimensional subspaces and then every subspace is quantized separately with the same number of codewords. However, the performance of PQ depends largely on the distribution of the original data. If the distributions of every subspace have larger difference, PQ will achieve bad results as shown in our experiments. In this paper, we propose a uniform variance product quantization (UVPQ) scheme to project the data by a uniform variance projection before decompose it, which can minimize the subspace distribution difference of the whole space. UVPQ can guarantee good results however the data rotate. Extensive experiments have verified the superiority of UVPQ over PQ for ANN search.


2014 ◽  
Vol 21 (3) ◽  
pp. 41-51 ◽  
Author(s):  
Benchang Wei ◽  
Tao Guan ◽  
Junqing Yu

Sign in / Sign up

Export Citation Format

Share Document