scholarly journals Content-Based Image Retrieval Based on Late Fusion of Binary and Local Descriptors

Author(s):  
Nouman Ali ◽  
Danish Ali Mazhar ◽  
Zeshan Iqbal ◽  
Rehan Ashraf ◽  
Jawad Ahmed ◽  
...  

One of the challenges in Content-Based Image Retrieval (CBIR) is to reduce the semantic gaps between low-level features and high-level semantic concepts. In CBIR, the images are represented in the feature space and the performance of CBIR depends on the type of selected feature representation. Late fusion also known as visual words integration is applied to enhance the performance of image retrieval. The recent advances in image retrieval diverted the focus of research towards the use of binary descriptors as they are reported computationally efficient. In this paper, we aim to investigate the late fusion of Fast Retina Keypoint (FREAK) and Scale Invariant Feature Transform (SIFT). The late fusion of binary and local descriptor is selected because among binary descriptors, FREAK has shown good results in classification-based problems while SIFT is robust to translation, scaling, rotation and small distortions. The late fusion of FREAK and SIFT integrates the performance of both feature descriptors for an effective image retrieval. Experimental results and comparisons show that the proposed late fusion enhances the performances of image retrieval.

Author(s):  
Nouman Ali ◽  
Danish Ali Mazhar ◽  
Zeshan Iqbal ◽  
Rehan Ashraf ◽  
Jawad Ahmed ◽  
...  

One of the challenges in Content-Based Image Retrieval (CBIR) is to reduce the semantic gaps between low-level features and high-level semantic concepts. In CBIR, the images are represented in the feature space and the performance of CBIR depends on the type of selected feature representation. Late fusion also known as visual words integration is applied to enhance the performance of image retrieval. The recent advances in image retrieval diverted the focus of research towards the use of binary descriptors as they are reported computationally efficient. In this paper, we aim to investigate the late fusion of Fast Retina Keypoint (FREAK) and Scale Invariant Feature Transform (SIFT). The late fusion of binary and local descriptor is selected because among binary descriptors, FREAK has shown good results in classification-based problems while SIFT is robust to translation, scaling, rotation and small distortions. The late fusion of FREAK and SIFT integrates the performance of both feature descriptors for an effective image retrieval. Experimental results and comparisons show that the proposed late fusion enhances the performances of image retrieval.


2018 ◽  
Vol 2018 ◽  
pp. 1-13 ◽  
Author(s):  
Muhammad Yousuf ◽  
Zahid Mehmood ◽  
Hafiz Adnan Habib ◽  
Toqeer Mahmood ◽  
Tanzila Saba ◽  
...  

Content-based image retrieval (CBIR) is a mechanism that is used to retrieve similar images from an image collection. In this paper, an effective novel technique is introduced to improve the performance of CBIR on the basis of visual words fusion of scale-invariant feature transform (SIFT) and local intensity order pattern (LIOP) descriptors. SIFT performs better on scale changes and on invariant rotations. However, SIFT does not perform better in the case of low contrast and illumination changes within an image, while LIOP performs better in such circumstances. SIFT performs better even at large rotation and scale changes, while LIOP does not perform well in such circumstances. Moreover, SIFT features are invariant to slight distortion as compared to LIOP. The proposed technique is based on the visual words fusion of SIFT and LIOP descriptors which overcomes the aforementioned issues and significantly improves the performance of CBIR. The experimental results of the proposed technique are compared with another proposed novel features fusion technique based on SIFT-LIOP descriptors as well as with the state-of-the-art CBIR techniques. The qualitative and quantitative analysis carried out on three image collections, namely, Corel-A, Corel-B, and Caltech-256, demonstrate the robustness of the proposed technique based on visual words fusion as compared to features fusion and the state-of-the-art CBIR techniques.


2021 ◽  
Vol 8 (7) ◽  
pp. 97-105
Author(s):  
Ali Ahmed ◽  
◽  
Sara Mohamed ◽  

Content-Based Image Retrieval (CBIR) systems retrieve images from the image repository or database in which they are visually similar to the query image. CBIR plays an important role in various fields such as medical diagnosis, crime prevention, web-based searching, and architecture. CBIR consists mainly of two stages: The first is the extraction of features and the second is the matching of similarities. There are several ways to improve the efficiency and performance of CBIR, such as segmentation, relevance feedback, expansion of queries, and fusion-based methods. The literature has suggested several methods for combining and fusing various image descriptors. In general, fusion strategies are typically divided into two groups, namely early and late fusion strategies. Early fusion is the combination of image features from more than one descriptor into a single vector before the similarity computation, while late fusion refers either to the combination of outputs produced by various retrieval systems or to the combination of different rankings of similarity. In this study, a group of color and texture features is proposed to be used for both methods of fusion strategies. Firstly, an early combination of eighteen color features and twelve texture features are combined into a single vector representation and secondly, the late fusion of three of the most common distance measures are used in the late fusion stage. Our experimental results on two common image datasets show that our proposed method has good performance retrieval results compared to the traditional way of using single features descriptor and also has an acceptable retrieval performance compared to some of the state-of-the-art methods. The overall accuracy of our proposed method is 60.6% and 39.07% for Corel-1K and GHIM-10K ‎datasets, respectively.


2018 ◽  
Vol 45 (1) ◽  
pp. 117-135 ◽  
Author(s):  
Amna Sarwar ◽  
Zahid Mehmood ◽  
Tanzila Saba ◽  
Khurram Ashfaq Qazi ◽  
Ahmed Adnan ◽  
...  

The advancements in the multimedia technologies result in the growth of the image databases. To retrieve images from such image databases using visual attributes of the images is a challenging task due to the close visual appearance among the visual attributes of these images, which also introduces the issue of the semantic gap. In this article, we recommend a novel method established on the bag-of-words (BoW) model, which perform visual words integration of the local intensity order pattern (LIOP) feature and local binary pattern variance (LBPV) feature to reduce the issue of the semantic gap and enhance the performance of the content-based image retrieval (CBIR). The recommended method uses LIOP and LBPV features to build two smaller size visual vocabularies (one from each feature), which are integrated together to build a larger size of the visual vocabulary, which also contains complementary features of both descriptors. Because for efficient CBIR, the smaller size of the visual vocabulary improves the recall, while the bigger size of the visual vocabulary improves the precision or accuracy of the CBIR. The comparative analysis of the recommended method is performed on three image databases, namely, WANG-1K, WANG-1.5K and Holidays. The experimental analysis of the recommended method on these image databases proves its robust performance as compared with the recent CBIR methods.


2012 ◽  
Vol 12 (3) ◽  
pp. 7-19 ◽  
Author(s):  
Letricia P. S. Avalhais ◽  
Sergio F. da Silva ◽  
Jose F. Rodrigues ◽  
Agma J. M. Traina ◽  
Caetano Traina

Author(s):  
HWEI-JEN LIN ◽  
YANG-TA KAO ◽  
FU-WEN YANG ◽  
PATRICK S. P. WANG

This paper proposes a Content-Based Image Retrieval (CBIR) system applicable in mobile devices. Due to the fact that different queries to a content-based image retrieval (CBIR) system emphasize different subsets of a large collection of features, most CBIR systems using only a few features are therefore only suitable for retrieving certain types of images. In this research we combine a wide range of features, including edge information, texture energy, and the HSV color distributions, forming a feature space of up to 1053 dimensions, in which the system can search for features most desired by the user. Through a training process using the AdaBoost algorithm9 our system can efficiently search for important features in a large set of features, as indicated by the user, and effectively retrieve the images according to these features. The characteristics of the system meet the requirements of mobile devices for performing image retrieval. The experimental results show that the performance of the proposed system is sufficiently applicable for mobile devices to retrieve images from a huge database.


Author(s):  
TIENWEI TSAI ◽  
YO-PING HUANG ◽  
TE-WEI CHIANG

In this paper, a two-stage content-based image retrieval (CBIR) approach is proposed to improve the retrieval performance. To develop a general retrieval scheme which is less dependent on domain-specific knowledge, the discrete cosine transform (DCT) is employed as a feature extraction method. In establishing the database, the DC coefficients of Y, U and V components are quantized such that the feature space is partitioned into a finite number of grids, each of which is mapped to a grid code (GC). When querying an image, at coarse classification stage, the grid-based classification (GBC) and the distance threshold pruning (DTP) serve as a filter to remove those candidates with widely distinct features. At the fine classification stage, only the remaining candidates need to be computed for the detailed similarity comparison. The experimental results show that both high efficacy and high efficiency can be achieved simultaneously using the proposed two-stage approach.


2014 ◽  
Vol 596 ◽  
pp. 388-393
Author(s):  
Guan Huang

This paper introduces a model for content based image retrieval. The proposed model extracts image color, texture and shape as feature vectors; and then the image feature space is divided into a group of search zones; during the image searching phase, the fractional order distance is utilized to evaluate the similarity between images. As the query image vector only needs to compare with library image vectors located in the same search zone, the time cost is largely reduced. Further more the fractional order distance is utilized to improve the vector matching accuracy. The experimental results demonstrated that the proposed model provides more accurate retrieval results with less time cost compared with other methods.


Sign in / Sign up

Export Citation Format

Share Document