The sparse image representation for automated image retrieval

Author(s):  
Pavel Praks ◽  
Radek Kucera ◽  
Ebroul Izquierdo
Author(s):  
Noureddine Abbadeni

This chapter describes an approach based on human perception to content-based image representation and retrieval. We consider textured images and propose to model the textural content of images by a set of features having a perceptual meaning and their application to content-based image retrieval. We present a new method to estimate a set of perceptual textural features, namely coarseness, directionality, contrast and busyness. The proposed computational measures are based on two representations: the original images representation and the autocovariance function (associated with images) representation. The correspondence of the proposed computational measures to human judgments is shown using a psychometric method based on the Spearman rank-correlation coefficient. The set of computational measures is applied to content-based image retrieval on a large image data set, the well-known Brodatz database. Experimental results show a strong correlation between the proposed computational textural measures and human perceptual judgments. The benchmarking of retrieval performance, done using the recall measure, shows interesting results. Furthermore, results merging/fusion returned by each of the two representations is shown to allow significant improvement in retrieval effectiveness.


2018 ◽  
Vol 70 (1) ◽  
pp. 47-65 ◽  
Author(s):  
Wei Lu ◽  
Heng Ding ◽  
Jiepu Jiang

Purpose The purpose of this paper is to utilize document expansion techniques for improving image representation and retrieval. This paper proposes a concise framework for tag-based image retrieval (TBIR). Design/methodology/approach The proposed approach includes three core components: a strategy of selecting expansion (similar) images from the whole corpus (e.g. cluster-based or nearest neighbor-based); a technique for assessing image similarity, which is adopted for selecting expansion images (text, image, or mixed); and a model for matching the expanded image representation with the search query (merging or separate). Findings The results show that applying the proposed method yields significant improvements in effectiveness, and the method obtains better performance on the top of the rank and makes a great improvement on some topics with zero score in baseline. Moreover, nearest neighbor-based expansion strategy outperforms the cluster-based expansion strategy, and using image features for selecting expansion images is better than using text features in most cases, and the separate method for calculating the augmented probability P(q|RD) is able to erase the negative influences of error images in RD. Research limitations/implications Despite these methods only outperform on the top of the rank instead of the entire rank list, TBIR on mobile platforms still can benefit from this approach. Originality/value Unlike former studies addressing the sparsity, vocabulary mismatch, and tag relatedness in TBIR individually, the approach proposed by this paper addresses all these issues with a single document expansion framework. It is a comprehensive investigation of document expansion techniques in TBIR.


2003 ◽  
Vol 03 (01) ◽  
pp. 119-143 ◽  
Author(s):  
ZHIYONG WANG ◽  
ZHERU CHI ◽  
DAGAN FENG ◽  
AH CHUNG TSOI

Content-based image retrieval has become an essential technique in multimedia data management. However, due to the difficulties and complications involved in the various image processing tasks, a robust semantic representation of image content is still very difficult (if not impossible) to achieve. In this paper, we propose a novel content-based image retrieval approach with relevance feedback using adaptive processing of tree-structure image representation. In our approach, each image is first represented with a quad-tree, which is segmentation free. Then a neural network model with the Back-Propagation Through Structure (BPTS) learning algorithm is employed to learn the tree-structure representation of the image content. This approach that integrates image representation and similarity measure in a single framework is applied to the relevance feedback of the content-based image retrieval. In our approach, an initial ranking of the database images is first carried out based on the similarity between the query image and each of the database images according to global features. The user is then asked to categorize the top retrieved images into similar and dissimilar groups. Finally, the BPTS neural network model is used to learn the user's intention for a better retrieval result. This process continues until satisfactory retrieval results are achieved. In the refining process, a fine similarity grading scheme can also be adopted to improve the retrieval performance. Simulations on texture images and scenery pictures have demonstrated promising results which compare favorably with the other relevance feedback methods tested.


2016 ◽  
Vol 366 ◽  
pp. 99-120 ◽  
Author(s):  
Nguyen Anh Tu ◽  
Dong-Luong Dinh ◽  
Mostofa Kamal Rasel ◽  
Young-Koo Lee

2015 ◽  
Vol 734 ◽  
pp. 596-599 ◽  
Author(s):  
Deng Ping Fan ◽  
Juan Wang ◽  
Xue Mei Liang

The Context-Aware Saliency (CA) model—is a new model used for saliency detection—has strong limitations: It is very time consuming. This paper improved the shortcoming of this model namely Fast-CA and proposed a novel framework for image retrieval and image representation. The proposed framework derives from Fast-CA and multi-texton histogram. And the mechanisms of visual attention are simulated and used to detect saliency areas of an image. Furthermore, a very simple threshold method is adopted to detect the dominant saliency areas. Color, texture and edge features are further extracted to describe image content within the dominant saliency areas, and then those features are integrated into one entity as image representation, where image representation is so called the dominant saliency areas histogram (DSAH) and used for image retrieval. Experimental results indicate that our algorithm outperform multi-texton histogram (MTH) and edge histogram descriptors (EHD) on Corel dataset with 10000 natural images.


Author(s):  
Wei-Bang Chen ◽  
Chengcui Zhang

Inaccurate image segmentation often has a negative impact on object-based image retrieval. Researchers have attempted to alleviate this problem by using hierarchical image representation. However, these attempts suffer from the inefficiency in building the hierarchical image representation and the high computational complexity in matching two hierarchically represented images. This paper presents an innovative multiple-object retrieval framework named Multiple-Object Image Retrieval (MOIR) on the basis of hierarchical image representation. This framework concurrently performs image segmentation and hierarchical tree construction, producing a hierarchical region tree to represent the image. In addition, an efficient hierarchical region tree matching algorithm is designed for multiple-object retrieval with a reasonably low time complexity. The experimental results demonstrate the efficacy and efficiency of the proposed approach.


2019 ◽  
Vol 2019 ◽  
pp. 1-21 ◽  
Author(s):  
Afshan Latif ◽  
Aqsa Rasheed ◽  
Umer Sajid ◽  
Jameel Ahmed ◽  
Nouman Ali ◽  
...  

Multimedia content analysis is applied in different real-world computer vision applications, and digital images constitute a major part of multimedia data. In last few years, the complexity of multimedia contents, especially the images, has grown exponentially, and on daily basis, more than millions of images are uploaded at different archives such as Twitter, Facebook, and Instagram. To search for a relevant image from an archive is a challenging research problem for computer vision research community. Most of the search engines retrieve images on the basis of traditional text-based approaches that rely on captions and metadata. In the last two decades, extensive research is reported for content-based image retrieval (CBIR), image classification, and analysis. In CBIR and image classification-based models, high-level image visuals are represented in the form of feature vectors that consists of numerical values. The research shows that there is a significant gap between image feature representation and human visual understanding. Due to this reason, the research presented in this area is focused to reduce the semantic gap between the image feature representation and human visual understanding. In this paper, we aim to present a comprehensive review of the recent development in the area of CBIR and image representation. We analyzed the main aspects of various image retrieval and image representation models from low-level feature extraction to recent semantic deep-learning approaches. The important concepts and major research studies based on CBIR and image representation are discussed in detail, and future research directions are concluded to inspire further research in this area.


Sign in / Sign up

Export Citation Format

Share Document