image query
Recently Published Documents


TOTAL DOCUMENTS

91
(FIVE YEARS 13)

H-INDEX

8
(FIVE YEARS 2)

Author(s):  
Sheung Wai Chan ◽  
Yiu-Ming Cheung

The existing image retrieval methods generally require at least one complete image as a query sample. From the practical point of view, a user may not have an image sample in hand for query. Instead, partial information from multiple image samples would be available. This paper therefore attempts to deal with this problem by presenting a novel framework that allows a user to make an image query composed of several partial information extracted from multiple image samples via Boolean operations (i.e. AND, OR and NOT). Based on the request from the query, a Descriptor Cluster Label Table (DCLT) is designed to efficiently find out the result of Boolean operations on partial information. Experiments show the promising result of the proposed framework on commodity query and criminal investigation, respectively, although it is essentially applicable to different scenarios as well by changing descriptors.


Author(s):  
Shaily Malik ◽  
Poonam Bansal

The real-world data is multimodal and to classify them by machine learning algorithms, features of both modalities must be transformed into common latent space. The high dimensional common space transformation of features lose their locality information and susceptible to noise. This research article has dealt with this issue of a semantic autoencoder and presents a novel algorithm with distinct mapped features with locality preservation into a commonly hidden space. We call it discriminative regularized semantic autoencoder (DRSAE). It maintains the low dimensional features in the manifold to manage the inter and intra-modality of the data. The data has multi labels, and these are transformed into an aware feature space. Conditional Principal label space transformation (CPLST) is used for it. With the two-fold proposed algorithm, we achieve a significant improvement in text retrieval form image query and image retrieval from the text query.


2020 ◽  
Vol 97 ◽  
pp. 106782
Author(s):  
Hongyang Yan ◽  
Mengqi Chen ◽  
Li Hu ◽  
Chunfu Jia
Keyword(s):  

2020 ◽  
Author(s):  
Mahmoud Elmezain ◽  
Hani M Ibrahem

Abstract This paper introduces a new approach to semantic image retrieval using shape descriptors as dispersion and moment in conjunction with discriminative classifier model of latent-dynamic conditional random fields (LDCRFs). The target region is firstly localized via the background subtraction model. Then the features of dispersion and moments are employed to k-means clustering to extract object’s feature as second stage. After that, the learning process is carried out by LDCRFs. Finally, simple protocol and RDF (resource description framework) query language (i.e. SPARQL) on input text or image query is to retrieve semantic image based on sequential processes of query engine, matching module and ontology manager. Experimental findings show that our approach can be successful to retrieve images against the mammal’s benchmark with retrieving rate of 98.11%. Such outcomes are likely to compare very positively with those accessible in the literature from other researchers.


Author(s):  
Kristina Hanspers ◽  
Anders Riutta ◽  
Martina Kutmon ◽  
Alexander R Pico

Background: Pathway diagrams are fundamental tools for describing biological processes in all aspects of science, including training, generating hypotheses, describing new knowledge and ultimately as communication tools in published work. Thousands of pathway diagrams are published each year as figures in papers. But as static images the pathway knowledge represented in figures is not accessible to researchers for computational queries and analyses. In this study, we aimed to identify pathway figures published in the past 25 years, to characterize the human gene content in figures by optical character recognition, and to describe their utility as a resource for pathway knowledge. Approach: To identify pathway figures representing 25 years of published research, we trained a machine learning service on manually-classified figures and applied it to 235,081 image query results from PubMed Central. Our previously described pipeline was utilized to extract human genes from the pathway figure images. These figures were characterized in terms of their parent papers, human gene content and enriched disease terms. Diverse use cases were explored for this newly accessible pathway resource. Results: We identified 64,643 pathway figures published between 1995 and 2019, depicting 1,112,551 instances of human genes (13,464 unique NCBI Genes) in various interactions and contexts. This represents more genes than found in the text of the same papers, as well as genes not found in any pathway database. We developed an interactive web tool to explore the results from the 65k set of figures, and used this tool to explore the history of scientific discovery of the Hippo Signaling pathway. We also defined a filtered set of 32k pathway figures useful for enrichment analysis.


2020 ◽  
Vol 2020 ◽  
pp. 1-17
Author(s):  
Guanhua Wang ◽  
Hua Ji ◽  
Dexin Kong ◽  
Na Zhang

Nowadays, the heterogeneity gap of different modalities is the key problem for cross-modal retrieval. In order to overcome heterogeneity gaps, potential correlations of different modalities need to be mined. At the same time, the semantic information of class labels is used to reduce the semantic gaps between different modalities data and realize the interdependence and interoperability of heterogeneous data. In order to fully exploit the potential correlation of different modalities, we propose a cross-modal retrieval framework based on graph regularization and modality dependence (GRMD). Firstly, considering the potential feature correlation and semantic correlation, different projection matrices are learned for different retrieval tasks, such as image query text (I2T) or text query image (T2I). Secondly, utilizing the internal structure of original feature space constructs an adjacent graph with semantic information constraints which can make different labels of heterogeneous data closer to the corresponding semantic information. The experimental results on three widely used datasets demonstrate the effectiveness of our method.


2019 ◽  
Vol 12 (1) ◽  
pp. 84 ◽  
Author(s):  
Yaxiong Chen ◽  
Xiaoqiang Lu

With the rapid progress of remote sensing (RS) observation technologies, cross-modal RS image-sound retrieval has attracted some attention in recent years. However, these methods perform cross-modal image-sound retrieval by leveraging high-dimensional real-valued features, which can require more storage than low-dimensional binary features (i.e., hash codes). Moreover, these methods cannot directly encode relative semantic similarity relationships. To tackle these issues, we propose a new, deep, cross-modal RS image-sound hashing approach, called deep triplet-based hashing (DTBH), to integrate hash code learning and relative semantic similarity relationship learning into an end-to-end network. Specially, the proposed DTBH method designs a triplet selection strategy to select effective triplets. Moreover, in order to encode relative semantic similarity relationships, we propose the objective function, which makes sure that that the anchor images are more similar to the positive sounds than the negative sounds. In addition, a triplet regularized loss term leverages approximate l1-norm of hash-like codes and hash codes and can effectively reduce the information loss between hash-like codes and hash codes. Extensive experimental results showed that the DTBH method could achieve a superior performance to other state-of-the-art cross-modal image-sound retrieval methods. For a sound query RS image task, the proposed approach achieved a mean average precision (mAP) of up to 60.13% on the UCM dataset, 87.49% on the Sydney dataset, and 22.72% on the RSICD dataset. For RS image query sound task, the proposed approach achieved a mAP of 64.27% on the UCM dataset, 92.45% on the Sydney dataset, and 23.46% on the RSICD dataset. Future work will focus on how to consider the balance property of hash codes to improve image-sound retrieval performance.


Sign in / Sign up

Export Citation Format

Share Document