scholarly journals 3D Radon Transform for Shape Retrieval Using Bag-of-Visual-Features

2019 ◽  
Vol 17 (4) ◽  
pp. 471-479
Author(s):  
Jinlin Ma ◽  
Ziping Ma

In order to improve the accuracy and efficiency of extracting features for 3D models retrieval, a novel approach using 3D radon transform and Bag-of-Visual-Features is proposed in this paper. Firstly the 3D radon transform is employed to obtain a view image using the different features in different angels. Then a set of local descriptor vectors are extracted by the SURF algorithm from the local features of the view. The similarity distance between geometrical transformed models is evaluated by using K-means algorithm to verify the geometric invariance of the proposed method. The numerical experiments are conducted to evaluate the retrieval efficiency compared to other typical methods. The experimental results show that the change of parameters has small effect on the retrieval performance of the proposed method

2018 ◽  
Vol 11 (8) ◽  
pp. 4627-4643 ◽  
Author(s):  
Simon Pfreundschuh ◽  
Patrick Eriksson ◽  
David Duncan ◽  
Bengt Rydberg ◽  
Nina Håkansson ◽  
...  

Abstract. A neural-network-based method, quantile regression neural networks (QRNNs), is proposed as a novel approach to estimating the a posteriori distribution of Bayesian remote sensing retrievals. The advantage of QRNNs over conventional neural network retrievals is that they learn to predict not only a single retrieval value but also the associated, case-specific uncertainties. In this study, the retrieval performance of QRNNs is characterized and compared to that of other state-of-the-art retrieval methods. A synthetic retrieval scenario is presented and used as a validation case for the application of QRNNs to Bayesian retrieval problems. The QRNN retrieval performance is evaluated against Markov chain Monte Carlo simulation and another Bayesian method based on Monte Carlo integration over a retrieval database. The scenario is also used to investigate how different hyperparameter configurations and training set sizes affect the retrieval performance. In the second part of the study, QRNNs are applied to the retrieval of cloud top pressure from observations by the Moderate Resolution Imaging Spectroradiometer (MODIS). It is shown that QRNNs are not only capable of achieving similar accuracy to standard neural network retrievals but also provide statistically consistent uncertainty estimates for non-Gaussian retrieval errors. The results presented in this work show that QRNNs are able to combine the flexibility and computational efficiency of the machine learning approach with the theoretically sound handling of uncertainties of the Bayesian framework. Together with this article, a Python implementation of QRNNs is released through a public repository to make the method available to the scientific community.


With an advent of technologya huge collection of digital images is formed as repositories on world wide web (WWW). The task of searching for similar images in the repository is difficult. In this paper, retrieval of similar images from www is demonstrated with the help of combination of image features as color and shape and then using Siamese neural network which is constructed to the requirement as a novel approach. Here, one-shot learning technique is used to test the Siamese Neural Network model for retrieval performance. Various experiments are conducted with both the methods and results obtained are tabulated. The performance of the system is evaluated with precision parameter and which is found to be high.Also, relative study is made with existing works.


2011 ◽  
Vol 328-330 ◽  
pp. 1763-1767
Author(s):  
Jian Qiang Shen ◽  
Xuan Zou

A novel approach is proposed for measuring fabric texture orientations and recognizing weave patterns. Wavelet transform is suited for fabric image decomposition and Radon Transform is fit for line detection in fabric texture. Since different weave patterns have their own regular orientations in original image and sub-band images decomposed by Wavelet transform, these orientations features are extracted and used as SOM and LVQ inputs to achieve automatic recognition of fabric weave. The experimental results show that the neural network of LVQ is more effective than SOM. The contribution of this study is that it not only can identify fundamental fabric weaves but also can classify double layer and some derivative twill weaves such as angular twill and pointed twill.


2015 ◽  
Vol 39 (1) ◽  
pp. 81-103
Author(s):  
Tho Thanh Quan ◽  
Xuan H. Luong ◽  
Thanh C. Nguyen ◽  
Hui Siu Cheung

Purpose – Most digital libraries (DL) are now available online. They also provide the Z39.50 standard protocol which allows computer-based systems to effectively retrieve information stored in the DLs. The major difficulty lies in inconsistency between database schemas of multiple DLs. The purpose of this paper is to present a system known as Argumentation-based Digital Library Search (ADLSearch), which facilitates information retrieval across multiple DLs. Design/methodology/approach – The proposed approach is based on argumentation theory for schema matching reconciliation from multiple schema matching algorithms. In addition, a distributed architecture is proposed for the ADLSearch system for information retrieval from multiple DLs. Findings – Initial performance results are promising. First, schema matching can improve the retrieval performance on DLs, as compared to the baseline technique. Subsequently, argumentation-based retrieval can yield better matching accuracy and retrieval efficiency than individual schema matching algorithms. Research limitations/implications – The work discussed in this paper has been implemented as a prototype supporting scholarly retrieval from about 800 DLs over the world. However, due to complexity of argumentation algorithm, the process of adding new DLs to the system cannot be performed in a real-time manner. Originality/value – In this paper, an argumentation-based approach is proposed for reconciling the conflicts from multiple schema matching algorithms in the context of information retrieval from multiple DL. Moreover, the proposed approach can also be applied for similar applications which require automatic mapping from multiple database schemas.


2013 ◽  
Vol 273 ◽  
pp. 796-799
Author(s):  
Yong Sheng Wang

This paper presents a novel approach to model 3D human face from multiple view 2D images in a fast mode. Our proposed method mainly includes three steps: 1) Face Recognition from 2D images, 2) Converting 2D images to 3D images, 3) Modeling 3D human face. To extract visual features of both 2D and 3D images, visual features adopted in 3D are described by Point Signature, and visual features utilized in 2D is represented by Gabor filter responses. Afterwards, 3D model is obtained by combining multiple view 2D images through calculating projections vector and translation vector. Experimental results show that our method can model 3D human face with high accuracy and efficiency.


2020 ◽  
Vol 20 (3) ◽  
pp. 75-85
Author(s):  
Shefali Dhingra ◽  
Poonam Bansal

AbstractContent Based Image Retrieval (CBIR) system is an efficient search engine which has the potentiality of retrieving the images from huge repositories by extracting the visual features. It includes color, texture and shape. Texture is the most eminent feature among all. This investigation focuses upon the classification complications that crop up in case of big datasets. In this, texture techniques are explored with machine learning algorithms in order to increase the retrieval efficiency. We have tested our system on three texture techniques using various classifiers which are Support vector machine, K-Nearest Neighbor (KNN), Naïve Bayes and Decision Tree (DT). Variant evaluation metrics precision, recall, false alarm rate, accuracy etc. are figured out to measure the competence of the designed CBIR system on two benchmark datasets, i.e. Wang and Brodatz. Result shows that with both these datasets the KNN and DT classifier hand over superior results as compared to others.


2019 ◽  
Author(s):  
Michael B. Bone ◽  
Fahad Ahmad ◽  
Bradley R. Buchsbaum

AbstractWhen recalling an experience of the past, many of the component features of the original episode may be, to a greater or lesser extent, reconstructed in the mind’s eye. There is strong evidence that the pattern of neural activity that occurred during an initial perceptual experience is recreated during episodic recall (neural reactivation), and that the degree of reactivation is correlated with the subjective vividness of the memory. However, while we know that reactivation occurs during episodic recall, we have lacked a way of precisely characterizing the contents—in terms of its featural constituents—of a reactivated memory. Here we present a novel approach, feature-specific informational connectivity (FSIC), that leverages hierarchical representations of image stimuli derived from a deep convolutional neural network to decode neural reactivation in fMRI data collected while participants performed an episodic recall task. We show that neural reactivation associated with low-level visual features (e.g. edges), high-level visual features (e.g. facial features), and semantic features (e.g. “terrier”) occur throughout the dorsal and ventral visual streams and extend into the frontal cortex. Moreover, we show that reactivation of both low- and high-level visual features correlate with the vividness of the memory, whereas only reactivation of low-level features correlates with recognition accuracy when the lure and target images are semantically similar. In addition to demonstrating the utility of FSIC for mapping feature-specific reactivation, these findings resolve the relative contributions of low- and high-level features to the vividness of visual memories, clarify the role of the frontal cortex during episodic recall, and challenge a strict interpretation the posterior-to-anterior visual hierarchy.


2020 ◽  
Vol 7 (4) ◽  
pp. 268-273
Author(s):  
Gibelli Daniele Maria ◽  
◽  
Poppa Pasquale ◽  
Cappella Annalisa ◽  
Rosati Riccardo ◽  
...  

Introduction The assessment of facial growth has always had a relevant importance in anatomy and morphological sciences. This article aims at presenting a method of facial superimposition between 3D models which provides a topographic map of those facial areas modified by growth. Methodology Eight children aged between 6 and 10 years were recruited. In December 2010 they underwent a 3D scan by the Vivid 910 laser scanner (Konica Minolta, Osaka, Japan). The same procedures were performed another five times, in June 2011, September 2011, January 2012 and September 2012; in total 6 analyses were performed on the same subjects in a time span of 21 months. Three-dimensional digital models belonging to the same individual were then superimposed on each other according to 11 facial landmarks. Three comparisons were performed for each individual, referring to the period between December 2010 and June 2011, between June 2011 and January 2012 and between January and September 2012. Results Results show that the protocol of superimposition gives a reliable image of facial growth with high sensibility: in detail, even the slight facial modifications due to different expressions are recorded. The method can also quantify the point-to-point difference between the two models, and therefore give an indication concerning the general increase or decrease of facial volume. Conclusion This approach may provide useful indications for the analysis of facial growth on a large sample and give a new point of view of the complex field of face development.


Author(s):  
Wei Li ◽  
Haiyu Song ◽  
Hongda Zhang ◽  
Houjie Li ◽  
Pengjie Wang

The ever-increasing size of images has made automatic image annotation one of the most important tasks in the fields of machine learning and computer vision. Despite continuous efforts in inventing new annotation algorithms and new models, results of the state-of-the-art image annotation methods are often unsatisfactory. In this paper, to further improve annotation refinement performance, a novel approach based on weighted mutual information to automatically refine the original annotations of images is proposed. Unlike the traditional refinement model using only visual feature, the proposed model use semantic embedding to properly map labels and visual features to a meaningful semantic space. To accurately measure the relevance between the particular image and its original annotations, the proposed model utilize all available information including image-to-image, label-to-label and image-to-label. Experimental results conducted on three typical datasets show not only the validity of the refinement, but also the superiority of the proposed algorithm over existing ones. The improvement largely benefits from our proposed mutual information method and utilizing all available information.


2020 ◽  
Vol 54 (2) ◽  
pp. 133-150
Author(s):  
Sumeer Gul ◽  
Sabha Ali ◽  
Aabid Hussain

PurposeThe purpose of this study is to assess the retrieval performance of three search engines, i.e. Google, Yahoo and Bing for navigational queries using two important retrieval measures, i.e. precision and relative recall in the field of life science and biomedicine.Design/methodology/approachTop three search engines namely Google, Yahoo and Bing were selected on the basis of their ranking as per Alexa, an analytical tool that provides ranking of global websites. Furthermore, the scope of study was confined to those search engines having interface in English. Clarivate Analytics' Web of Science was used for the extraction of navigational queries in the field of life science and biomedicine. Navigational queries (classified as one-word, two-word and three-word queries) were extracted from the keywords of the papers representing the top 100 contributing authors in the select field. Keywords were also checked for the duplication. Two important evaluation parameters, i.e. precision and relative recall were used to calculate the performance of search engines on the navigational queries.FindingsThe mean precision for Google scores high (2.30) followed by Yahoo (2.29) and Bing (1.68), while mean relative recall also scores high for Google (0.36) followed by Yahoo (0.33) and Bing (0.31) respectively.Research limitations/implicationsThe study is of great help to the researchers and academia in determining the retrieval efficiency of Google, Yahoo and Bing in terms of navigational query execution in the field of life science and biomedicine. The study can help users to focus on various search processes and the query structuring and its execution across the select search engines for achieving desired result list in a professional search environment. The study can also act as a ready reference source for exploring navigational queries and how these queries can be managed in the context of information retrieval process. It will also help to showcase the retrieval efficiency of various search engines on the basis of subject diversity (life science and biomedicine) highlighting the same in terms of query intention.Originality/valueThough many studies have been conducted highlighting the retrieval efficiency of search engines the current work is the first of its kind to study the retrieval effectiveness of Google, Yahoo and Bing on navigational queries in the field of life science and biomedicine. The study will help in understanding various methods and approaches to be adopted by the users for the navigational query execution across a professional search environment, i.e. “life science and biomedicine”


Sign in / Sign up

Export Citation Format

Share Document