local descriptor
Recently Published Documents


TOTAL DOCUMENTS

238
(FIVE YEARS 73)

H-INDEX

20
(FIVE YEARS 6)

Author(s):  
Cristina Romero-González ◽  
Ismael García-Varea ◽  
Jesus Martínez-Gómez

AbstractMany of the research problems in robot vision involve the detection of keypoints, areas with salient information in the input images and the generation of local descriptors, that encode relevant information for such keypoints. Computer vision solutions have recently relied on Deep Learning techniques, which make extensive use of the computational capabilities available. In autonomous robots, these capabilities are usually limited and, consequently, images cannot be processed adequately. For this reason, some robot vision tasks still benefit from a more classic approach based on keypoint detectors and local descriptors. In 2D images, the use of binary representations for visual tasks has shown that, with lower computational requirements, they can obtain a performance comparable to classic real-value techniques. However, these achievements have not been fully translated to 3D images, where research is mainly focused on real-value approaches. Thus, in this paper, we propose a keypoint detector and local descriptor based on 3D binary patterns. The experimentation demonstrates that our proposal is competitive against state-of-the-art techniques, while its processing can be performed more efficiently.


Facial expression plays an important role in communicating emotions. In this paper, a robust method for recognizing facial expressions is proposed using the combination of appearance features. Traditionally, appearance features mainly divide any face image into regular matrices for the computation of facial expression recognition. However, in this paper, we have computed appearance features in specific regions by extracting facial components such as eyes, nose, mouth, and forehead, etc. The proposed approach mainly has five stages to detect facial expression viz. face detection and regions of interest extraction, feature extraction, pattern analysis using a local descriptor, the fusion of appearance features and finally classification using a Multiclass Support Vector Machine (MSVM). Results of the proposed method are compared with the earlier holistic representations for recognizing facial expressions, and it is found that the proposed method outperforms state-of-the-art methods.


2021 ◽  
Author(s):  
Usman Muhammad ◽  
Md. Ziaul Hoque ◽  
Weiqiang Wang ◽  
Mourad Oussalah

The bag-of-words (BoW) model is one of the most popular representation methods for image classification. However, the lack of spatial information, change of illumination, and inter-class similarity among scene categories impair its performance in the remote-sensing domain. To alleviate these issues, this paper proposes to explore the spatial dependencies between different image regions and introduce a neighborhood-based collaborative learning (NBCL) for remote-sensing scene classification. Particularly, our proposed method employs multilevel features learning based on small, medium, and large neighborhood regions to enhance the discriminative power of image representation. To achieve this, image patches are selected through a fixed-size sliding window where each image is represented by four independent image region sequences. Apart from multilevel learning, we explicitly impose Gaussian pyramids to magnify the visual information of the scene images and optimize their position and scale parameters locally. Motivated by this, a local descriptor is exploited to extract multilevel and multiscale features that we represent in terms of codewords histogram by performing k-means clustering. Finally, a simple fusion strategy is proposed to balance the contribution of these features, and the fused features are incorporated into a Bidirectional Long Short-Term Memory (BiLSTM) network for constructing the final representation for classification. Experimental results on NWPU-RESISC45, AID, UC-Merced, and WHU-RS datasets demonstrate that the proposed approach not only surpasses the conventional bag-of-words approaches but also yields significantly higher classification performance than the existing state-of-the-art deep learning methods used nowadays.


2021 ◽  
Author(s):  
Usman Muhammad ◽  
Md. Ziaul Hoque ◽  
Weiqiang Wang ◽  
Mourad Oussalah

The bag-of-words (BoW) model is one of the most popular representation methods for image classification. However, the lack of spatial information, change of illumination, and inter-class similarity among scene categories impair its performance in the remote-sensing domain. To alleviate these issues, this paper proposes to explore the spatial dependencies between different image regions and introduce a neighborhood-based collaborative learning (NBCL) for remote-sensing scene classification. Particularly, our proposed method employs multilevel features learning based on small, medium, and large neighborhood regions to enhance the discriminative power of image representation. To achieve this, image patches are selected through a fixed-size sliding window where each image is represented by four independent image region sequences. Apart from multilevel learning, we explicitly impose Gaussian pyramids to magnify the visual information of the scene images and optimize their position and scale parameters locally. Motivated by this, a local descriptor is exploited to extract multilevel and multiscale features that we represent in terms of codewords histogram by performing k-means clustering. Finally, a simple fusion strategy is proposed to balance the contribution of these features, and the fused features are incorporated into a Bidirectional Long Short-Term Memory (BiLSTM) network for constructing the final representation for classification. Experimental results on NWPU-RESISC45, AID, UC-Merced, and WHU-RS datasets demonstrate that the proposed approach not only surpasses the conventional bag-of-words approaches but also yields significantly higher classification performance than the existing state-of-the-art deep learning methods used nowadays.


2021 ◽  
Author(s):  
Chengzhang Shi ◽  
Chung-Ming Own ◽  
Ching-Chih Chou ◽  
Bailu Guo

Sign in / Sign up

Export Citation Format

Share Document