gaussian derivative
Recently Published Documents


TOTAL DOCUMENTS

52
(FIVE YEARS 9)

H-INDEX

10
(FIVE YEARS 2)

Author(s):  
Tony Lindeberg

AbstractThis paper presents a hybrid approach between scale-space theory and deep learning, where a deep learning architecture is constructed by coupling parameterized scale-space operations in cascade. By sharing the learnt parameters between multiple scale channels, and by using the transformation properties of the scale-space primitives under scaling transformations, the resulting network becomes provably scale covariant. By in addition performing max pooling over the multiple scale channels, or other permutation-invariant pooling over scales, a resulting network architecture for image classification also becomes provably scale invariant. We investigate the performance of such networks on the MNIST Large Scale dataset, which contains rescaled images from the original MNIST dataset over a factor of 4 concerning training data and over a factor of 16 concerning testing data. It is demonstrated that the resulting approach allows for scale generalization, enabling good performance for classifying patterns at scales not spanned by the training data.


2021 ◽  
Vol 3 (3) ◽  
pp. 240-254
Author(s):  
Subarna Shakya

Face recognition at a distance (FRAD) is one of the most difficult types of face recognition applications, particularly at a distance. Due to the poor resolution of facial image, it is difficult to identify faces from a distance. Recently, while recording individuals, the camera view is broad and just a small portion of a person's face is visible in the image. To ensure that the facial image has a low resolution, which deteriorates both face detection and identification engines, the facial image is constantly at low resolution. As an immediate solution, employing a high-definition camera is considered as a simple and practical approach to improve the reliability of algorithm and perform well on low-resolution facial images. While facial detection will be somewhat decreased, a picture with higher quality will result in a slower face detection rate. The proposed work aims to recognize faces with good accuracy even at a distance. The eye localization works for the face and eye location in the face of a human being with varied sizes at multiple distances. This process is used to detect the face quickly with a comparatively high accuracy. The Gaussian derivative filter is used to reduce the feature size in the storage element, which improves the speed of the recognition ratio. Besides, the proposed work includes benchmark datasets to evaluate the recognition process. As a result, the proposed system has achieved a 93.24% average accuracy of face recognition.


2020 ◽  
Vol 2020 (28) ◽  
pp. 215-220
Author(s):  
Hiroaki Kotera

The edge response in retinal image is the first step for human vision recognizing the outside world. A variety of receptive field models for describing the impulse response have been proposed. Which satisfies the uncertain principle? occupied the interest from a point of minimizing the product (Δx)(Δ w) both in spatial and spectral. Among the typical edge response models, finally Gabor function and 2nd. Gaussian Derivative GD2 remained as strong candidates. While famous D. Marr and R. Young support GD2, many vision researchers prefer Gabor. The retinal edge response model is used for image sharpening.<br/> Different from the conventional image sharpening filters, this paper proposes a novel image sharpening filter by modifying the Lanczos resampling filter. The Lanczos filter is used for image scaling to resize digital images. Usually it works to interpolate the discrete sampled points like as a kind of smoothing filter not as sharpening. The Lanczos kernel is given by the product of sampling Sinc function and the scaled Sinc function. The scaled Sinc function expanded by the scale "s" plays a role of window function. The author noticed that the inverse scaling of Lanczos window can be used not for smoothing but for sharpening filter.<br/> This paper demonstrates how the proposed model works effectively in comparison with Gabor and GD2.


Sensors ◽  
2020 ◽  
Vol 20 (3) ◽  
pp. 785 ◽  
Author(s):  
Hansung Lee ◽  
So-Hee Park ◽  
Jang-Hee Yoo ◽  
Se-Hoon Jung ◽  
Jun-Ho Huh

Although access control based on human face recognition has become popular in consumer applications, it still has several implementation issues before it can realize a stand-alone access control system. Owing to a lack of computational resources, lightweight and computationally efficient face recognition algorithms are required. The conventional access control systems require significant active cooperation from the users despite its non-aggressive nature. The lighting/illumination change is one of the most difficult and challenging problems for human-face-recognition-based access control applications. This paper presents the design and implementation of a user-friendly, stand-alone access control system based on human face recognition at a distance. The local binary pattern (LBP)-AdaBoost framework was employed for face and eyes detection, which is fast and invariant to illumination changes. It can detect faces and eyes of varied sizes at a distance. For fast face recognition with a high accuracy, the Gabor-LBP histogram framework was modified by substituting the Gabor wavelet with Gaussian derivative filters, which reduced the facial feature size by 40% of the Gabor-LBP-based facial features, and was robust to significant illumination changes and complicated backgrounds. The experiments on benchmark datasets produced face recognition accuracies of 97.27% on an E-face dataset and 99.06% on an XM2VTS dataset, respectively. The system achieved a 91.5% true acceptance rate with a 0.28% false acceptance rate and averaged a 5.26 frames/sec processing speed on a newly collected face image and video dataset in an indoor office environment.


Sensors ◽  
2019 ◽  
Vol 19 (11) ◽  
pp. 2479 ◽  
Author(s):  
Lifu Chen ◽  
Xianliang Cui ◽  
Zhenhong Li ◽  
Zhihui Yuan ◽  
Jin Xing ◽  
...  

Synthetic Aperture Radar (SAR) scene classification is challenging but widely applied, in which deep learning can play a pivotal role because of its hierarchical feature learning ability. In the paper, we propose a new scene classification framework, named Feature Recalibration Network with Multi-scale Spatial Features (FRN-MSF), to achieve high accuracy in SAR-based scene classification. First, a Multi-Scale Omnidirectional Gaussian Derivative Filter (MSOGDF) is constructed. Then, Multi-scale Spatial Features (MSF) of SAR scenes are generated by weighting MSOGDF, a Gray Level Gradient Co-occurrence Matrix (GLGCM) and Gabor transformation. These features were processed by the Feature Recalibration Network (FRN) to learn high-level features. In the network, the Depthwise Separable Convolution (DSC), Squeeze-and-Excitation (SE) Block and Convolution Neural Network (CNN) are integrated. Finally, these learned features will be classified by the Softmax function. Eleven types of SAR scenes obtained from four systems combining different bands and resolutions were trained and tested, and a mean accuracy of 98.18% was obtained. To validate the generality of FRN-MSF, five types of SAR scenes sampled from two additional large-scale Gaofen-3 and TerraSAR-X images were evaluated for classification. The mean accuracy of the five types reached 94.56%; while the mean accuracy for the same five types of the former tested 11 types of scene was 96%. The high accuracy indicates that the FRN-MSF is promising for SAR scene classification without losing generality.


Complexity ◽  
2018 ◽  
Vol 2018 ◽  
pp. 1-23
Author(s):  
Jinping Liu ◽  
Jiezhou He ◽  
Zhaohui Tang ◽  
Pengfei Xu ◽  
Wuxia Zhang ◽  
...  

Texture pattern classification has long been an essential issue in computer vision (CV). However, texture is a kind of perceptual concept of human beings in scene observation or content understanding, which cannot be defined or described clearly in CV. Visually, the visual appearance of the complex spatial structure (CSS) of texture pattern (TP) generally depends on the random organization (or layout) of local homogeneous fragments (LHFs) in the imaged surface. Hence, it is essential to investigate the latent statistical distribution (LSD) behavior of LHFs for distinctive CSS feature characterization to achieve good classification performance. This work presents an image statistical modeling-based TP identification (ISM-TPI) method. It firstly makes a theoretical explanation of the Weibull distribution (WD) behavior of the LHFs of the imaged surface in the imaging process based on the sequential fragmentation theory (SFT), which consequently derives a symmetrical WD model (SWDM) to characterize the LSD of the TP’s SS. Multidirectional and multiscale TP features are then characterized by the SWDM parameters based on the oriented differential operators; in other words, texture images are convolved with multiscale and multidirectional Gaussian derivative filters (GDFs), including the steerable isotropic GDFs (SIGDFs) and the oriented anisotropic GDFs (OAGDFs), for the omnidirectional and multiscale SS detail exhibition with low computational complexity. Finally, SWDM-based TP feature parameters, demonstrated to be directly related to the human vision perception system with significant physical perception meaning, are extracted and used to TP classification with a partial least squares-discriminant analysis- (PLS-DA-) based classifier. The effectiveness of the proposed ISM-TPI method is verified by extensive experiments on three texture image databases. The classification results demonstrate the superiority of the proposed methods over several state-of-the-art TP classification methods.


Sign in / Sign up

Export Citation Format

Share Document