scholarly journals FSD-BRIEF: A Distorted BRIEF Descriptor for Fisheye Image Based on Spherical Perspective Model

Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1839
Author(s):  
Yutong Zhang ◽  
Jianmei Song ◽  
Yan Ding ◽  
Yating Yuan ◽  
Hua-Liang Wei

Fisheye images with a far larger Field of View (FOV) have severe radial distortion, with the result that the associated image feature matching process cannot achieve the best performance if the traditional feature descriptors are used. To address this challenge, this paper reports a novel distorted Binary Robust Independent Elementary Feature (BRIEF) descriptor for fisheye images based on a spherical perspective model. Firstly, the 3D gray centroid of feature points is designed, and the position and direction of the feature points on the spherical image are described by a constructed feature point attitude matrix. Then, based on the attitude matrix of feature points, the coordinate mapping relationship between the BRIEF descriptor template and the fisheye image is established to realize the computation associated with the distorted BRIEF descriptor. Four experiments are provided to test and verify the invariance and matching performance of the proposed descriptor for a fisheye image. The experimental results show that the proposed descriptor works well for distortion invariance and can significantly improve the matching performance in fisheye images.

2021 ◽  
Vol 5 (4) ◽  
pp. 783-793
Author(s):  
Muhammad Muttabi Hudaya ◽  
Siti Saadah ◽  
Hendy Irawan

needs a solid validation that has verification and matching uploaded images. To solve this problem, this paper implementing a detection model using Faster R-CNN and a matching method using ORB (Oriented FAST and Rotated BRIEF) and KNN-BFM (K-Nearest Neighbor Brute Force Matcher). The goal of the implementations is to reach both an 80% mark of accuracy and prove matching using ORB only can be a replaced OCR technique. The implementation accuracy results in the detection model reach mAP (Mean Average Precision) of 94%. But, the matching process only achieves an accuracy of 43,46%. The matching process using only image feature matching underperforms the previous OCR technique but improves processing time from 4510ms to 60m). Image matching accuracy has proven to increase by using a high-quality dan high quantity dataset, extracting features on the important area of EKTP card images.


2019 ◽  
Vol 2019 ◽  
pp. 1-15
Author(s):  
Buhai Shi ◽  
Qingming Zhang ◽  
Haibo Xu

This paper presents a geometrical-information-assisted approach for matching local features. With the aid of Bayes’ theorem, it is found that the posterior confidence of matched features can be improved by introducing global geometrical information given by distances between feature points. Based on this result, we work out an approach to obtain the geometrical information and apply it to assist matching features. The pivotal techniques in this paper include (1) exploiting elliptic parameters of feature descriptors to estimate transformations that map feature points in images to points in an assumed plane; (2) projecting feature points to the assumed plane and finding a reliable referential point in it; (3) computing differences of the distances between the projected points and the referential point. Our new approach employs these differences to assist matching features, reaching better performance than the nearest neighbor-based approach in precision versus the number of matched features.


Symmetry ◽  
2018 ◽  
Vol 10 (12) ◽  
pp. 725
Author(s):  
Wei Zhang ◽  
Guoying Zhang

Image feature description and matching is widely used in computer vision, such as camera pose estimation. Traditional feature descriptions lack the semantic and spatial information, and give rise to a large number of feature mismatches. In order to improve the accuracy of image feature matching, a feature description and matching method, based on local semantic information fusion and feature spatial consistency, is proposed in this paper. Once object detection is used on images, feature points are then extracted, and image patches with various sizes surrounding these points are clipped. These patches are sent into the Siamese convolution network to get their semantic vectors. Then, semantic fusion description of feature points is obtained by weighted sum of the semantic vectors, and their weights optimized by particle swarm optimization (PSO) algorithm. When matching these feature points using their descriptions, feature spatial consistency is calculated based on the spatial consistency of matched objects, and the orientation and distance constraint of adjacent points within matched objects. With the description and matching method, the feature points are matched accurately and effectively. Our experiment results showed the efficiency of our methods.


2014 ◽  
Vol 644-650 ◽  
pp. 4157-4161
Author(s):  
Xin Zhang ◽  
Ya Sheng Zhang ◽  
Hong Yao

In the process of image matching, it is involved such as image rotation, scale zooming, brightness change and other problems. In order to improve the precision of image matching, image matching algorithm based on SIFT feature point is proposed. First, the method of generating scale space is introduced. Then, the scale and position of feature points are determined through three dimension quadratic function and feature vectors are determined through gradient distribution characteristic of neighborhood pixels of feature points. Finally, feature matching is completed based on the Euclidean distance. The experiment result indicates that using SIFT feature point can achieve image matching effectively.


2021 ◽  
Author(s):  
Aikui Tian ◽  
Kangtao Wang ◽  
liye zhang ◽  
Bingcai Wei

Abstract Aiming at the problem of inaccurate extraction of feature points by the traditional image matching method, low robustness, and problems such as diffculty in inentifying feature points in area with poor texture. This paper proposes a new local image feature matching method, which replaces the traditional sequential image feature detection, description and matching steps. First, extract the coarse features with a resolution of 1/8 from the original image, then tile to a one-dimensional vector plus the positional encoding, feed them to the self-attention layer and cross-attention layer in the Transformer module, and finally get through the Differentiable Matching Layer and confidence matrix, after setting the threshold and the mutual closest standard, a Coarse-Level matching prediction is obtained. Secondly the fine matching is refined at the Fine-level match, after the Fine-level match is established, the image overlapped area is aligned by transforming the matrix to a unified coordinate, and finally the image is fused by the weighted fusion algorithm to realize the unification of seamless mosaic of images. This paper uses the self-attention layer and cross-attention layer in Transformers to obtain the feature descriptor of the image. Finally, experiments show that in terms of feature point extraction, LoFTR algorithm is more accurate than the traditional SIFT algorithm in both low-texture regions and regions with rich textures. At the same time, the image mosaic effect obtained by this method is more accurate than that of the traditional classic algorithms, the experimental effect is more ideal.


Author(s):  
Marizuana Mat Daud ◽  
Zulaikha Kadim ◽  
Shang Li Yuen ◽  
Hon Hock Woon ◽  
Ibrahima Faye ◽  
...  

2011 ◽  
Vol 65 ◽  
pp. 497-502
Author(s):  
Yan Wei Wang ◽  
Hui Li Yu

A feature matching algorithm based on wavelet transform and SIFT is proposed in this paper, Firstly, Biorthogonal wavelet transforms algorithm is used for medical image to delaminating, and restoration the processed image. Then the SIFT (Scale Invariant Feature Transform) applied in this paper to abstracting key point. Experimental results show that our algorithm compares favorably in high-compressive ratio, the rapid matching speed and low storage of the image, especially for the tilt and rotation conditions.


Sign in / Sign up

Export Citation Format

Share Document