Recognition of Birds in Blurred and Illumination Images by Local Features

Author(s):  
Suresha .M ◽  
. Sandeep

Local features are of great importance in computer vision. It performs feature detection and feature matching are two important tasks. In this paper concentrates on the problem of recognition of birds using local features. Investigation summarizes the local features SURF, FAST and HARRIS against blurred and illumination images. FAST and Harris corner algorithm have given less accuracy for blurred images. The SURF algorithm gives best result for blurred image because its identify strongest local features and time complexity is less and experimental demonstration shows that SURF algorithm is robust for blurred images and the FAST algorithms is suitable for images with illumination.

2020 ◽  
Vol 17 (9) ◽  
pp. 4419-4424
Author(s):  
Venkat P. Patil ◽  
C. Ram Singla

Image mosaicing is a method that combines several images or pictures of the superposing field of view to create a panoramic high-resolution picture. In the field of medical imagery, satellite data, computer vision, military automatic target recognition can be seen the importance of image mosaicing. The present domains of studies in computer vision, computer graphics and photo graphics are image stitching and video stitching. The registration of images includes five primary phases: feature detection and description; matching feature; rejection of outliers; transformation function derivation; image replication. Stitching images from specific scenes is a difficult job when images can be picked up under different noise. In this paper, we examine an algorithm for seamless stitching of images in order to resolve all such problems by employing dehazing methods to the collected images, and before defining image features and bound energy characteristics that match image-based features of the SIFT-Scale Invariant Feature Transform. The proposed method experimentation is compared with the conventional methods of stitching of image using squared distance to match the feature. The proposed seamless stitching technique is assessed on the basis of the metrics, HSGV and VSGV. The analysis of this stitching algorithm aims to minimize the amount of computation time and discrepancies in the final stitched results obtained.


2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
Di Wang ◽  
Hongying Zhang ◽  
Yanhua Shao

The precise evaluation of camera position and orientation is a momentous procedure of most machine vision tasks, especially visual localization. Aiming at the shortcomings of local features of dealing with changing scenes and the problem of realizing a robust end-to-end network that worked from feature detection to matching, an invariant local feature matching method for changing scene image pairs is proposed, which is a network that integrates feature detection, descriptor constitution, and feature matching. In the feature point detection and descriptor construction stage, joint training is carried out based on a neural network. In the feature point extraction and descriptor construction stage, joint training is carried out based on a neural network. To obtain local features with solid robustness to viewpoint and illumination changes, the Vector of Locally Aggregated Descriptors based on Neural Network (NetVLAD) module is introduced to compute the degree of correlation of description vectors from one image to another counterpart. Then, to enhance the relationship between relevant local features of image pairs, the attentional graph neural network (AGNN) is introduced, and the Sinkhorn algorithm is used to match them; finally, the local feature matching results between image pairs are output. The experimental results show that, compared with the existed algorithms, the proposed method enhances the robustness of local features of varying sights, performs better in terms of homography estimation, matching precision, and recall, and when meeting the requirements of the visual localization system to the environment, the end-to-end network tasks can be realized.


2012 ◽  
Vol 2012 ◽  
pp. 1-7 ◽  
Author(s):  
Huiwen Cai ◽  
Xiaoyan Wang ◽  
Ming Xia ◽  
Yangsheng Wang

Maximally stable extremal regions (MSER) is a state-of-the-art method in local feature detection. However, this method is sensitive to blurring because, in blurred images, the intensity values in region boundary will vary more slowly, and this will undermine the stability criterion that the MSER relies on. In this paper, we propose a method to improve MSER, making it more robust to image blurring. To find back the regions missed by MSER in the blurred image, we utilize the fact that the entropy of probability distribution function of intensity values increases rapidly when the local region expands across the boundary, while the entropy in the central part remains small. We use the entropy averaged by the regional area as a measure to reestimate regions missed by MSER. Experiments show that, when dealing with blurred images, the proposed method has better performance than the original MSER, with little extra computational effort.


2017 ◽  
Vol 2 (1) ◽  
pp. 80-87
Author(s):  
Puyda V. ◽  
◽  
Stoian. A.

Detecting objects in a video stream is a typical problem in modern computer vision systems that are used in multiple areas. Object detection can be done on both static images and on frames of a video stream. Essentially, object detection means finding color and intensity non-uniformities which can be treated as physical objects. Beside that, the operations of finding coordinates, size and other characteristics of these non-uniformities that can be used to solve other computer vision related problems like object identification can be executed. In this paper, we study three algorithms which can be used to detect objects of different nature and are based on different approaches: detection of color non-uniformities, frame difference and feature detection. As the input data, we use a video stream which is obtained from a video camera or from an mp4 video file. Simulations and testing of the algoritms were done on a universal computer based on an open-source hardware, built on the Broadcom BCM2711, quad-core Cortex-A72 (ARM v8) 64-bit SoC processor with frequency 1,5GHz. The software was created in Visual Studio 2019 using OpenCV 4 on Windows 10 and on a universal computer operated under Linux (Raspbian Buster OS) for an open-source hardware. In the paper, the methods under consideration are compared. The results of the paper can be used in research and development of modern computer vision systems used for different purposes. Keywords: object detection, feature points, keypoints, ORB detector, computer vision, motion detection, HSV model color


2019 ◽  
Vol 31 (6) ◽  
pp. 844-850 ◽  
Author(s):  
Kevin T. Huang ◽  
Michael A. Silva ◽  
Alfred P. See ◽  
Kyle C. Wu ◽  
Troy Gallerani ◽  
...  

OBJECTIVERecent advances in computer vision have revolutionized many aspects of society but have yet to find significant penetrance in neurosurgery. One proposed use for this technology is to aid in the identification of implanted spinal hardware. In revision operations, knowing the manufacturer and model of previously implanted fusion systems upfront can facilitate a faster and safer procedure, but this information is frequently unavailable or incomplete. The authors present one approach for the automated, high-accuracy classification of anterior cervical hardware fusion systems using computer vision.METHODSPatient records were searched for those who underwent anterior-posterior (AP) cervical radiography following anterior cervical discectomy and fusion (ACDF) at the authors’ institution over a 10-year period (2008–2018). These images were then cropped and windowed to include just the cervical plating system. Images were then labeled with the appropriate manufacturer and system according to the operative record. A computer vision classifier was then constructed using the bag-of-visual-words technique and KAZE feature detection. Accuracy and validity were tested using an 80%/20% training/testing pseudorandom split over 100 iterations.RESULTSA total of 321 total images were isolated containing 9 different ACDF systems from 5 different companies. The correct system was identified as the top choice in 91.5% ± 3.8% of the cases and one of the top 2 or 3 choices in 97.1% ± 2.0% and 98.4 ± 13% of the cases, respectively. Performance persisted despite the inclusion of variable sizes of hardware (i.e., 1-level, 2-level, and 3-level plates). Stratification by the size of hardware did not improve performance.CONCLUSIONSA computer vision algorithm was trained to classify at least 9 different types of anterior cervical fusion systems using relatively sparse data sets and was demonstrated to perform with high accuracy. This represents one of many potential clinical applications of machine learning and computer vision in neurosurgical practice.


2019 ◽  
Vol 8 (4) ◽  
pp. 8231-8236

A restoration and classification computation for blurred image which depends on obscure identification and characterization is proposed in this paper. Initially, new obscure location calculation is proposed to recognize the Gaussian, Motion and Defocus based blurred locales in the image. The degradation-restoration model referred with pre-processing followed by binarization and features extraction/classification algorithm applied on obscure images. At this point, support vector machine (SVM) classification algorithm is proposed to cluster the blurred images. Once the obscure class of the locales is affirmed, the structure of the obscure kernels of the blurred images are affirmed. At that point, the obscure kernel estimation techniques are embraced to appraise the obscure kernels. At last, the blurred locales are re-established utilizing nonblind image deblurring calculation and supplant the blurred images with the restored images. The simulation results demonstrate that the proposed calculation performs well


Automatic image registration (IR) is very challenging and very important in the field of hyperspectral remote sensing data. Efficient autonomous IR method is needed with high precision, fast, and robust. A key operation of IR is to align the multiple images in single co-ordinate system for extracting and identifying variation between images considered. In this paper, presented a feature descriptor by combining features from both Feature from Accelerated Segment Test (FAST) and Binary Robust Invariant Scalable Key point (BRISK). The proposed hybrid invariant local features (HILF) descriptor extract useful and similar feature sets from reference and source images. The feature matching method allows finding precise relationship or matching among two feature sets. An experimental analysis described the outcome BRISK, FASK and proposed HILF in terms of inliers ratio and repeatability evaluation metrics.


2010 ◽  
Vol 9 (4) ◽  
pp. 29-34 ◽  
Author(s):  
Achim Weimert ◽  
Xueting Tan ◽  
Xubo Yang

In this paper, we present a novel feature detection approach designed for mobile devices, showing optimized solutions for both detection and description. It is based on FAST (Features from Accelerated Segment Test) and named 3D FAST. Being robust, scale-invariant and easy to compute, it is a candidate for augmented reality (AR) applications running on low performance platforms. Using simple calculations and machine learning, FAST is a feature detection algorithm known to be efficient but not very robust in addition to its lack of scale information. Our approach relies on gradient images calculated for different scale levels on which a modified9 FAST algorithm operates to obtain the values of the corner response function. We combine the detection with an adapted version of SURF (Speed Up Robust Features) descriptors, providing a system with all means to implement feature matching and object detection. Experimental evaluation on a Symbian OS device using a standard image set and comparison with SURF using Hessian matrix-based detector is included in this paper, showing improvements in speed (compared to SURF) and robustness (compared to FAST)


2021 ◽  
Author(s):  
Longfei Zhou ◽  
Lin Zhang ◽  
Nicholas Konz

Computer vision techniques have played an important role in promoting the informatization, digitization and intelligence of industrial manufacturing systems. Considering the rapid development of computer vision techniques, we present a comprehensive review of the state-of-the-art of these techniques and their applications in manufacturing industries. We survey the most common methods, including feature detection, recognition, segmentation and 3D modeling. A system framework of computer vision in the manufacturing environment is proposed, consisting of a lighting module, a manufacturing system, a sensing module, computer vision algorithms, a decision-making module, and an actuator. Applications of computer vision to different stages of the entire product life cycle are then explored, including product design, modeling and simulation, planning and scheduling, the production process, inspection and quality control, assembly, transportation, and disassembly. Challenges include algorithm implementation, data pre-processing, data labeling, and benchmarks. Future directions include building benchmarks, developing methods for non-annotated data processing, developing effective data pre-processing mechanisms, customizing computer vision models, and opportunities aroused by 5G.


Sign in / Sign up

Export Citation Format

Share Document