pyramid matching
Recently Published Documents


TOTAL DOCUMENTS

136
(FIVE YEARS 29)

H-INDEX

15
(FIVE YEARS 3)

2022 ◽  
Vol 70 (3) ◽  
pp. 5039-5058
Author(s):  
Khuram Nawaz Khayam ◽  
Zahid Mehmood ◽  
Hassan Nazeer Chaudhry ◽  
Muhammad Usman Ashraf ◽  
Usman Tariq ◽  
...  

Electronics ◽  
2021 ◽  
Vol 10 (24) ◽  
pp. 3177
Author(s):  
Venkat Anil Adibhatla ◽  
Yu-Chieh Huang ◽  
Ming-Chung Chang ◽  
Hsu-Chi Kuo ◽  
Abhijeet Utekar ◽  
...  

Deep learning methods are currently used in industries to improve the efficiency and quality of the product. Detecting defects on printed circuit boards (PCBs) is a challenging task and is usually solved by automated visual inspection, automated optical inspection, manual inspection, and supervised learning methods, such as you only look once (YOLO) of tiny YOLO, YOLOv2, YOLOv3, YOLOv4, and YOLOv5. Previously described methods for defect detection in PCBs require large numbers of labeled images, which is computationally expensive in training and requires a great deal of human effort to label the data. This paper introduces a new unsupervised learning method for the detection of defects in PCB using student–teacher feature pyramid matching as a pre-trained image classification model used to learn the distribution of images without anomalies. Hence, we extracted the knowledge into a student network which had same architecture as the teacher network. This one-step transfer retains key clues as much as possible. In addition, we incorporated a multi-scale feature matching strategy into the framework. A mixture of multi-level knowledge from the features pyramid passes through a better supervision, known as hierarchical feature alignment, which allows the student network to receive it, thereby allowing for the detection of various sizes of anomalies. A scoring function reflects the probability of the occurrence of anomalies. This framework helped us to achieve accurate anomaly detection. Apart from accuracy, its inference speed also reached around 100 frames per second.


Author(s):  
Santhoshkumar SP ◽  
Kumar M Praveen ◽  
Beaulah H Lilly

Video has more information than the isolated images. Processing, analyzing and understanding of contents present in videos are becoming very important. Consumer videos are generally captured by amateurs using handheld cameras of events and it contains considerable camera motion, occlusion, cluttered background, and large intraclass variations within the same type of events, making their visual cues highly variable and less discriminant. So visual event recognition is an extremely challenging task in computer vision. A visual event recognition framework for consumer videos is framed by leveraging a large amount of loosely labeled web videos. The videos are divided into training and testing sets manually. A simple method called the Aligned Space-Time Pyramid Matching method was proposed to effectively measure the distances between two video clips from different domains. Each video is divided into space-time volumes over multiple levels. A new transfer learning method is referred to as Adaptive Multiple Kernel Learning fuse the information from multiple pyramid levels, features, and copes with the considerable variation in feature distributions between videos from two domains web video domain and consumer video domain.With the help of MATLAB Simulink videos are divided and compared with web domain videos. The inputs are taken from the Kodak data set and the results are given in the form of MATLAB simulation.


2021 ◽  
Vol 72 (6) ◽  
pp. 374-380
Author(s):  
Bhavinkumar Gajjar ◽  
Hiren Mewada ◽  
Ashwin Patani

Abstract Support vector machine (SVM) techniques and deep learning have been prevalent in object classification for many years. However, deep learning is computation-intensive and can require a long training time. SVM is significantly faster than Convolution Neural Network (CNN). However, the SVM has limited its applications in the mid-size dataset as it requires proper tuning. Recently the parameterization of multiple kernels has shown greater flexibility in the characterization of the dataset. Therefore, this paper proposes a sparse coded multi-scale approach to reduce training complexity and tuning of SVM using a non-linear fusion of kernels for large class natural scene classification. The optimum features are obtained by parameterizing the dictionary, Scale Invariant Feature Transform (SIFT) parameters, and fusion of multiple kernels. Experiments were conducted on a large dataset to examine the multi-kernel space capability to find distinct features for better classification. The proposed approach founds to be promising than the linear multi-kernel SVM approaches achieving 91.12 % maximum accuracy.


2021 ◽  
Vol 67 (1) ◽  
Author(s):  
Sung-Wook Hwang ◽  
Junji Sugiyama

AbstractAlthough wood cross sections contain spatiotemporal information regarding tree growth, computer vision-based wood identification studies have traditionally favored disordered image representations that do not take such information into account. This paper describes image partitioning strategies that preserve the spatial information of wood cross-sectional images. Three partitioning strategies are designed, namely grid partitioning based on spatial pyramid matching and its variants, radial and tangential partitioning, and their recognition performance is evaluated for the Fagaceae micrograph dataset. The grid and radial partitioning strategies achieve better recognition performance than the bag-of-features model that constitutes their underlying framework. Radial partitioning, which is a strategy for preserving spatial information from pith to bark, further improves the performance, especially for radial-porous species. The Pearson correlation and autocorrelation coefficients produced from radially partitioned sub-images have the potential to be used as auxiliaries in the construction of multi-feature datasets. The contribution of image partitioning strategies is found to be limited to species recognition and is unremarkable at the genus level.


Sensors ◽  
2021 ◽  
Vol 21 (3) ◽  
pp. 923
Author(s):  
Ente Guo ◽  
Zhifeng Chen ◽  
Yanlin Zhou ◽  
Dapeng Oliver Wu

Estimating the depth of image and egomotion of agent are important for autonomous and robot in understanding the surrounding environment and avoiding collision. Most existing unsupervised methods estimate depth and camera egomotion by minimizing photometric error between adjacent frames. However, the photometric consistency sometimes does not meet the real situation, such as brightness change, moving objects and occlusion. To reduce the influence of brightness change, we propose a feature pyramid matching loss (FPML) which captures the trainable feature error between a current and the adjacent frames and therefore it is more robust than photometric error. In addition, we propose the occlusion-aware mask (OAM) network which can indicate occlusion according to change of masks to improve estimation accuracy of depth and camera pose. The experimental results verify that the proposed unsupervised approach is highly competitive against the state-of-the-art methods, both qualitatively and quantitatively. Specifically, our method reduces absolute relative error (Abs Rel) by 0.017–0.088.


Author(s):  
Kai-Chao Yao ◽  
Shih-Feng Fu ◽  
Wei-Tzer Huang ◽  
Cheng-Chun Wu

This article uses LabVIEW, a software program to develop a whitefly feature identification and counting technology, and machine learning algorithms for whitefly monitoring, identification, and counting applications. In addition, a high-magnification CCD camera is used for on-demand image photography, and then the functional programs of the VI library of LabVIEW NI-DAQ and LabVIEW NI Vision Development Module are used to develop image recognition functions. The grayscale-value pyramid-matching algorithm is used for image conversion and recognition in the machine learning mode. The built graphical user interface and device hardware provide convenient and effective whitefly feature identification and sample counting. This monitoring technology exhibits features such as remote monitoring, counting, data storage, and statistical analysis.


2020 ◽  
Vol 0 (0) ◽  
Author(s):  
Junfeng Jing ◽  
Huanhuan Ren

AbstractTo solve the problem of defect detection in printed fabrics caused by abundant colors and varied patterns, a defect detection method based on RGB accumulative average method (RGBAAM) and image pyramid matching is proposed. First, the minimum period of the printed fabric is calculated by the RGBAAM. Second, a Gaussian pyramid is constructed for the template image and the detected image by using the minimum period as a template. Third, the similarity measurement method is used to match the template image and the detected image. Finally, the position of the printed fabric defect is marked in the image to be detected by using the Laplacian pyramid restoration. The experimental results show that the method can accurately segment the printed fabric periodic unit and locate the defect position. The calculation cost is low for the method proposed in this article.


Sign in / Sign up

Export Citation Format

Share Document