scholarly journals Pin-missing defect recognition based on feature fusion and spatial attention mechanism

2022 ◽  
Vol 8 ◽  
pp. 656-663
Author(s):  
Hui He ◽  
Yuchen Li ◽  
Jing Yang ◽  
Zeli Wang ◽  
Bo Chen ◽  
...  
2021 ◽  
Vol 13 (10) ◽  
pp. 1950
Author(s):  
Cuiping Shi ◽  
Xin Zhao ◽  
Liguo Wang

In recent years, with the rapid development of computer vision, increasing attention has been paid to remote sensing image scene classification. To improve the classification performance, many studies have increased the depth of convolutional neural networks (CNNs) and expanded the width of the network to extract more deep features, thereby increasing the complexity of the model. To solve this problem, in this paper, we propose a lightweight convolutional neural network based on attention-oriented multi-branch feature fusion (AMB-CNN) for remote sensing image scene classification. Firstly, we propose two convolution combination modules for feature extraction, through which the deep features of images can be fully extracted with multi convolution cooperation. Then, the weights of the feature are calculated, and the extracted deep features are sent to the attention mechanism for further feature extraction. Next, all of the extracted features are fused by multiple branches. Finally, depth separable convolution and asymmetric convolution are implemented to greatly reduce the number of parameters. The experimental results show that, compared with some state-of-the-art methods, the proposed method still has a great advantage in classification accuracy with very few parameters.


2021 ◽  
Vol 50 ◽  
pp. 101392
Author(s):  
Yucheng Wang ◽  
Xinyu Li ◽  
Yiping Gao ◽  
Lijian Wang ◽  
Liang Gao

2021 ◽  
Author(s):  
Zhangfang Hu ◽  
Yanling Xia ◽  
Yuan Luo ◽  
Lan Wang

2021 ◽  
Author(s):  
Zongbao Liang ◽  
Xing Liu ◽  
Bo Chen ◽  
YunFei Yuan ◽  
Yang Song ◽  
...  

2020 ◽  
Vol 133 ◽  
pp. 327-333 ◽  
Author(s):  
Heng Zhou ◽  
Zhijun Fang ◽  
Yongbin Gao ◽  
Bo Huang ◽  
Cengsi Zhong ◽  
...  

2020 ◽  
Vol 10 (12) ◽  
pp. 4312 ◽  
Author(s):  
Jie Xu ◽  
Haoliang Wei ◽  
Linke Li ◽  
Qiuru Fu ◽  
Jinhong Guo

Video description plays an important role in the field of intelligent imaging technology. Attention perception mechanisms are extensively applied in video description models based on deep learning. Most existing models use a temporal-spatial attention mechanism to enhance the accuracy of models. Temporal attention mechanisms can obtain the global features of a video, whereas spatial attention mechanisms obtain local features. Nevertheless, because each channel of the convolutional neural network (CNN) feature maps has certain spatial semantic information, it is insufficient to merely divide the CNN features into regions and then apply a spatial attention mechanism. In this paper, we propose a temporal-spatial and channel attention mechanism that enables the model to take advantage of various video features and ensures the consistency of visual features between sentence descriptions to enhance the effect of the model. Meanwhile, in order to prove the effectiveness of the attention mechanism, this paper proposes a video visualization model based on the video description. Experimental results show that, our model has achieved good performance on the Microsoft Video Description (MSVD) dataset and a certain improvement on the Microsoft Research-Video to Text (MSR-VTT) dataset.


Sign in / Sign up

Export Citation Format

Share Document