Research on Acceleration Methods of Semi-training Color Stripping DehazeNet

Author(s):  
Dahai Ji ◽  
Ye Wang
Keyword(s):  
2020 ◽  
Vol 62 (3) ◽  
pp. 30-38
Author(s):  
Achilleas Marinakis ◽  
Panagiotis J. Papakanellos ◽  
George Fikioris

2021 ◽  
Author(s):  
Dominika Przewlocka ◽  
Marcin Kowalczyk ◽  
Tomasz Kryjak

Deep learning algorithms are a key component of many state-of-the-art vision systems, especially as Convolutional Neural Networks (CNN) outperform most solutions in the sense of accuracy. To apply such algorithms in real-time applications, one has to address the challenges of memory and computational complexity. To deal with the first issue, we use networks with reduced precision, specifically a binary neural network (also known as XNOR). To satisfy the computational requirements, we propose to use highly parallel and low-power FPGA devices. In this work, we explore the possibility of accelerating XNOR networks for traffic sign classification. The trained binary networks are implemented on the ZCU 104 development board, equipped with a Zynq UltraScale+ MPSoC device using two different approaches. Firstly, we propose a custom HDL accelerator for XNOR networks, which enables the inference with almost 450 fps. Even better results are obtained with the second method - the Xilinx FINN accelerator - enabling to process input images with around 550 frame rate. Both approaches provide over 96% accuracy on the test set.


2009 ◽  
Vol 82 (4) ◽  
pp. 165-170
Author(s):  
Takahisa ISHIMURA ◽  
Takayuki HONDA ◽  
Rong LU ◽  
Tetsuo MIYAKOSHI
Keyword(s):  

2020 ◽  
Vol 34 (07) ◽  
pp. 10957-10964
Author(s):  
Lanqing He ◽  
Zhongdao Wang ◽  
Yali Li ◽  
Shengjin Wang

The softmax loss and its variants are widely used as objectives for embedding learning applications like face recognition. However, the intra- and inter-class objectives in Softmax are entangled, therefore a well-optimized inter-class objective leads to relaxation on the intra-class objective, and vice versa. In this paper, we propose to dissect Softmax into independent intra- and inter-class objective (D-Softmax) with a clear understanding. It is straightforward to tune each part to the best state with D-Softmax as objective.Furthermore, we find the computation of the inter-class part is redundant and propose sampling-based variants of D-Softmax to reduce the computation cost. The face recognition experiments on regular-scale data show D-Softmax is favorably comparable to existing losses such as SphereFace and ArcFace. Experiments on massive-scale data show the fast variants significantly accelerates the training process (such as 64×) with only a minor sacrifice in performance, outperforming existing acceleration methods of Softmax in terms of both performance and efficiency.


Sign in / Sign up

Export Citation Format

Share Document