Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation

2021 ◽  
Vol 20 (5s) ◽  
pp. 1-21
Author(s):  
Fateme S. Hosseini ◽  
Fanruo Meng ◽  
Chengmo Yang ◽  
Wujie Wen ◽  
Rosario Cammarota

Hardware accelerators are essential to the accommodation of ever-increasing Deep Neural Network (DNN) workloads on the resource-constrained embedded devices. While accelerators facilitate fast and energy-efficient DNN operations, their accuracy is threatened by faults in their on-chip and off-chip memories, where millions of DNN weights are held. The use of emerging Non-Volatile Memories (NVM) further exposes DNN accelerators to a non-negligible rate of permanent defects due to immature fabrication, limited endurance, and aging. To tolerate defects in NVM-based DNN accelerators, previous work either requires extra redundancy in hardware or performs defect-aware retraining, imposing significant overhead. In comparison, this paper proposes a set of algorithms that exploit the flexibility in setting the fault-free bits in weight memory to effectively approximate weight values, so as to mitigate defect-induced accuracy drop. These algorithms can be applied as a one-step solution when loading the weights to embedded devices. They only require trivial hardware support and impose negligible run-time overhead. Experiments on popular DNN models show that the proposed techniques successfully boost inference accuracy even in the face of elevated defect rates in the weight memory.

Author(s):  
Hendrik Wohrle ◽  
Mariela De Lucas Alvarez ◽  
Fabian Schlenke ◽  
Alexander Walsemann ◽  
Michael Karagounis ◽  
...  

2020 ◽  
Vol 10 (4) ◽  
pp. 33
Author(s):  
Pramesh Pandey ◽  
Noel Daniel Gundi ◽  
Prabal Basu ◽  
Tahmoures Shabanian ◽  
Mitchell Craig Patrick ◽  
...  

AI evolution is accelerating and Deep Neural Network (DNN) inference accelerators are at the forefront of ad hoc architectures that are evolving to support the immense throughput required for AI computation. However, much more energy efficient design paradigms are inevitable to realize the complete potential of AI evolution and curtail energy consumption. The Near-Threshold Computing (NTC) design paradigm can serve as the best candidate for providing the required energy efficiency. However, NTC operation is plagued with ample performance and reliability concerns arising from the timing errors. In this paper, we dive deep into DNN architecture to uncover some unique challenges and opportunities for operation in the NTC paradigm. By performing rigorous simulations in TPU systolic array, we reveal the severity of timing errors and its impact on inference accuracy at NTC. We analyze various attributes—such as data–delay relationship, delay disparity within arithmetic units, utilization pattern, hardware homogeneity, workload characteristics—and uncover unique localized and global techniques to deal with the timing errors in NTC.


Author(s):  
Hongwu Jiang ◽  
Shanshi Huang ◽  
Xiaochen Peng ◽  
Jian-Wei Su ◽  
Yen-Chi Chou ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document