An Energy-Efficient Embedded Deep Neural Network Processor for High Speed Visual Attention in Mobile Vision Recognition SoC

Author(s):  
Seongwook Park ◽  
Injoon Hong ◽  
Junyoung Park ◽  
Hoi-Jun Yoo
2019 ◽  
Vol 52 (24) ◽  
pp. 135-139 ◽  
Author(s):  
Yuanjie Zhang ◽  
Na Qin ◽  
Deqing Huang ◽  
Kaiwei Liang

2020 ◽  
Vol 10 (4) ◽  
pp. 33
Author(s):  
Pramesh Pandey ◽  
Noel Daniel Gundi ◽  
Prabal Basu ◽  
Tahmoures Shabanian ◽  
Mitchell Craig Patrick ◽  
...  

AI evolution is accelerating and Deep Neural Network (DNN) inference accelerators are at the forefront of ad hoc architectures that are evolving to support the immense throughput required for AI computation. However, much more energy efficient design paradigms are inevitable to realize the complete potential of AI evolution and curtail energy consumption. The Near-Threshold Computing (NTC) design paradigm can serve as the best candidate for providing the required energy efficiency. However, NTC operation is plagued with ample performance and reliability concerns arising from the timing errors. In this paper, we dive deep into DNN architecture to uncover some unique challenges and opportunities for operation in the NTC paradigm. By performing rigorous simulations in TPU systolic array, we reveal the severity of timing errors and its impact on inference accuracy at NTC. We analyze various attributes—such as data–delay relationship, delay disparity within arithmetic units, utilization pattern, hardware homogeneity, workload characteristics—and uncover unique localized and global techniques to deal with the timing errors in NTC.


Photonics ◽  
2021 ◽  
Vol 8 (9) ◽  
pp. 363
Author(s):  
Qi Zhang ◽  
Zhuangzhuang Xing ◽  
Duan Huang

We demonstrate a pruned high-speed and energy-efficient optical backpropagation (BP) neural network. The micro-ring resonator (MRR) banks, as the core of the weight matrix operation, are used for large-scale weighted summation. We find that tuning a pruned MRR weight banks model gives an equivalent performance in training with the model of random initialization. Results show that the overall accuracy of the optical neural network on the MNIST dataset is 93.49% after pruning six-layer MRR weight banks on the condition of low insertion loss. This work is scalable to much more complex networks, such as convolutional neural networks and recurrent neural networks, and provides a potential guide for truly large-scale optical neural networks.


2021 ◽  
pp. 1-1
Author(s):  
Noriaki Kaneda ◽  
Chun-Yen Chuang ◽  
Ziyi Zhu ◽  
Amitkumar Mahadevan ◽  
Bob Farah ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document