systolic array
Recently Published Documents


TOTAL DOCUMENTS

931
(FIVE YEARS 78)

H-INDEX

30
(FIVE YEARS 3)

2021 ◽  
Author(s):  
Geonhwa Jeong ◽  
Eric Qin ◽  
Ananda Samajdar ◽  
Christopher J. Hughes ◽  
Sreenivas Subramoney ◽  
...  
Keyword(s):  

2021 ◽  
Author(s):  
Mehdi Safarpour ◽  
Reza Inanlou ◽  
Olli Silven

An energy efficient architecture for TPUs that is based on reduced voltage operation. The errors are captured and corrected by utilizing ABFT and hence aggressive voltage scaling is made possible.


2021 ◽  
Vol 20 (5s) ◽  
pp. 1-20
Author(s):  
Hyungmin Cho

Depthwise convolutions are widely used in convolutional neural networks (CNNs) targeting mobile and embedded systems. Depthwise convolution layers reduce the computation loads and the number of parameters compared to the conventional convolution layers. Many deep neural network (DNN) accelerators adopt an architecture that exploits the high data-reuse factor of DNN computations, such as a systolic array. However, depthwise convolutions have low data-reuse factor and under-utilize the processing elements (PEs) in systolic arrays. In this paper, we present a DNN accelerator design called RiSA, which provides a novel mechanism that boosts the PE utilization for depthwise convolutions on a systolic array with minimal overheads. In addition, the PEs in systolic arrays can be efficiently used only if the data items ( tensors ) are arranged in the desired layout. Typical DNN accelerators provide various types of PE interconnects or additional modules to flexibly rearrange the data items and manage data movements during DNN computations. RiSA provides a lightweight set of tensor management tasks within the PE array itself that eliminates the need for an additional module for tensor reshaping tasks. Using this embedded tensor reshaping, RiSA supports various DNN models, including convolutional neural networks and natural language processing models while maintaining a high area efficiency. Compared to Eyeriss v2, RiSA improves the area and energy efficiency for MobileNet-V1 inference by 1.91× and 1.31×, respectively.


2021 ◽  
Author(s):  
Jeong-Jun Lee ◽  
Jianhao Chen ◽  
Wenrui Zhang ◽  
Peng Li

Sign in / Sign up

Export Citation Format

Share Document