adaptive parallelism
Recently Published Documents


TOTAL DOCUMENTS

28
(FIVE YEARS 2)

H-INDEX

6
(FIVE YEARS 0)

Electronics ◽  
2021 ◽  
Vol 10 (3) ◽  
pp. 230
Author(s):  
Jaechan Cho ◽  
Yongchul Jung ◽  
Seongjoo Lee ◽  
Yunho Jung

Binary neural networks (BNNs) have attracted significant interest for the implementation of deep neural networks (DNNs) on resource-constrained edge devices, and various BNN accelerator architectures have been proposed to achieve higher efficiency. BNN accelerators can be divided into two categories: streaming and layer accelerators. Although streaming accelerators designed for a specific BNN network topology provide high throughput, they are infeasible for various sensor applications in edge AI because of their complexity and inflexibility. In contrast, layer accelerators with reasonable resources can support various network topologies, but they operate with the same parallelism for all the layers of the BNN, which degrades throughput performance at certain layers. To overcome this problem, we propose a BNN accelerator with adaptive parallelism that offers high throughput performance in all layers. The proposed accelerator analyzes target layer parameters and operates with optimal parallelism using reasonable resources. In addition, this architecture is able to fully compute all types of BNN layers thanks to its reconfigurability, and it can achieve a higher area–speed efficiency than existing accelerators. In performance evaluation using state-of-the-art BNN topologies, the designed BNN accelerator achieved an area–speed efficiency 9.69 times higher than previous FPGA implementations and 24% higher than existing VLSI implementations for BNNs.


2018 ◽  
Author(s):  
Florian Spenke ◽  
Karsten Balzer ◽  
Sascha Frick ◽  
Bernd Hartke ◽  
Johannes M. Dieterich

A new parallel high-performance computing setup, which can use every little bit of computing resources left over by traditional scheduling, regardless how small or big it may be. This enables HPC providers to achieve 100 percent load on their machines at all times, and it enables HPC users to get substantial computing time on HPC systems that are "full" with traditional jobs.<br>


Author(s):  
Jiayuan Meng ◽  
Thomas D. Uram ◽  
Vitali Morozov ◽  
Venkatram Vishwanath ◽  
Kalyan Kumaran
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document