scholarly journals High-Performance FPGA-based Accelerator for Bayesian Neural Networks

Author(s):  
Hongxiang Fan ◽  
Martin Ferianc ◽  
Miguel Rodrigues ◽  
Hongyu Zhou ◽  
Xinyu Niu ◽  
...  
Entropy ◽  
2021 ◽  
Vol 23 (2) ◽  
pp. 223
Author(s):  
Yen-Ling Tai ◽  
Shin-Jhe Huang ◽  
Chien-Chang Chen ◽  
Henry Horng-Shing Lu

Nowadays, deep learning methods with high structural complexity and flexibility inevitably lean on the computational capability of the hardware. A platform with high-performance GPUs and large amounts of memory could support neural networks having large numbers of layers and kernels. However, naively pursuing high-cost hardware would probably drag the technical development of deep learning methods. In the article, we thus establish a new preprocessing method to reduce the computational complexity of the neural networks. Inspired by the band theory of solids in physics, we map the image space into a noninteraction physical system isomorphically and then treat image voxels as particle-like clusters. Then, we reconstruct the Fermi–Dirac distribution to be a correction function for the normalization of the voxel intensity and as a filter of insignificant cluster components. The filtered clusters at the circumstance can delineate the morphological heterogeneity of the image voxels. We used the BraTS 2019 datasets and the dimensional fusion U-net for the algorithmic validation, and the proposed Fermi–Dirac correction function exhibited comparable performance to other employed preprocessing methods. By comparing to the conventional z-score normalization function and the Gamma correction function, the proposed algorithm can save at least 38% of computational time cost under a low-cost hardware architecture. Even though the correction function of global histogram equalization has the lowest computational time among the employed correction functions, the proposed Fermi–Dirac correction function exhibits better capabilities of image augmentation and segmentation.


Electronics ◽  
2021 ◽  
Vol 10 (14) ◽  
pp. 1614
Author(s):  
Jonghun Jeong ◽  
Jong Sung Park ◽  
Hoeseok Yang

Recently, the necessity to run high-performance neural networks (NN) is increasing even in resource-constrained embedded systems such as wearable devices. However, due to the high computational and memory requirements of the NN applications, it is typically infeasible to execute them on a single device. Instead, it has been proposed to run a single NN application cooperatively on top of multiple devices, a so-called distributed neural network. In the distributed neural network, workloads of a single big NN application are distributed over multiple tiny devices. While the computation overhead could effectively be alleviated by this approach, the existing distributed NN techniques, such as MoDNN, still suffer from large traffics between the devices and vulnerability to communication failures. In order to get rid of such big communication overheads, a knowledge distillation based distributed NN, called Network of Neural Networks (NoNN), was proposed, which partitions the filters in the final convolutional layer of the original NN into multiple independent subsets and derives smaller NNs out of each subset. However, NoNN also has limitations in that the partitioning result may be unbalanced and it considerably compromises the correlation between filters in the original NN, which may result in an unacceptable accuracy degradation in case of communication failure. In this paper, in order to overcome these issues, we propose to enhance the partitioning strategy of NoNN in two aspects. First, we enhance the redundancy of the filters that are used to derive multiple smaller NNs by means of averaging to increase the immunity of the distributed NN to communication failure. Second, we propose a novel partitioning technique, modified from Eigenvector-based partitioning, to preserve the correlation between filters as much as possible while keeping the consistent number of filters distributed to each device. Throughout extensive experiments with the CIFAR-100 (Canadian Institute For Advanced Research-100) dataset, it has been observed that the proposed approach maintains high inference accuracy (over 70%, 1.53× improvement over the state-of-the-art approach), on average, even when a half of eight devices in a distributed NN fail to deliver their partial inference results.


Author(s):  
Withit Chatlatanagulchai ◽  
Peter H. Meckl

Flexibility at the joint of a manipulator is an intrinsic property. Even “rigid-joint” robots, in fact, possess a certain amount of flexibility. Previous experiments confirmed that joint flexibility should be explicitly included in the model when designing a high-performance controller for a manipulator because the flexibility, if not dealt with, can excite system natural frequencies and cause severe damage. However, control design for a flexible-joint robot manipulator is still an open problem. Besides being described by a complicated system model for which the passivity property does not hold, the manipulator is also underactuated, that is, the control input does not drive the link directly, but through the flexible dynamics. Our work offers another possible solution to this open problem. We use three-layer neural networks to represent the system model. Their weights are adapted in real time and from scratch, which means we do not need the mathematical model of the robot in our control algorithm. All uncertainties are handled by variable-structure control. Backstepping structure allows input efforts to be applied to each subsystem where they are needed. Control laws to adjust all adjustable parameters are devised using Lyapunov’s second method to ensure that error trajectories are globally uniformly ultimately bounded. We present two state-feedback schemes: first, when neural networks are used to represent the unknown plant, and second, when neural networks are used to represent the unknown parts of the control laws. In the former case, we also design an observer to enable us to design a control law using only output signals—the link positions. We use simulations to compare our algorithms with some other well-known techniques. We use experiments to demonstrate the practicality of our algorithms.


2021 ◽  
pp. 1-11
Author(s):  
Oscar Herrera ◽  
Belém Priego

Traditionally, a few activation functions have been considered in neural networks, including bounded functions such as threshold, sigmoidal and hyperbolic-tangent, as well as unbounded ReLU, GELU, and Soft-plus, among other functions for deep learning, but the search for new activation functions still being an open research area. In this paper, wavelets are reconsidered as activation functions in neural networks and the performance of Gaussian family wavelets (first, second and third derivatives) are studied together with other functions available in Keras-Tensorflow. Experimental results show how the combination of these activation functions can improve the performance and supports the idea of extending the list of activation functions to wavelets which can be available in high performance platforms.


2021 ◽  
Author(s):  
Jiaru Zhang ◽  
Yang Hua ◽  
Zhengui Xue ◽  
Tao Song ◽  
Chengyu Zheng ◽  
...  

2018 ◽  
Vol 2018 ◽  
pp. 1-16 ◽  
Author(s):  
Shubin Zheng ◽  
Qianwen Zhong ◽  
Lele Peng ◽  
Xiaodong Chai

Electricity load forecasting is becoming one of the key issues to solve energy crisis problem, and time-series Bayesian Neural Network is one popular method used in load forecast models. However, it has long running time and relatively strong dependence on time and weather factors at a residential level. To solve these problems, this article presents an improved Bayesian Neural Networks (IBNN) forecast model by augmenting historical load data as inputs based on simple feedforward structure. From the load time delays correlations and impact factors analysis, containing different inputs, number of hidden neurons, historic period of data, forecasting time range, and range requirement of sample data, some advices are given on how to better choose these factors. To validate the performance of improved Bayesian Neural Networks model, several residential sample datasets of one whole year from Ausgrid have been selected to build the improved Bayesian Neural Networks model. The results compared with the time-series load forecast model show that the improved Bayesian Neural Networks model can significantly reduce calculating time by more than 30 times and even when the time or meteorological factors are missing, it can still predict the load with a high accuracy. Compared with other widely used prediction methods, the IBNN also performs a better accuracy and relatively shorter computing time. This improved Bayesian Neural Networks forecasting method can be applied in residential energy management.


Sign in / Sign up

Export Citation Format

Share Document