Towards Verification-Aware Knowledge Distillation for Neural-Network Controlled Systems: Invited Paper

Author(s):  
Jiameng Fan ◽  
Chao Huang ◽  
Wenchao Li ◽  
Xin Chen ◽  
Qi Zhu
Electronics ◽  
2021 ◽  
Vol 10 (14) ◽  
pp. 1614
Author(s):  
Jonghun Jeong ◽  
Jong Sung Park ◽  
Hoeseok Yang

Recently, the necessity to run high-performance neural networks (NN) is increasing even in resource-constrained embedded systems such as wearable devices. However, due to the high computational and memory requirements of the NN applications, it is typically infeasible to execute them on a single device. Instead, it has been proposed to run a single NN application cooperatively on top of multiple devices, a so-called distributed neural network. In the distributed neural network, workloads of a single big NN application are distributed over multiple tiny devices. While the computation overhead could effectively be alleviated by this approach, the existing distributed NN techniques, such as MoDNN, still suffer from large traffics between the devices and vulnerability to communication failures. In order to get rid of such big communication overheads, a knowledge distillation based distributed NN, called Network of Neural Networks (NoNN), was proposed, which partitions the filters in the final convolutional layer of the original NN into multiple independent subsets and derives smaller NNs out of each subset. However, NoNN also has limitations in that the partitioning result may be unbalanced and it considerably compromises the correlation between filters in the original NN, which may result in an unacceptable accuracy degradation in case of communication failure. In this paper, in order to overcome these issues, we propose to enhance the partitioning strategy of NoNN in two aspects. First, we enhance the redundancy of the filters that are used to derive multiple smaller NNs by means of averaging to increase the immunity of the distributed NN to communication failure. Second, we propose a novel partitioning technique, modified from Eigenvector-based partitioning, to preserve the correlation between filters as much as possible while keeping the consistent number of filters distributed to each device. Throughout extensive experiments with the CIFAR-100 (Canadian Institute For Advanced Research-100) dataset, it has been observed that the proposed approach maintains high inference accuracy (over 70%, 1.53× improvement over the state-of-the-art approach), on average, even when a half of eight devices in a distributed NN fail to deliver their partial inference results.


1994 ◽  
Vol 27 (8) ◽  
pp. 605-610
Author(s):  
K. Kumamaru ◽  
K. Inoue ◽  
S. Nonaka ◽  
H. Ono ◽  
T. Söderström

2021 ◽  
Author(s):  
O.V. Druzhinina ◽  
E.R. Korepanov ◽  
V.V. Belousov ◽  
O.N. Masina ◽  
A.A. Petrov

The development of tools for solving research problems with the use of domestic software and hardware is an urgent direction. Such tasks include the tasks of neural network modeling of nonlinear controlled systems. The paper provides an extended analysis of the capabilities of the Elbrus architecture and the blocks of the built-in EML library for mathematical modeling of nonlinear systems. A comparative analysis of the instrumentation and efficiency of computational experiments is performed, taking into account the use of an 8-core processor and the potential capabilities of a 16-core processor. The specifics of the EML library blocks in relation to solving specific types of scientific problems is considered and the optimized software is analyzed. The design of generalized models of nonlinear systems with switching is proposed. For generalized models, a new switching algorithm has been developed that can be adapted to the Elbrus computing platform. An algorithmic tree is constructed, and algorithmic and software are developed for the study of models with switching. The results of adaptation of the modules of the software package for modeling managed systems to the elements of the platform are presented. The results of computer modeling of nonlinear systems based on the Elbrus 801-RS computing platform are systematized and generalized. The results can be used in problems of creating algorithmic and software for solving research modeling problems, in problems of synthesis and analysis of models of controlled technical systems with switching modes of operation, as well as in problems of neural network modeling and machine learning.


2019 ◽  
Vol 9 (16) ◽  
pp. 3396 ◽  
Author(s):  
Jianfeng Wu ◽  
Yongzhu Hua ◽  
Shengying Yang ◽  
Hongshuai Qin ◽  
Huibin Qin

This paper presents a new deep neural network (DNN)-based speech enhancement algorithm by integrating the distilled knowledge from the traditional statistical-based method. Unlike the other DNN-based methods, which usually train many different models on the same data and then average their predictions, or use a large number of noise types to enlarge the simulated noisy speech, the proposed method does not train a whole ensemble of models and does not require a mass of simulated noisy speech. It first trains a discriminator network and a generator network simultaneously using the adversarial learning method. Then, the discriminator network and generator network are re-trained by distilling knowledge from the statistical method, which is inspired by the knowledge distillation in a neural network. Finally, the generator network is fine-tuned using real noisy speech. Experiments on CHiME4 data sets demonstrate that the proposed method achieves a more robust performance than the compared DNN-based method in terms of perceptual speech quality.


Sign in / Sign up

Export Citation Format

Share Document