Fine-tuning CLB placement to speed up reconfigurations in NVM-based FPGAs

Author(s):  
Yuan Xue ◽  
Patrick Cronin ◽  
Chengmo Yang ◽  
Jingtong Hu
Keyword(s):  
2020 ◽  
Vol 34 (05) ◽  
pp. 7839-7846
Author(s):  
Junliang Guo ◽  
Xu Tan ◽  
Linli Xu ◽  
Tao Qin ◽  
Enhong Chen ◽  
...  

Non-autoregressive translation (NAT) models remove the dependence on previous target tokens and generate all target tokens in parallel, resulting in significant inference speedup but at the cost of inferior translation accuracy compared to autoregressive translation (AT) models. Considering that AT models have higher accuracy and are easier to train than NAT models, and both of them share the same model configurations, a natural idea to improve the accuracy of NAT models is to transfer a well-trained AT model to an NAT model through fine-tuning. However, since AT and NAT models differ greatly in training strategy, straightforward fine-tuning does not work well. In this work, we introduce curriculum learning into fine-tuning for NAT. Specifically, we design a curriculum in the fine-tuning process to progressively switch the training from autoregressive generation to non-autoregressive generation. Experiments on four benchmark translation datasets show that the proposed method achieves good improvement (more than 1 BLEU score) over previous NAT baselines in terms of translation accuracy, and greatly speed up (more than 10 times) the inference process over AT baselines.


2022 ◽  
Vol 2022 ◽  
pp. 1-11
Author(s):  
Zijin Wu

With the development of the country’s economy, there is a flourishing situation in the field of culture and art. However, the diversification of artistic expressions has not brought development to folk music. On the contrary, it brought a huge impact, and some national music even fell into the dilemma of being lost. This article is mainly aimed at the recognition and classification of folk music emotions and finds the model that can make the classification accuracy rate as high as possible. The classification model used in this article is mainly after determining the use of Support Vector Machine (SVM) classification method, a variety of attempts have been made to feature extraction, and good results have been achieved. Explore the Deep Belief Network (DBN) pretraining and reverse fine-tuning process, using DBN to learn the fusion characteristics of music. According to the abstract characteristics learned by them, the recognition and classification of folk music emotions are carried out. The DBN is improved by adding “Dropout” to each Restricted Boltzmann Machine (RBM) and adjusting the increase standard of weight and bias. The improved network can avoid the overfitting problem and speed up the training of the network. Through experiments, it is found that using the fusion features proposed in this paper, through classification, the classification accuracy has been improved.


2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Yisu Ge ◽  
Shufang Lu ◽  
Fei Gao

Many current convolutional neural networks are hard to meet the practical application requirement because of the enormous network parameters. For accelerating the inference speed of networks, more and more attention has been paid to network compression. Network pruning is one of the most efficient and simplest ways to compress and speed up the networks. In this paper, a pruning algorithm for the lightweight task is proposed, and a pruning strategy based on feature representation is investigated. Different from other pruning approaches, the proposed strategy is guided by the practical task and eliminates the irrelevant filters in the network. After pruning, the network is compacted to a smaller size and is easy to recover accuracy with fine-tuning. The performance of the proposed pruning algorithm is validated on the acknowledged image datasets, and the experimental results prove that the proposed algorithm is more suitable to prune the irrelevant filters for the fine-tuning dataset.


1998 ◽  
Vol 4 (2) ◽  
pp. 311-316
Author(s):  
Enzo Paci

The activities of WTO are focused on promoting a creative approach by National Tourism Administrations, Statistical Offices and local authorities to encourage countries to collect more reliable and more complete tourism statistics in line with WTO definitions, so as to improve their international comparability. WTO also emphasizes the need to speed up the production and publication of these statistics at country level in order to provide the means of identifying tourism trends by month and fine-tuning promotion and marketing policies. Computerization and the successful effort to develop standard definitions and classifications for tourism have given renewed force to WTO's work in statistics. WTO has expanded activities with Member States to implement the WTO Recommendations on Tourism Statistics, adopted by the United Nations Statistical Commission in 1993, through manuals, seminars and an ambitious statistical development programme to assess the economic importance of tourism. The programme includes the holding of a World Conference on the measurement of the economic impact o tourism in Nice (France) towards the end of May/beginning of June 1999. The objective of the Conference is to develop a core of indicators for the assessment of the net economic impact of tourism at both national and international level, thereby enhancing the credibility of the tourism industry.


Author(s):  
Weijie Chen ◽  
Yuan Zhang ◽  
Di Xie ◽  
Shiliang Pu

Neuron pruning is an efficient method to compress the network into a slimmer one for reducing the computational cost and storage overhead. Most of state-of-the-art results are obtained in a layer-by-layer optimization mode. It discards the unimportant input neurons and uses the survived ones to reconstruct the output neurons approaching to the original ones in a layer-by-layer manner. However, an unnoticed problem arises that the information loss is accumulated as layer increases since the survived neurons still do not encode the entire information as before. A better alternative is to propagate the entire useful information to reconstruct the pruned layer instead of directly discarding the less important neurons. To this end, we propose a novel Layer DecompositionRecomposition Framework (LDRF) for neuron pruning, by which each layer’s output information is recovered in an embedding space and then propagated to reconstruct the following pruned layers with useful information preserved. We mainly conduct our experiments on ILSVRC-12 benchmark with VGG-16 and ResNet-50. What should be emphasized is that our results before end-to-end fine-tuning are significantly superior owing to the information-preserving property of our proposed framework. With end-to-end fine-tuning, we achieve state-of-the-art results of 5.13× and 3× speed-up with only 0.5% and 0.65% top-5 accuracy drop respectively, which outperform the existing neuron pruning methods.


1998 ◽  
Vol 4 (1) ◽  
pp. 3-9
Author(s):  
Enzo Paci

The activities of WTO are focused on promoting a creative approach by National Tourism Administrations, Statistical Offices and local authorities to encourage countries to collect more reliable and more complete tourism statistics in line with WTO definitions, so as to improve their international comparability. WTO also emphasizes the need to speed up the production and publication of these statistics at country level in order to provide the means of identifying tourism trends by month and fine-tuning promotion and marketing policies. Computerization and the successful effort to develop standard definitions and classifications for tourism have given renewed force to WTO’s work in statistics. WTO has expanded activities with Member States to implement the WTO Recommendations on Tourism Statistics, adopted by the United Nations Statistical Commission in 1993, through manuals, seminars and an ambitious statistical development programme to assess the economic importance of tourism. The programme includes the holding of a World Conference on the measurement of the economic impact of tourism in Nice (France) towards the end of May/beginning of June 1999. The objective of the Conference is to develop a core of indicators for the assessment of the net economic impact of tourism at both national and international level, thereby enhancing the credibility of the tourism industry.


In the accident insurance industry, settling the claim is a time-consuming process since it is a manual process and there is a gap between the optimal and the actual settlement. Using deep learning models, we are not only trying to speed up the process but also provide better customer service and increase the profitability of insurance companies. In this paper we are using various pretrained models such as VGG 16, VGG 19, Resnet50 and Densenet and based on these models, selecting the best performing models. We initially check whether the car is damaged or not using the Resnet50 model and if it’s a damaged one we use the WPOD-net model to detect the license plate. To identify the damaged region, we use the YOLO model. At last, comes the damage severity which is implemented using the Densenet model. After implementing various models, we find out that transfer learning gives better results than fine-tuning. In addition to that we propose a framework that integrates all of this into one application and in turn helps in the automation of the insurance industry


2020 ◽  
Vol 6 ◽  
pp. e274
Author(s):  
Maxim Borisyak ◽  
Tatiana Gaintseva ◽  
Andrey Ustyuzhanin

Adversarial Optimization provides a reliable, practical way to match two implicitly defined distributions, one of which is typically represented by a sample of real data, and the other is represented by a parameterized generator. Matching of the distributions is achieved by minimizing a divergence between these distribution, and estimation of the divergence involves a secondary optimization task, which, typically, requires training a model to discriminate between these distributions. The choice of the model has its trade-off: high-capacity models provide good estimations of the divergence, but, generally, require large sample sizes to be properly trained. In contrast, low-capacity models tend to require fewer samples for training; however, they might provide biased estimations. Computational costs of Adversarial Optimization becomes significant when sampling from the generator is expensive. One of the practical examples of such settings is fine-tuning parameters of complex computer simulations. In this work, we introduce a novel family of divergences that enables faster optimization convergence measured by the number of samples drawn from the generator. The variation of the underlying discriminator model capacity during optimization leads to a significant speed-up. The proposed divergence family suggests using low-capacity models to compare distant distributions (typically, at early optimization steps), and the capacity gradually grows as the distributions become closer to each other. Thus, it allows for a significant acceleration of the initial stages of optimization. This acceleration was demonstrated on two fine-tuning problems involving Pythia event generator and two of the most popular black-box optimization algorithms: Bayesian Optimization and Variational Optimization. Experiments show that, given the same budget, adaptive divergences yield results up to an order of magnitude closer to the optimum than Jensen-Shannon divergence. While we consider physics-related simulations, adaptive divergences can be applied to any stochastic simulation.


Author(s):  
Brian Cross

A relatively new entry, in the field of microscopy, is the Scanning X-Ray Fluorescence Microscope (SXRFM). Using this type of instrument (e.g. Kevex Omicron X-ray Microprobe), one can obtain multiple elemental x-ray images, from the analysis of materials which show heterogeneity. The SXRFM obtains images by collimating an x-ray beam (e.g. 100 μm diameter), and then scanning the sample with a high-speed x-y stage. To speed up the image acquisition, data is acquired "on-the-fly" by slew-scanning the stage along the x-axis, like a TV or SEM scan. To reduce the overhead from "fly-back," the images can be acquired by bi-directional scanning of the x-axis. This results in very little overhead with the re-positioning of the sample stage. The image acquisition rate is dominated by the x-ray acquisition rate. Therefore, the total x-ray image acquisition rate, using the SXRFM, is very comparable to an SEM. Although the x-ray spatial resolution of the SXRFM is worse than an SEM (say 100 vs. 2 μm), there are several other advantages.


Author(s):  
A. G. Jackson ◽  
M. Rowe

Diffraction intensities from intermetallic compounds are, in the kinematic approximation, proportional to the scattering amplitude from the element doing the scattering. More detailed calculations have shown that site symmetry and occupation by various atom species also affects the intensity in a diffracted beam. [1] Hence, by measuring the intensities of beams, or their ratios, the occupancy can be estimated. Measurement of the intensity values also allows structure calculations to be made to determine the spatial distribution of the potentials doing the scattering. Thermal effects are also present as a background contribution. Inelastic effects such as loss or absorption/excitation complicate the intensity behavior, and dynamical theory is required to estimate the intensity value.The dynamic range of currents in diffracted beams can be 104or 105:1. Hence, detection of such information requires a means for collecting the intensity over a signal-to-noise range beyond that obtainable with a single film plate, which has a S/N of about 103:1. Although such a collection system is not available currently, a simple system consisting of instrumentation on an existing STEM can be used as a proof of concept which has a S/N of about 255:1, limited by the 8 bit pixel attributes used in the electronics. Use of 24 bit pixel attributes would easily allowthe desired noise range to be attained in the processing instrumentation. The S/N of the scintillator used by the photoelectron sensor is about 106 to 1, well beyond the S/N goal. The trade-off that must be made is the time for acquiring the signal, since the pattern can be obtained in seconds using film plates, compared to 10 to 20 minutes for a pattern to be acquired using the digital scan. Parallel acquisition would, of course, speed up this process immensely.


Sign in / Sign up

Export Citation Format

Share Document