inference process
Recently Published Documents


TOTAL DOCUMENTS

183
(FIVE YEARS 77)

H-INDEX

18
(FIVE YEARS 4)

Entropy ◽  
2022 ◽  
Vol 24 (1) ◽  
pp. 125
Author(s):  
Damián G. Hernández ◽  
Inés Samengo

Inferring the value of a property of a large stochastic system is a difficult task when the number of samples is insufficient to reliably estimate the probability distribution. The Bayesian estimator of the property of interest requires the knowledge of the prior distribution, and in many situations, it is not clear which prior should be used. Several estimators have been developed so far in which the proposed prior us individually tailored for each property of interest; such is the case, for example, for the entropy, the amount of mutual information, or the correlation between pairs of variables. In this paper, we propose a general framework to select priors that is valid for arbitrary properties. We first demonstrate that only certain aspects of the prior distribution actually affect the inference process. We then expand the sought prior as a linear combination of a one-dimensional family of indexed priors, each of which is obtained through a maximum entropy approach with constrained mean values of the property under study. In many cases of interest, only one or very few components of the expansion turn out to contribute to the Bayesian estimator, so it is often valid to only keep a single component. The relevant component is selected by the data, so no handcrafted priors are required. We test the performance of this approximation with a few paradigmatic examples and show that it performs well in comparison to the ad-hoc methods previously proposed in the literature. Our method highlights the connection between Bayesian inference and equilibrium statistical mechanics, since the most relevant component of the expansion can be argued to be that with the right temperature.


2021 ◽  
Vol 13 (24) ◽  
pp. 5132
Author(s):  
Xiaolan Huang ◽  
Kai Xu ◽  
Chuming Huang ◽  
Chengrui Wang ◽  
Kun Qin

The key to fine-grained aircraft recognition is discovering the subtle traits that can distinguish different subcategories. Early approaches leverage part annotations of fine-grained objects to derive rich representations. However, manual labeling part information is cumbersome. In response to this issue, previous CNN-based methods reuse the backbone network to extract part-discrimination features, the inference process of which consumes much time. Therefore, we introduce generalized multiple instance learning (MIL) into fine-grained recognition. In generalized MIL, an aircraft is assumed to consist of multiple instances (such as head, tail, and body). Firstly, instance-level representations are obtained by the feature extractor and instance conversion component. Secondly, the obtained instance features are scored by an MIL classifier, which can yield high-level part semantics. Finally, a fine-grained object label is inferred by a MIL pooling function that aggregates multiple instance scores. The proposed approach is trained end-to-end without part annotations and complex location networks. Experimental evidence is conducted to prove the feasibility and effectiveness of our approach on combined aircraft images (CAIs).


Information ◽  
2021 ◽  
Vol 12 (12) ◽  
pp. 513
Author(s):  
Andreas Maniatopoulos ◽  
Nikolaos Mitianoudis

In neural networks, a vital component in the learning and inference process is the activation function. There are many different approaches, but only nonlinear activation functions allow such networks to compute non-trivial problems by using only a small number of nodes, and such activation functions are called nonlinearities. With the emergence of deep learning, the need for competent activation functions that can enable or expedite learning in deeper layers has emerged. In this paper, we propose a novel activation function, combining many features of successful activation functions, achieving 2.53% higher accuracy than the industry standard ReLU in a variety of test cases.


Mathematics ◽  
2021 ◽  
Vol 9 (23) ◽  
pp. 3130
Author(s):  
Bharathwaj Suresh ◽  
Kamlesh Pillai ◽  
Gurpreet Singh Kalsi ◽  
Avishaii Abuhatzera ◽  
Sreenivas Subramoney

Deep Neural Networks (DNNs) have set state-of-the-art performance numbers in diverse fields of electronics (computer vision, voice recognition), biology, bioinformatics, etc. However, the process of learning (training) from the data and application of the learnt information (inference) process requires huge computational resources. Approximate computing is a common method to reduce computation cost, but it introduces loss in task accuracy, which limits their application. Using an inherent property of Rectified Linear Unit (ReLU), a popular activation function, we propose a mathematical model to perform MAC operation using reduced precision for predicting negative values early. We also propose a method to perform hierarchical computation to achieve the same results as IEEE754 full precision compute. Applying this method on ResNet50 and VGG16 shows that up to 80% of ReLU zeros (which is 50% of all ReLU outputs) can be predicted and detected early by using just 3 out of 23 mantissa bits. This method is equally applicable to other floating-point representations.


2021 ◽  
Vol 51 (4) ◽  
pp. 107-141
Author(s):  
Paweł Lindstedt ◽  
Edward Rokicki ◽  
Maciej Deliś ◽  
Kamila Dobosz ◽  
Andrzej Czarnecki

Abstract In the machine operating process, there are certain interactions between its operational use and wear. The current wear is increased by the current intensity of operational use, and usable potential is reduced by the current wear rate. In the diagnostic inference process, static characteristics and trajectories of state from the experiment are compared with different matrices determined for various assumed configurations of changes. As a result, the approximated diagnosis is created. This method is not universal. It applies only to the continuous progressive state, more or less increased wear rate of the machine.


Author(s):  
Lu Xiang ◽  
Junnan Zhu ◽  
Yang Zhao ◽  
Yu Zhou ◽  
Chengqing Zong

Cross-lingual dialogue systems are increasingly important in e-commerce and customer service due to the rapid progress of globalization. In real-world system deployment, machine translation (MT) services are often used before and after the dialogue system to bridge different languages. However, noises and errors introduced in the MT process will result in the dialogue system's low robustness, making the system's performance far from satisfactory. In this article, we propose a novel MT-oriented noise enhanced framework that exploits multi-granularity MT noises and injects such noises into the dialogue system to improve the dialogue system's robustness. Specifically, we first design a method to automatically construct multi-granularity MT-oriented noises and multi-granularity adversarial examples, which contain abundant noise knowledge oriented to MT. Then, we propose two strategies to incorporate the noise knowledge: (i) Utterance-level adversarial learning and (ii) Knowledge-level guided method. The former adopts adversarial learning to learn a perturbation-invariant encoder, guiding the dialogue system to learn noise-independent hidden representations. The latter explicitly incorporates the multi-granularity noises, which contain the noise tokens and their possible correct forms, into the training and inference process, thus improving the dialogue system's robustness. Experimental results on three dialogue models, two dialogue datasets, and two language pairs have shown that the proposed framework significantly improves the performance of the cross-lingual dialogue system.


2021 ◽  
Author(s):  
Zeyu Wang ◽  
Ziqun Zhou ◽  
Haibin Shen ◽  
Qi Xu ◽  
Kejie Huang

<div>Electroencephalography (EEG) emotion recognition, an important task in Human-Computer Interaction (HCI), has made a great breakthrough with the help of deep learning algorithms. Although the application of attention mechanism on conventional models has improved its performance, most previous research rarely focused on multiplex EEG features jointly, lacking a compact model with unified attention modules. This study proposes Joint-Dimension-Aware Transformer (JDAT), a robust model based on squeezed Multi-head Self-Attention (MSA) mechanism for EEG emotion recognition. The adaptive squeezed MSA applied on multidimensional features enables JDAT to focus on diverse EEG information, including space, frequency, and time. Under the joint attention, JDAT is sensitive to the complicated brain activities, such as signal activation, phase-intensity couplings, and resonance. Moreover, its gradually compressed structure contains no recurrent or parallel modules, greatly reducing the memory and complexity, and accelerating the inference process. The proposed JDAT is evaluated on DEAP, DREAMER, and SEED datasets, and experimental results show that it outperforms state-of-the-art methods along with stronger flexibility.</div>


2021 ◽  
Author(s):  
Zeyu Wang ◽  
Ziqun Zhou ◽  
Haibin Shen ◽  
Qi Xu ◽  
Kejie Huang

<div>Electroencephalography (EEG) emotion recognition, an important task in Human-Computer Interaction (HCI), has made a great breakthrough with the help of deep learning algorithms. Although the application of attention mechanism on conventional models has improved its performance, most previous research rarely focused on multiplex EEG features jointly, lacking a compact model with unified attention modules. This study proposes Joint-Dimension-Aware Transformer (JDAT), a robust model based on squeezed Multi-head Self-Attention (MSA) mechanism for EEG emotion recognition. The adaptive squeezed MSA applied on multidimensional features enables JDAT to focus on diverse EEG information, including space, frequency, and time. Under the joint attention, JDAT is sensitive to the complicated brain activities, such as signal activation, phase-intensity couplings, and resonance. Moreover, its gradually compressed structure contains no recurrent or parallel modules, greatly reducing the memory and complexity, and accelerating the inference process. The proposed JDAT is evaluated on DEAP, DREAMER, and SEED datasets, and experimental results show that it outperforms state-of-the-art methods along with stronger flexibility.</div>


2021 ◽  
Author(s):  
Yuki Kobayashi

Murray (2020) recently introduced a novel computational lightness model, Markov Illuminance and Reflectance (MIR), a Bayesian observer model that represents input information and prior assumption with conditional random field (CRF) and that can account for many lightness illusions and phenomena. In the original MIR’s inference process, however, it did not utilize all the links in its CRF. Thus, this letter reports that a simple modification to the original MIR’s inference process improves its performance. MIR is a highly extensible model, so I recommend future research use the proposed version to attain further sophistication.


2021 ◽  
Vol 10 (5) ◽  
pp. 416-423
Author(s):  
Ukyo Yoshimura ◽  
Toshiyuki Inoue ◽  
Akira Tsuchiya ◽  
Keiji Kishine

Sign in / Sign up

Export Citation Format

Share Document