scholarly journals Classification with Label Distribution Learning

Author(s):  
Jing Wang ◽  
Xin Geng

Label Distribution Learning (LDL) is a novel learning paradigm, aim of which is to minimize the distance between the model output and the ground-truth label distribution. We notice that, in real-word applications, the learned label distribution model is generally treated as a classification model, with the label corresponding to the highest model output as the predicted label, which unfortunately prompts an inconsistency between the training phrase and the test phrase. To solve the inconsistency, we propose in this paper a new Label Distribution Learning algorithm for Classification (LDL4C). Firstly, instead of KL-divergence, absolute loss is applied as the measure for LDL4C. Secondly, samples are re-weighted with information entropy. Thirdly, large margin classifier is adapted to boost discrimination precision. We then reveal that theoretically LDL4C seeks a balance between generalization and discrimination. Finally, we compare LDL4C with existing LDL algorithms on 17 real-word datasets, and experimental results demonstrate the effectiveness of LDL4C in classification.

Author(s):  
Tingting Ren ◽  
Xiuyi Jia ◽  
Weiwei Li ◽  
Shu Zhao

Label distribution learning (LDL) can be viewed as the generalization of multi-label learning. This novel paradigm focuses on the relative importance of different labels to a particular instance. Most previous LDL methods either ignore the correlation among labels, or only exploit the label correlations in a global way. In this paper, we utilize both the global and local relevance among labels to provide more information for training model and propose a novel label distribution learning algorithm. In particular, a label correlation matrix based on low-rank approximation is applied to capture the global label correlations. In addition, the label correlation among local samples are adopted to modify the label correlation matrix. The experimental results on real-world data sets show that the proposed algorithm outperforms state-of-the-art LDL methods.


Author(s):  
Wenfang Zhu ◽  
Xiuyi Jia ◽  
Weiwei Li

Label distribution learning has attracted more and more attention in view of its more generalized ability to express the label ambiguity. However, it is much more expensive to obtain the label distribution information of the data rather than the logical labels. Thus, label enhancement is proposed to recover the label distributions from the logical labels. In this paper, we propose a novel label enhancement method by using privileged information. We first apply a multi-label learning model to implicitly capture the complex structural information between instances and generate the privileged information. Second, we adopt LUPI (learning with privileged information) paradigm to utilize the privileged information and employ RSVM+ as the prediction model. Finally, comparison experiments on 12 datasets demonstrate that our proposal can better fit the ground-truth label distributions.


2021 ◽  
Author(s):  
Mofei Song ◽  
Han Xu ◽  
Xiao Fan Liu ◽  
Qian Li

This paper proposes an image-based visibility estimation method with deep label distribution learning. To train an accurate model for visibility estimation, it is important to obtain the precise ground truth for every image. However, the ground-truth visibility is difficult to be labeled due to its high ambiguity. To solve this problem, we associate a label distribution to each image. The label distribution contains all the possible visibilities with their probabilities. To learn from such annotation, we employ a CNN-RNN model for visibility-aware feature extraction and a conditional probability neural network for distribution prediction. Our experiment shows that labeling the image with visibility distribution can not only overcome the inaccurate annotation problem, but also boost the learning performance without the increase of training examples.


2021 ◽  
Author(s):  
Mofei Song ◽  
Han Xu ◽  
Xiao Fan Liu ◽  
Qian Li

This paper proposes an image-based visibility estimation method with deep label distribution learning. To train an accurate model for visibility estimation, it is important to obtain the precise ground truth for every image. However, the ground-truth visibility is difficult to be labeled due to its high ambiguity. To solve this problem, we associate a label distribution to each image. The label distribution contains all the possible visibilities with their probabilities. To learn from such annotation, we employ a CNN-RNN model for visibility-aware feature extraction and a conditional probability neural network for distribution prediction. Our experiment shows that labeling the image with visibility distribution can not only overcome the inaccurate annotation problem, but also boost the learning performance without the increase of training examples.


Energies ◽  
2021 ◽  
Vol 14 (11) ◽  
pp. 3117
Author(s):  
Junghwan Kim

Engine knock determination has been conducted in various ways for spark timing calibration. In the present study, a knock classification model was developed using a machine learning algorithm. Wavelet packet decomposition (WPD) and ensemble empirical mode decomposition (EEMD) were employed for the characterization of the in-cylinder pressure signals from the experimental engine. The WPD was used to calculate 255 features from seven decomposition levels. EEMD provided total 70 features from their intrinsic mode functions (IMF). The experimental engine was operated at advanced spark timings to induce knocking under various engine speeds and load conditions. Three knock intensity metrics were employed to determine that the dataset included 4158 knock cycles out of a total of 66,000 cycles. The classification model trained with 66,000 cycles achieved an accuracy of 99.26% accuracy in the knock cycle detection. The neighborhood component analysis revealed that seven features contributed significantly to the classification. The classification model retrained with the seven significant features achieved an accuracy of 99.02%. Although the misclassification rate increased in the normal cycle detection, the feature selection decreased the model size from 253 to 8.25 MB. Finally, the compact classification model achieved an accuracy of 99.95% with the second dataset obtained at the knock borderline (KBL) timings, which validates that the model is sufficient for the KBL timing determination.


2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Peter M. Maloca ◽  
Philipp L. Müller ◽  
Aaron Y. Lee ◽  
Adnan Tufail ◽  
Konstantinos Balaskas ◽  
...  

AbstractMachine learning has greatly facilitated the analysis of medical data, while the internal operations usually remain intransparent. To better comprehend these opaque procedures, a convolutional neural network for optical coherence tomography image segmentation was enhanced with a Traceable Relevance Explainability (T-REX) technique. The proposed application was based on three components: ground truth generation by multiple graders, calculation of Hamming distances among graders and the machine learning algorithm, as well as a smart data visualization (‘neural recording’). An overall average variability of 1.75% between the human graders and the algorithm was found, slightly minor to 2.02% among human graders. The ambiguity in ground truth had noteworthy impact on machine learning results, which could be visualized. The convolutional neural network balanced between graders and allowed for modifiable predictions dependent on the compartment. Using the proposed T-REX setup, machine learning processes could be rendered more transparent and understandable, possibly leading to optimized applications.


2021 ◽  
Vol 13 (15) ◽  
pp. 2935
Author(s):  
Chunhua Qian ◽  
Hequn Qiang ◽  
Feng Wang ◽  
Mingyang Li

Building a high-precision, stable, and universal automatic extraction model of the rocky desertification information is the premise for exploring the spatiotemporal evolution of rocky desertification. Taking Guizhou province as the research area and based on MODIS and continuous forest inventory data in China, we used a machine learning algorithm to build a rocky desertification model with bedrock exposure rate, temperature difference, humidity, and other characteristic factors and considered improving the model accuracy from the spatial and temporal dimensions. The results showed the following: (1) The supervised classification method was used to build a rocky desertification model, and the logical model, RF model, and SVM model were constructed separately. The accuracies of the models were 73.8%, 78.2%, and 80.6%, respectively, and the kappa coefficients were 0.61, 0.672, and 0.707, respectively. SVM performed the best. (2) Vegetation types and vegetation seasonal phases are closely related to rocky desertification. After combining them, the model accuracy and kappa coefficient improved to 91.1% and 0.861. (3) The spatial distribution characteristics of rocky desertification in Guizhou are obvious, showing a pattern of being heavy in the west, light in the east, heavy in the south, and light in the north. Rocky desertification has continuously increased from 2001 to 2019. In conclusion, combining the vertical spatial structure of vegetation and the differences in seasonal phase is an effective method to improve the modeling accuracy of rocky desertification, and the SVM model has the highest rocky desertification classification accuracy. The research results provide data support for exploring the spatiotemporal evolution pattern of rocky desertification in Guizhou.


2021 ◽  
Vol 436 ◽  
pp. 12-21
Author(s):  
Xinyue Dong ◽  
Shilin Gu ◽  
Wenzhang Zhuge ◽  
Tingjin Luo ◽  
Chenping Hou

Sign in / Sign up

Export Citation Format

Share Document