A case study by using Python to implement data and dimensionality reduction

Author(s):  
Huang Chih-Chien ◽  
Hsu Chung-Chian ◽  
Wang Suefen ◽  
Pon YuShun ◽  
Liao WenWei
2021 ◽  
Author(s):  
Cheng Yuanyuan

Abstract:Purpose: To study the effect of the application of the dimensionality reduction in logical judgment (or logical reasoning, logical inference) programs. Methods: Use enumeration and dimensionality reduction methods to solve logical judgment problems.The effect of the two methods is illustrated in the form of a case study. Results: For logical judgmentproblems, using enumeration method to find the best answer is a comprehensive and fundamental method, but the disadvantage is that it is computationally intensive and computationally inefficient. Compared with the ideas of parallel treatment of known conditions by enumeration method, the application of dimensionality reduction thinking was built on the basis of fully mining information for feature extraction and feature selection. Conclusions: The dimensionality reduction method was applied to the logical judgment problems, and on the basis of fully mining information, the dimensionality reduction principle of statistics were applied to stratify and merge variables with the same or similar characteristics to achieve the purpose of streamlining variables, simplifying logical judgment steps, reducing computation and improving algorithm efficiency.


2018 ◽  
Vol 10 (10) ◽  
pp. 1564 ◽  
Author(s):  
Patrick Bradley ◽  
Sina Keller ◽  
Martin Weinmann

In this paper, we investigate the potential of unsupervised feature selection techniques for classification tasks, where only sparse training data are available. This is motivated by the fact that unsupervised feature selection techniques combine the advantages of standard dimensionality reduction techniques (which only rely on the given feature vectors and not on the corresponding labels) and supervised feature selection techniques (which retain a subset of the original set of features). Thus, feature selection becomes independent of the given classification task and, consequently, a subset of generally versatile features is retained. We present different techniques relying on the topology of the given sparse training data. Thereby, the topology is described with an ultrametricity index. For the latter, we take into account the Murtagh Ultrametricity Index (MUI) which is defined on the basis of triangles within the given data and the Topological Ultrametricity Index (TUI) which is defined on the basis of a specific graph structure. In a case study addressing the classification of high-dimensional hyperspectral data based on sparse training data, we demonstrate the performance of the proposed unsupervised feature selection techniques in comparison to standard dimensionality reduction and supervised feature selection techniques on four commonly used benchmark datasets. The achieved classification results reveal that involving supervised feature selection techniques leads to similar classification results as involving unsupervised feature selection techniques, while the latter perform feature selection independently from the given classification task and thus deliver generally versatile features.


Author(s):  
Max Losch ◽  
Mario Fritz ◽  
Bernt Schiele

AbstractToday’s deep learning systems deliver high performance based on end-to-end training but are notoriously hard to inspect. We argue that there are at least two reasons making inspectability challenging: (i) representations are distributed across hundreds of channels and (ii) a unifying metric quantifying inspectability is lacking. In this paper, we address both issues by proposing Semantic Bottlenecks (SB), which can be integrated into pretrained networks, to align channel outputs with individual visual concepts and introduce the model agnostic Area Under inspectability Curve (AUiC) metric to measure the alignment. We present a case study on semantic segmentation to demonstrate that SBs improve the AUiC up to six-fold over regular network outputs. We explore two types of SB-layers in this work. First, concept-supervised SB-layers (SSB), which offer inspectability w.r.t. predefined concepts that the model is demanded to rely on. And second, unsupervised SBs (USB), which offer equally strong AUiC improvements by restricting distributedness of representations across channels. Importantly, for both SB types, we can recover state of the art segmentation performance across two different models despite a drastic dimensionality reduction from 1000s of non aligned channels to 10s of semantics-aligned channels that all downstream results are based on.


Sign in / Sign up

Export Citation Format

Share Document