scholarly journals A New Method to Construct the KD Tree Based on Presorted Results

Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-7
Author(s):  
Yu Cao ◽  
Huizan Wang ◽  
Wenjing Zhao ◽  
Boheng Duan ◽  
Xiaojiang Zhang

Searching is one of the most fundamental operations in many complex systems. However, the complexity of the search process would increase dramatically in high-dimensional space. K-dimensional (KD) tree, as a classical data structure, has been widely used in high-dimensional vital data search. However, at present, common methods proposed for KD tree construction are either unstable or time-consuming. This paper proposed a new algorithm to construct a balanced KD tree based on presorted results. Compared with previous similar method, the new algorithm could reduce the complexity of the construction process (excluding the presorting process) from O (KNlog2N) level to O (Nlog2N) level, where K is the number of dimensions and N is the number of data. In addition, with the help of presorted results, the performance of the new method is no longer subject to the initial conditions, which expands the application scope of KD tree.

2020 ◽  
Author(s):  
Maximilian Gelbrecht ◽  
Jürgen Kurths ◽  
Frank Hellmann

<p>Many high-dimensional complex systems such as climate models exhibit an enormously complex landscape of possible asymptotic state. On most occasions these are challenging to analyse with traditional bifurcation analysis methods. Often, one is also more broadly interested in classes of asymptotic states. Here, we present a novel numerical approach prepared for analysing such high-dimensional multistable complex systems: Monte Carlo Basin Bifurcation Analysis (MCBB).<span>  </span>Based on random sampling and clustering methods, we identify the type of dynamic regimes with the largest basins of attraction and track how the volume of these basins change with the system parameters. In order to due this suitable, easy to compute, statistics of trajectories with randomly generated initial conditions and parameters are clustered by an algorithm such as DBSCAN. Due to the modular and flexible nature of the method, it has a wide range of possible applications. While initially oscillator networks were one of the main applications of this methods, here we present an analysis of a simple conceptual climate model setup up by coupling an energy balance model to the Lorenz96 system. The method is available to use as a package for the Julia language.<span> </span></p>


2014 ◽  
Vol 590 ◽  
pp. 688-692
Author(s):  
Bei Chen ◽  
Kun Song

Overlap information usually exits in the high-dimensional data. Misclassified points may be more when affinity propagation clustering is applied to these data. Concerning this problem, a new method combining principal components analysis and affinity propagation clustering is proposed. In this method, dimensionality of the original data is reduced on the premise of reserving most information of the variables. Then, affinity propagation clustering is implemented in the low-dimensional space. Thus, because the redundant information is deleted, the classification is accurate. Experiment is done by using this new method, the results of the experiment explain that this method is effective.


2021 ◽  
pp. 1-12
Author(s):  
Jian Zheng ◽  
Jianfeng Wang ◽  
Yanping Chen ◽  
Shuping Chen ◽  
Jingjin Chen ◽  
...  

Neural networks can approximate data because of owning many compact non-linear layers. In high-dimensional space, due to the curse of dimensionality, data distribution becomes sparse, causing that it is difficulty to provide sufficient information. Hence, the task becomes even harder if neural networks approximate data in high-dimensional space. To address this issue, according to the Lipschitz condition, the two deviations, i.e., the deviation of the neural networks trained using high-dimensional functions, and the deviation of high-dimensional functions approximation data, are derived. This purpose of doing this is to improve the ability of approximation high-dimensional space using neural networks. Experimental results show that the neural networks trained using high-dimensional functions outperforms that of using data in the capability of approximation data in high-dimensional space. We find that the neural networks trained using high-dimensional functions more suitable for high-dimensional space than that of using data, so that there is no need to retain sufficient data for neural networks training. Our findings suggests that in high-dimensional space, by tuning hidden layers of neural networks, this is hard to have substantial positive effects on improving precision of approximation data.


2021 ◽  
Vol 47 (1) ◽  
Author(s):  
Fabian Laakmann ◽  
Philipp Petersen

AbstractWe demonstrate that deep neural networks with the ReLU activation function can efficiently approximate the solutions of various types of parametric linear transport equations. For non-smooth initial conditions, the solutions of these PDEs are high-dimensional and non-smooth. Therefore, approximation of these functions suffers from a curse of dimension. We demonstrate that through their inherent compositionality deep neural networks can resolve the characteristic flow underlying the transport equations and thereby allow approximation rates independent of the parameter dimension.


Sensors ◽  
2019 ◽  
Vol 19 (20) ◽  
pp. 4454 ◽  
Author(s):  
Marek Piorecky ◽  
Vlastimil Koudelka ◽  
Jan Strobl ◽  
Martin Brunovsky ◽  
Vladimir Krajca

Simultaneous recordings of electroencephalogram (EEG) and functional magnetic resonance imaging (fMRI) are at the forefront of technologies of interest to physicians and scientists because they combine the benefits of both modalities—better time resolution (hdEEG) and space resolution (fMRI). However, EEG measurements in the scanner contain an electromagnetic field that is induced in leads as a result of gradient switching slight head movements and vibrations, and it is corrupted by changes in the measured potential because of the Hall phenomenon. The aim of this study is to design and test a methodology for inspecting hidden EEG structures with respect to artifacts. We propose a top-down strategy to obtain additional information that is not visible in a single recording. The time-domain independent component analysis algorithm was employed to obtain independent components and spatial weights. A nonlinear dimension reduction technique t-distributed stochastic neighbor embedding was used to create low-dimensional space, which was then partitioned using the density-based spatial clustering of applications with noise (DBSCAN). The relationships between the found data structure and the used criteria were investigated. As a result, we were able to extract information from the data structure regarding electrooculographic, electrocardiographic, electromyographic and gradient artifacts. This new methodology could facilitate the identification of artifacts and their residues from simultaneous EEG in fMRI.


Sign in / Sign up

Export Citation Format

Share Document