A statistical neural network for high-dimensional vector classification

Author(s):  
M. Verleysen ◽  
J.-L. Voz ◽  
P. Thissen ◽  
J.-D. Legat
Mathematics ◽  
2021 ◽  
Vol 9 (7) ◽  
pp. 779
Author(s):  
Ruriko Yoshida

A tropical ball is a ball defined by the tropical metric over the tropical projective torus. In this paper we show several properties of tropical balls over the tropical projective torus and also over the space of phylogenetic trees with a given set of leaf labels. Then we discuss its application to the K nearest neighbors (KNN) algorithm, a supervised learning method used to classify a high-dimensional vector into given categories by looking at a ball centered at the vector, which contains K vectors in the space.


Author(s):  
J. Nichols ◽  
Albert Cohen ◽  
Peter Binev ◽  
Olga Mula

Parametric PDEs of the general form $$ \mathcal{P}(u,a)=0 $$ are commonly used to describe many physical processes, where $\mathcal{P}$ is a differential operator, a is a high-dimensional vector of parameters and u is the unknown solution belonging to some Hilbert space V. Typically one observes m linear measurements of u(a) of the form $\ell_i(u)=\langle w_i,u \rangle$, $i=1,\dots,m$, where $\ell_i\in V'$ and $w_i$ are the Riesz representers, and we write $W_m = \text{span}\{w_1,\ldots,w_m\}$. The goal is to recover an approximation $u^*$ of u from the measurements. The solutions u(a) lie in a manifold within V which we can approximate by a linear space $V_n$, where n is of moderate dimension. The structure of the PDE ensure that for any a the solution is never too far away from $V_n$, that is, $\text{dist}(u(a),V_n)\le \varepsilon$. In this setting, the observed measurements and $V_n$ can be combined to produce an approximation $u^*$ of u up to accuracy $$ \Vert u -u^*\Vert \leq \beta^{-1}(V_n,W_m) \, \varepsilon $$ where $$ \beta(V_n,W_m) := \inf_{v\in V_n} \frac{\Vert P_{W_m}v\Vert}{\Vert v \Vert} $$ plays the role of a stability constant. For a given $V_n$, one relevant objective is to guarantee that $\beta(V_n,W_m)\geq \gamma >0$ with a number of measurements $m\geq n$ as small as possible. We present results in this direction when the measurement functionals $\ell_i$ belong to a complete dictionary.


2014 ◽  
Vol 5 (3) ◽  
pp. 82-96 ◽  
Author(s):  
Marijana Zekić-Sušac ◽  
Sanja Pfeifer ◽  
Nataša Šarlija

Abstract Background: Large-dimensional data modelling often relies on variable reduction methods in the pre-processing and in the post-processing stage. However, such a reduction usually provides less information and yields a lower accuracy of the model. Objectives: The aim of this paper is to assess the high-dimensional classification problem of recognizing entrepreneurial intentions of students by machine learning methods. Methods/Approach: Four methods were tested: artificial neural networks, CART classification trees, support vector machines, and k-nearest neighbour on the same dataset in order to compare their efficiency in the sense of classification accuracy. The performance of each method was compared on ten subsamples in a 10-fold cross-validation procedure in order to assess computing sensitivity and specificity of each model. Results: The artificial neural network model based on multilayer perceptron yielded a higher classification rate than the models produced by other methods. The pairwise t-test showed a statistical significance between the artificial neural network and the k-nearest neighbour model, while the difference among other methods was not statistically significant. Conclusions: Tested machine learning methods are able to learn fast and achieve high classification accuracy. However, further advancement can be assured by testing a few additional methodological refinements in machine learning methods.


2019 ◽  
Vol 29 (5) ◽  
pp. 311-319
Author(s):  
V. G. Mikhailov

Abstract We consider a special function ρ(H) of the subset H of n-dimensional vector linear space over the field K. This function is used in the estimates of accuracy of the Poisson approximation for the distribution of the number of solutions of systems of random equations and random inclusions over K. For the case when K = GF(2) and H is a sphere or ball (in the Hamming metric) in {0, 1}n we obtain explicit and approximate formulas for ρ(H) for sufficiently large values of n.


Sign in / Sign up

Export Citation Format

Share Document