minimal norm
Recently Published Documents


TOTAL DOCUMENTS

110
(FIVE YEARS 21)

H-INDEX

9
(FIVE YEARS 1)

Author(s):  
Nicola Gigli ◽  
Francesco Nobili

AbstractWe review the theory of Gradient Flows in the framework of convex and lower semicontinuous functionals on $$\textsf {CAT} (\kappa )$$ CAT ( κ ) -spaces and prove that they can be characterized by the same differential inclusion $$y_t'\in -\partial ^-\textsf {E} (y_t)$$ y t ′ ∈ - ∂ - E ( y t ) one uses in the smooth setting and more precisely that $$y_t'$$ y t ′ selects the element of minimal norm in $$-\partial ^-\textsf {E} (y_t)$$ - ∂ - E ( y t ) . This generalizes previous results in this direction where the energy was also assumed to be Lipschitz. We then apply such result to the Korevaar–Schoen energy functional on the space of $$L^2$$ L 2 and (0) valued maps: we define the Laplacian of such $$L^2$$ L 2 map as the element of minimal norm in $$-\partial ^-\textsf {E} (u)$$ - ∂ - E ( u ) , provided it is not empty. The theory of gradient flows ensures that the set of maps admitting a Laplacian is $$L^2$$ L 2 -dense. Basic properties of this Laplacian are then studied.


2021 ◽  
Vol 28 (2) ◽  
pp. 186-197
Author(s):  
Mikhail Viktorovich Nevskii

Let  $B$ be a Euclidean ball in ${\mathbb R}^n$ and let $C(B)$ be a space of continuos functions $f:B\to{\mathbb R}$ with the uniform norm $\|f\|_{C(B)}:=\max_{x\in B}|f(x)|.$ By $\Pi_1\left({\mathbb R}^n\right)$ we mean a set of polynomials of degree $\leq 1$, i.e., a set of linear functions upon ${\mathbb R}^n$. The interpolation projector  $P:C(B)\to \Pi_1({\mathbb R}^n)$ with the nodes $x^{(j)}\in B$ is defined by the equalities $Pf\left(x^{(j)}\right)=f\left(x^{(j)}\right)$,  $j=1,\ldots, n+1$.The norm of $P$ as an operator from $C(B)$ to $C(B)$ can be calculated by the formula $\|P\|_B=\max_{x\in B}\sum |\lambda_j(x)|.$ Here $\lambda_j$ are the basic Lagrange polynomials corresponding to the $n$-dimensional nondegenerate simplex $S$ with the vertices $x^{(j)}$. Let $P^\prime$ be a projector having the nodes in the vertices of a regular simplex inscribed into the ball. We describe the points $y\in B$ with the property $\|P^\prime\|_B=\sum |\lambda_j(y)|$. Also we formulate some geometric conjecture which implies that $\|P^\prime\|_B$ is equal to the minimal norm of an interpolation projector with nodes in $B$.  We prove that this conjecture holds true at least for $n=1,2,3,4$. 


Acta Numerica ◽  
2021 ◽  
Vol 30 ◽  
pp. 87-201
Author(s):  
Peter L. Bartlett ◽  
Andrea Montanari ◽  
Alexander Rakhlin

The remarkable practical success of deep learning has revealed some major surprises from a theoretical perspective. In particular, simple gradient methods easily find near-optimal solutions to non-convex optimization problems, and despite giving a near-perfect fit to training data without any explicit effort to control model complexity, these methods exhibit excellent predictive accuracy. We conjecture that specific principles underlie these phenomena: that overparametrization allows gradient methods to find interpolating solutions, that these methods implicitly impose regularization, and that overparametrization leads to benign overfitting, that is, accurate predictions despite overfitting training data. In this article, we survey recent progress in statistical learning theory that provides examples illustrating these principles in simpler settings. We first review classical uniform convergence results and why they fall short of explaining aspects of the behaviour of deep learning methods. We give examples of implicit regularization in simple settings, where gradient methods lead to minimal norm functions that perfectly fit the training data. Then we review prediction methods that exhibit benign overfitting, focusing on regression problems with quadratic loss. For these methods, we can decompose the prediction rule into a simple component that is useful for prediction and a spiky component that is useful for overfitting but, in a favourable setting, does not harm prediction accuracy. We focus specifically on the linear regime for neural networks, where the network can be approximated by a linear model. In this regime, we demonstrate the success of gradient flow, and we consider benign overfitting with two-layer networks, giving an exact asymptotic analysis that precisely demonstrates the impact of overparametrization. We conclude by highlighting the key challenges that arise in extending these insights to realistic deep learning settings.


Author(s):  
Ya-Nan Li ◽  
Yun-Zhang Li

The concept of Hilbert–Schmidt frame (HS-frame) was first introduced by Sadeghi and Arefijamaal in 2012. It is more general than [Formula: see text]-frames, and thus, covers many generalizations of frames. This paper addresses the theory of HS-frames. We present a parametric and algebraic formula for all duals of an arbitrarily given HS-frame; prove that the canonical HS-dual induces a minimal-norm expression of the elements in Hilbert spaces; characterize when an HS-frame is an HS-Riesz basis, and when an HS-Bessel sequence is an HS-Riesz sequence (HS-Riesz basis) in terms of Gram matrices.


2021 ◽  
Vol 7 (4) ◽  
pp. 5029-5048
Author(s):  
Anli Wei ◽  
◽  
Ying Li ◽  
Wenxv Ding ◽  
Jianli Zhao ◽  
...  

<abstract><p>In this paper, we propose an efficient method for some special solutions of the quaternion matrix equation $ AXB+CYD = E $. By integrating real representation of a quaternion matrix with $ \mathcal{H} $-representation, we investigate the minimal norm least squares solution of the previous quaternion matrix equation over different constrained matrices and obtain their expressions. In this way, we first apply $ \mathcal{H} $-representation to solve quaternion matrix equation with special structure, which not only broadens the application scope of $ \mathcal{H} $-representation, but further expands the research idea of solving quaternion matrix equation. The algorithms only include real operations. Consequently, it is very simple and convenient, and it can be applied to all kinds of quaternion matrix equation with similar problems. The numerical example is provided to illustrate the feasibility of our algorithms.</p></abstract>


Author(s):  
Pedro Pérez-Aros ◽  
David Salas ◽  
Emilio Vilches

Sign in / Sign up

Export Citation Format

Share Document