scholarly journals Information Bottleneck Analysis by a Conditional Mutual Information Bound

Entropy ◽  
2021 ◽  
Vol 23 (8) ◽  
pp. 974
Author(s):  
Taro Tezuka ◽  
Shizuma Namekawa

Task-nuisance decomposition describes why the information bottleneck loss I(z;x)−βI(z;y) is a suitable objective for supervised learning. The true category y is predicted for input x using latent variables z. When n is a nuisance independent from y, I(z;n) can be decreased by reducing I(z;x) since the latter upper bounds the former. We extend this framework by demonstrating that conditional mutual information I(z;x|y) provides an alternative upper bound for I(z;n). This bound is applicable even if z is not a sufficient representation of x, that is, I(z;y)≠I(x;y). We used mutual information neural estimation (MINE) to estimate I(z;x|y). Experiments demonstrated that I(z;x|y) is smaller than I(z;x) for layers closer to the input, matching the claim that the former is a tighter bound than the latter. Because of this difference, the information plane differs when I(z;x|y) is used instead of I(z;x).

2018 ◽  
Vol 39 (11) ◽  
pp. 3015-3030
Author(s):  
BALÁZS GERENCSÉR ◽  
VIKTOR HARANGI

This paper is concerned with factors of independent and identically distributed processes on the $d$ -regular tree for $d\geq 3$ . We study the mutual information of values on two given vertices. If the vertices are neighbors (i.e. their distance is $1$ ), then a known inequality between the entropy of a vertex and the entropy of an edge provides an upper bound for the (normalized) mutual information. In this paper we obtain upper bounds for vertices at an arbitrary distance $k$ , of order $(d-1)^{-k/2}$ . Although these bounds are sharp, we also show that an interesting phenomenon occurs here: for any fixed process, the rate of mutual information decay is much faster, essentially of order $(d-1)^{-k}$ .


2021 ◽  
Vol 2021 (8) ◽  
Author(s):  
Sotaro Sugishita

Abstract We consider entanglement of first-quantized identical particles by adopting an algebraic approach. In particular, we investigate fermions whose wave functions are given by the Slater determinants, as for singlet sectors of one-matrix models. We show that the upper bounds of the general Rényi entropies are N log 2 for N particles or an N × N matrix. We compute the target space entanglement entropy and the mutual information in a free one-matrix model. We confirm the area law: the single-interval entropy for the ground state scales as $$ \frac{1}{3} $$ 1 3 log N in the large N model. We obtain an analytical $$ \mathcal{O}\left({N}^0\right) $$ O N 0 expression of the mutual information for two intervals in the large N expansion.


1996 ◽  
Vol 321 ◽  
pp. 335-370 ◽  
Author(s):  
R. R. Kerswell

Rigorous upper bounds on the viscous dissipation rate are identified for two commonly studied precessing fluid-filled configurations: an oblate spheroid and a long cylinder. The latter represents an interesting new application of the upper-bounding techniques developed by Howard and Busse. A novel ‘background’ method recently introduced by Doering & Constantin is also used to deduce in both instances an upper bound which is independent of the fluid's viscosity and the forcing precession rate. Experimental data provide some evidence that the observed viscous dissipation rate mirrors this behaviour at sufficiently high precessional forcing. Implications are then discussed for the Earth's precessional response.


Author(s):  
Indranil Biswas ◽  
Ajneet Dhillon ◽  
Nicole Lemire

AbstractWe find upper bounds on the essential dimension of the moduli stack of parabolic vector bundles over a curve. When there is no parabolic structure, we improve the known upper bound on the essential dimension of the usual moduli stack. Our calculations also give lower bounds on the essential dimension of the semistable locus inside the moduli stack of vector bundles of rank r and degree d without parabolic structure.



2021 ◽  
Vol 2021 (9) ◽  
Author(s):  
Alex May

Abstract We prove a theorem showing that the existence of “private” curves in the bulk of AdS implies two regions of the dual CFT share strong correlations. A private curve is a causal curve which avoids the entanglement wedge of a specified boundary region $$ \mathcal{U} $$ U . The implied correlation is measured by the conditional mutual information $$ I\left({\mathcal{V}}_1:\left.{\mathcal{V}}_2\right|\mathcal{U}\right) $$ I V 1 : V 2 U , which is O(1/GN) when a private causal curve exists. The regions $$ {\mathcal{V}}_1 $$ V 1 and $$ {\mathcal{V}}_2 $$ V 2 are specified by the endpoints of the causal curve and the placement of the region $$ \mathcal{U} $$ U . This gives a causal perspective on the conditional mutual information in AdS/CFT, analogous to the causal perspective on the mutual information given by earlier work on the connected wedge theorem. We give an information theoretic argument for our theorem, along with a bulk geometric proof. In the geometric perspective, the theorem follows from the maximin formula and entanglement wedge nesting. In the information theoretic approach, the theorem follows from resource requirements for sending private messages over a public quantum channel.


Author(s):  
Greg Ver Steeg

Learning by children and animals occurs effortlessly and largely without obvious supervision. Successes in automating supervised learning have not translated to the more ambiguous realm of unsupervised learning where goals and labels are not provided. Barlow (1961) suggested that the signal that brains leverage for unsupervised learning is dependence, or redundancy, in the sensory environment. Dependence can be characterized using the information-theoretic multivariate mutual information measure called total correlation. The principle of Total Cor-relation Ex-planation (CorEx) is to learn representations of data that "explain" as much dependence in the data as possible. We review some manifestations of this principle along with successes in unsupervised learning problems across diverse domains including human behavior, biology, and language.


Sign in / Sign up

Export Citation Format

Share Document