Next Article in Journal
Computational-Intelligence-Based Scheduling with Edge Computing in Cyber–Physical Production Systems
Next Article in Special Issue
A Measure of Synergy Based on Union Information
Previous Article in Journal
A Communication Anti-Jamming Scheme Assisted by RIS with Angular Response
Previous Article in Special Issue
Arrow Contraction and Expansion in Tropical Diagrams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Conditioning in Tropical Probability Theory

by
Rostislav Matveev
1,* and
Jacobus W. Portegies
2,*
1
Max Planck Institute for Mathematics in the Sciences, 04103 Leipzig, Germany
2
Department of Mathematics and Computer Science, Eindhoven University of Technology, 5600 MB Eindhoven, The Netherlands
*
Authors to whom correspondence should be addressed.
Entropy 2023, 25(12), 1641; https://doi.org/10.3390/e25121641
Submission received: 6 November 2023 / Revised: 4 December 2023 / Accepted: 7 December 2023 / Published: 9 December 2023

Abstract

:
We define a natural operation of conditioning of tropical diagrams of probability spaces and show that it is Lipschitz continuous with respect to the asymptotic entropy distance.

1. Introduction

In [1,2], we have initiated the study of tropical probability spaces and their diagrams. In [1], we endowed (commutative) diagrams of probability spaces with the intrinsic entropy distance and, in [2], we defined tropical diagrams as points in the asymptotic cone of the metric space. They are represented by certain sequences of diagrams of probability spaces.
We expect that tropical diagrams will be helpful in the study of information optimization problems, such as the ones considered in [3,4,5,6,7,8], and we have indeed applied them to derive a dimension-reduction result for the shape of the entropic cone in [9].
In this present article, we introduce the notion of conditioning on a space in a tropical diagram and show that the operation is Lipschitz continuous with respect to the asymptotic entropy distance.
It is a rather technical result, and we have, therefore, decided to treat it in this separate article, but it is an important ingredient in the theory and, in particular, we need it for the dimension-reduction result mentioned before.
Given a tuple of finite-valued random variables ( X i ) i = 1 n and a random variable Y , one may "condition" the collection ( X i ) on Y . The result of this operation is a family of n-tuples of random variables denoted ( X i | Y ) i = 1 n parameterized by those values of Y that have positive probability. Each tuple of random variables in this family is defined on a separate probability space.
When passing to the tropical setting, the situation is different in the sense that when we condition a tropical diagram [ X ] on a space [ Y ] , the result is again a tropical diagram [ X | Y ] rather than a family. After recalling some preliminaries in Section 2, we describe the operation of conditioning and prove that the result depends in a Lipschitz way on the original diagram in Section 3.

2. Preliminaries

Our main objects of study are commutative diagrams of probability spaces and their tropical counterparts. In this section, we recall briefly the main definitions and results.

2.1. Probability Spaces and Their Diagrams

2.1.1. Probability Spaces

By a finite probability space, we mean a set with a probability measure that has finite support. A reduction from one probability space to another is an equivalence class of measure-preserving maps. Two maps are equivalent if they coincide on a set of full measures. We call a point x in a probability space X = ( X ̲ , p ) an atom if it has positive weight, and we write x X to mean x is an atom in X (as opposed to x X ̲ for points in the underlying set). For a probability space X, we denote by | X | the cardinality of the support of the probability measure.

2.1.2. Indexing Categories

To record the combinatorial structure of commutative diagrams of probability spaces and reductions, we use an object that we call an indexing category. By an indexing category, we mean a finite category G such that for any pair of objects i , j G , there is at most one morphism between them either way. In addition, we will assume it satisfies one additional property that we will describe after introducing some terminology. For a pair of objects i , j G such that there is a morphism γ i j : i j , object i will be called an ancestor of j and object j will be called a descendant of i. The subcategory of all descendants of an object i G is called an ideal generated by i and will be denoted i , while we will call the subcategory consisting of all ancestors of i together with all the morphisms in it a co-ideal generated by i and denote it by i . (The term filter is also used for a co-ideal in the literature about lattices).
The additional property that an indexing category has to satisfy is that for any pair of objects i , j G , there exists a minimal common ancestor  𝚤 ^ , and 𝚤 ^ is an ancestor for both i and j and any other ancestor of them both is also an ancestor of 𝚤 ^ ; in other words, G is an upper semi-lattice.
An equivalent formulation of the property above is the following: the intersection of the co-ideals generated by two objects i , j G is also a co-ideal generated by some object 𝚤 ^ G .
Any indexing category G is necessarily initial, which means that there exists an initial object, that is an object i 0 such that G = i 0 .
A fan in a category is a pair of morphisms with the same domain. A fan ( i k j ) is called minimal if for any other fan ( i l j ) included in a commutative diagram
Entropy 25 01641 i001
the vertical arrow must be an isomorphism; in other words, k is a minimal common ancestor of i and j.
For any pair of objects i , j in an indexing category G , there exists a unique minimal fan ( i 𝚤 ^ j ) in G .

2.1.3. Diagrams

We denote by Prob the category of finite probability spaces and reductions, i.e., the equivalence classes of measure-preserving maps. For an indexing category G = i ; γ i j , a G -diagram is a functor X : G Prob . A reduction f from one G -diagram X = X i ; χ i j to another Y = Y i ; υ i j is a natural transformation between the functors. It amounts to a collection of reductions f i : X i Y i , such that the big diagram consisting of all spaces X i , Y i and all morphisms χ i j , υ i j and f i is commutative. The category of G -diagrams and reductions will be denoted as Prob G . The construction of diagrams could be iterated; thus, we can consider H -diagrams of G -diagrams and denote the corresponding category Prob G H = Prob G , H . Every H -diagram of a G -diagram can also be considered as G -diagrams of H -diagrams; thus, there is a natural equivalence of categories Prob G , H Prob H , G .
A G -diagram X will be called minimal if it maps minimal fans in G to minimal fans in the target category. The subspace of all minimal G -diagrams will be denoted Prob G m . In [1], we have shown that for any fan in Prob or in Prob G , its minimization exists and is unique up to isomorphism.

2.1.4. Tensor Product

The tensor product of two probability spaces X = ( X ̲ , p ) and Y = ( Y ̲ , q ) is their independent product X Y : = ( X ̲ × Y ̲ , p q ) . For two G -diagrams X = X i ; χ i j and Y = Y i ; υ i j , we define their tensor product to be X Y = X i Y ; χ i j × υ i j .

2.1.5. Constant Diagrams

Given an indexing category G and a probability space, we can form a constant diagram X G that has all spaces equal to X and all reductions equal to the identity isomorphism. Sometimes, when such a constant diagram is included in a diagram with other G -diagrams (such as, for example, a reduction X X G ), we will write simply X in place of X G .

2.1.6. Entropy

Evaluating entropy on every space in a G -diagram, we obtain a tuple of non-negative numbers indexed by objects in G ; thus, entropy gives a map
Ent * : Prob G R G ,
where the target space R G is a space of real-valued functions on the set of objects in G endowed with the 1 -norm. Entropy is a homomorphism in that it satisfies
Ent * ( X Y ) = Ent * ( X ) + Ent * ( Y ) .

2.1.7. Entropy Distance

Let G be an indexing category and K = ( X Z Y ) be a fan of G -diagrams. We define the entropy distance as
kd ( K ) : = Ent * Z Ent * X 1 + Ent * Z Ent * Y 1 .
The intrinsic entropy distance between two G -diagrams is defined to be the infimal entropy distance of all fans with terminal diagrams X and Y :
k ( X , Y ) : = inf kd ( K ) : K = ( X Z Y ) .
The intrinsic entropy distance was introduced in [10,11] for probability spaces.
In [1], it is shown that the infimum is attained, that the optimal fan is minimal, that k is a pseudo-distance, which vanishes if, and only if, X and Y are isomorphic, and that Ent * is a 1-Lipschitz linear functional with respect to k .

2.2. Diagrams of Sets, Distributions, and Empirical Reductions

2.2.1. Distributions on Sets

For a set S, we denote by Δ S the collection of all finitely-supported probability distributions on S. For a pair of distributions π 1 , π 2 Δ S , we denote by π 1 π 2 1 the total variation distance between them.
For a map f : S S between two sets, we denote by f * : Δ S Δ S the induced affine map (the map-preserving convex combinations).
For n N , we define the empirical map q : S n Δ S by the assignment below. For s ¯ = ( s 1 , , s n ) S n and A S , we define
q ( s ¯ ) ( A ) : = 1 n · | k : s k A | .
For a finite probability space X = ( S , p ) , the empirical distribution on Δ X is the push-forward τ n : = q * p n . Thus,
q : X n ( Δ X , τ n )
is a reduction of finite probability spaces. The construction of empirical reduction is functorial, which is for a reduction between two probability spaces f : X Y , the diagram of the reductions
Entropy 25 01641 i002
commutes.

2.2.2. Distributions on Diagrams of Sets

Let Set denote the category of sets and surjective maps. For an indexing category G , we denote by Set G the category of G -diagrams in Set . The objects in Set G are commutative diagrams of sets indexed by G , and the spaces in such a diagram are sets, where the arrows represent surjective maps, subject to commutativity relations.
For a diagram of sets S = S i ; σ i j , we define the space of distributions on the diagram  S by
Δ S : = ( π i ) i Δ S i : ( σ i j ) * π i = π j .
If S 0 is the initial set of S , then there is an isomorphism of
Δ S 0 Δ S Δ S 0 π 0 ( σ 0 i ) * π 0 Δ S Δ S 0 π 0 π i Δ S .
Given a G -diagram of sets S = S i ; σ i j and an element π Δ S , we can construct a G -diagram of probability spaces ( S , π ) : = ( S i , π i ) ; σ i j . Note that any diagram X of probability spaces has this form.

2.3. Conditioning

Consider a G -diagram of probability spaces X = ( S , π ) , where S is a diagram of sets and π Δ S . Let X 0 = ( S 0 , π 0 ) be the initial space in X and U : = X i be another space in X . Since S 0 is initial, there is a map σ 0 , i : S 0 S i . Fix an atom u U and define the conditioned distribution π 0 ( · | u ) on S 0 as the distribution supported in σ 0 , i 1 ( u ) and for every s σ 0 , i 1 ( u ) defined by
π 0 ( s | u ) : = π 0 ( s ) π 0 ( σ 0 , i 1 ( u ) ) .
Let π ( · | u ) Δ S be the distribution corresponding to π 0 ( · | u ) under the isomorphism in (1). We define the conditioned  G -diagram as X | u : = ( S , π ( · | u ) ) .

2.4. The Slicing Lemma

In [1], we prove the so-called Slicing Lemma that allows us to estimate the intrinsic entropy distance between two diagrams in terms of distances between conditioned diagrams. Among the corollaries of the Slicing Lemma is the following inequality.
Proposition 1. 
Let ( X X ^ U G ) Prob G , Λ 2 be a fan of G -diagrams of probability spaces and Y Prob G be another diagram. Then,
k ( X , Y ) U k ( X | u , Y ) d p ( u ) + 2 [ [ G ] ] · Ent U .
The fan in the assumption of the proposition above can often be constructed in the following manner. Suppose X is a G -diagram and U : = X ι is a space in it for some ι G . We can construct a fan ( X f X ^ g U G ) Prob G , Λ 2 by assigning X ^ i to be the initial space of the (unique) minimal fan in X with terminal spaces X i and U and f i and g i to be left and right reductions in that fan for any i G .

2.5. Tropical Diagrams

A detailed discussion of the topics in this section can be found in [2].
The asymptotic entropy distance between two diagrams of the same combinatorial type is defined by
κ ( X , Y ) : = lim 1 n k ( X n , Y n ) .
A tropical G -diagram is an equivalence class of certain sequences of G -diagrams of probability spaces. Below, we describe the type of sequences and the equivalence relation.
A function φ : R 1 R 0 is called an admissible function if φ is non-decreasing and there is a constant D φ , such that for any t 1 :
t · t φ ( s ) s 2 d s D φ · φ ( t ) .
An example of an admissible function will be φ ( t ) = t α for α [ 0 , 1 ) .
A sequence X ¯ = ( X ( n ) : n N 0 ) of diagrams of probability spaces will be called quasi-linear with the defect bounded by an admissible function φ if for some C > 0 and all m , n N , it satisfies
κ X ( n + m ) , X ( n ) X ( m ) C · φ ( n + m ) .
For example for a diagram X , the sequence X : = ( X n : n N 0 ) is φ -quasi-linear for φ 0 (and for any admissible φ ). Sequences with zero defect are called linear, and the space of all linear sequences in Prob G is denoted by L ( Prob G ) .
The asymptotic entropic distance between two φ -quasi-linear sequences X ¯ = X ( n ) : n N 0 and Y ¯ = Y ( n ) : n N 0 is defined to be
κ ( X ¯ , Y ¯ ) : = lim n 1 n k ( X ( n ) , Y ( n ) ) ,
and the sequences are called asymptotically equivalent if κ ( X ¯ , Y ¯ ) = 0 . An equivalence class of a sequence X ¯ will be denoted as [ X ] , and the totality of all the classes as Prob [ G ] . We have shown in [2] that the space of equivalence classes of φ -quasi-linear sequences does not depend on the choice of a non-zero admissible function φ .
The sum of two such equivalence classes is defined to be the equivalence class of the sequence obtained by tensor-multiplying representative sequences of the summands term-wise. In addition, there is a doubly transitive action of R 0 on Prob [ G ] . In [2], the following theorem is proven.
Theorem 1. 
Let G be an indexing category. Then:
1. 
The space Prob [ G ] does not depend on the choice of a positive admissible function φ up to isometry.
2. 
The space Prob [ G ] is metrically complete.
3. 
The map X X is a κ - κ -isometric embedding. The space of linear sequences, i.e., the image of the map above, is dense in Prob [ G ] .
4. 
There is a distance-preserving homomorphism from Prob [ G ] into a Banach space B, whose image is a closed convex cone in B.
5. 
The entropy functional
Ent * : Prob [ G ] R G [ X ( n ) n N 0 ] lim n 1 n Ent * X ( n )
is a well-defined 1-Lipschitz linear map.

2.6. Asymptotic Equipartition Property for Diagrams

Among all G -diagrams, there is a special class of maximally symmetric ones. We call such diagrams homogeneous; see below for the definition. Homogeneous diagrams come in very handy in many considerations, because their structure is easier to describe than that of general diagrams. We show below that among the tropical diagrams, those that have homogeneous representatives are dense. It means, in particular, that when considering continuous functionals in the space of diagrams, it suffices to only study them in the space of all homogeneous diagrams.

2.6.1. Homogeneous Diagrams

A G -diagram X is called homogeneous if the automorphism group Aut ( X ) acts transitively on every space in X , by which we mean that the action is transitive on the support of the probability measure. Homogeneous probability spaces are isomorphic to uniform spaces. For more complex indexing categories, this simple description is not sufficient.

2.6.2. Tropical Homogeneous Diagrams

The subcategory of all homogeneous G -diagrams will be denoted Prob G h , and we write Prob G h , m for the category of minimal homogeneous G -diagrams. These spaces are invariant under the tensor product; thus, they are metric Abelian monoids, and the general "tropicalization" described in [2] can be performed. Passing to the tropical limit, we obtain spaces of tropical (minimal) homogeneous diagrams, which we denote by Prob [ G ] h and Prob [ G ] h , m , respectively.

2.6.3. Asymptotic Equipartition Property

For an indexing category G , denote by [ [ G ] ] the number of objects in G . In [1], the following theorem is proven.
Theorem 2. 
Suppose X Prob G is a G -diagram of probability spaces for some fixed indexing category G . Then, there exists a sequence H ¯ = ( H n ) n = 0 of homogeneous G -diagrams, such that
1 n k ( X n , H n ) C ( | X 0 | , [ [ G ] ] ) · ln 3 n n ,
where C ( | X 0 | , [ [ G ] ] ) is a constant only depending on | X 0 | and [ [ G ] ] .
The approximating sequence of homogeneous diagrams is evidently quasi-linear with the defect bounded by the admissible function
φ ( t ) : = 2 C ( | X 0 | , [ [ G ] ] ) · t 3 / 4 2 C ( | X 0 | , [ [ G ] ] ) · t 1 / 2 · ln 3 / 2 t .
Thus, Theorem 2 above states that L ( Prob G ) Prob [ G ] h . On the other hand, we have shown in [2] that the space of linear sequences L ( Prob G ) is dense in Prob [ G ] . Combining the two statements, we obtain the following theorem.
Theorem 3. 
For any indexing category G , the space Prob [ G ] h is dense in Prob [ G ] . Similarly, the space Prob [ G ] h , m is dense in Prob [ G ] m .

3. Conditioning of Tropical Diagrams

3.1. Motivation

Let X Prob G be a G -diagram of probability spaces containing probability space U = X i 0 indexed by an object i 0 G .
Given an atom u U , we can define a conditioned diagram X | u . If the diagram X is homogeneous, then the isomorphism class of X | u is independent of u, so that ( X | u : u U ) is a constant family. On the other hand, we have shown that the power of any diagram can be approximated by homogeneous diagrams, thus suggesting that in the tropical setting X | U should be a well-defined tropical diagram, rather than a family. Below, we give a definition of the tropical conditioning operation and prove its consistency.

3.2. Classical-Tropical Conditioning

Here, we define the operation of conditioning of the classical diagram, such that the result is a tropical diagram. Let X be a G -diagram of probability spaces and U be a space in X . We define the conditioning map
[ · | · ] : Prob G Prob [ G ]
by conditioning X by u U and averaging the corresponding tropical diagrams:
[ X | U ] : = u U ( X | u ) d p U ( u ) ,
where ( X | u ) is the tropical diagram represented by a linear sequence generated by X | u ; see Section 2.5. Note that the integral on the right-hand side is just a finite convex combination of tropical diagrams. Expanding all the definitions, we will obtain for [ Y ] : = [ X | U ] , the representative sequence
Y ( n ) = u U ( X | u ) n · p ( u ) .

3.3. Properties

3.3.1. Conditioning of Homogeneous Diagrams

If the diagram X is homogeneous, then for any atom u U , with a positive weight,
[ X | U ] = ˇ ( X | u ) .

3.3.2. Entropy

By definition, the conditioned entropy is
Ent * ( X | U ) : = U Ent * ( X | u ) d p U ( u ) .
Now that [ X | U ] is a tropical diagram, the expression Ent * ( X | U ) can be interpreted in two, a priori different, ways: by the formula above and as the entropy of the object introduced in the previous subsection. Fortunately, the numeric value of it does not depend on the interpretation since the entropy is a linear functional on Prob [ G ] .

3.3.3. Additivity

If X and Y are two G -diagrams with U : = X ι , V : = Y ι for some ι G , then
[ ( X Y ) | ( U V ) ] = [ X | U ] + [ Y | V ] .
Proof. 
[ ( X Y ) | ( U V ) ] = U V ( X Y ) | ( u , v ) d p ( u ) d p ( v ) = U V ( X | u + Y | v ) d p ( u ) d p ( v ) = U X | u d p ( u ) + V Y | v d p ( v ) = [ X | U ] + [ Y | V ]

3.3.4. Homogeneity

It follows that for any diagram X , a space U in X and n N 0 holds
[ X n | U n ] = n · [ X | U ] .

3.4. Continuity and Lipschitz Property

Proposition 2. 
Let G be an indexing category, X , Y Prob G be two G diagrams, and U : = X ι and V : = Y ι be two spaces in X and Y , respectively, indexed by some ι G . Then,
κ [ X | U ] , [ Y | V ] ( 2 · [ [ G ] ] + 1 ) · k X , Y .
Using the homogeneity property of conditioning, Section 3.3.4, we can obtain the following stronger inequality.
Corollary 1. 
In the setting of Proposition 2, the following holds:
κ [ X | U ] , [ Y | V ] ( 2 · [ [ G ] ] + 1 ) · κ X , Y .
Before we prove Proposition 2, we will need some preparatory lemmas.
Lemma 1. 
Let A be a G -diagram of probability spaces and E be a space in it. Let q : E n ( Δ E , τ n ) be the empirical reduction. Then, for any n N and any e ¯ , e ¯ , E n
k ( A n | e ¯ , A n | e ¯ ) n · Ent * ( A ) 1 · q ( e ¯ ) q ( e ¯ ) 1 .
Proof. 
To prove the lemma, we construct a coupling between A n | e ¯ and A n | e ¯ in the following manner. Note that there exists a permutation σ S n , such that
| i : e i e σ i | = n 2 · q ( e ¯ ) q ( e ¯ ) 1 .
Let
I = i : e i = e σ i I ˜ = i : e i e σ i .
Using that | I ˜ | = n 2 · q ( e ¯ ) q ( e ¯ ) 1 , we can estimate
k A n | e ¯ , A n | e ¯ = k i = 1 n ( A | e i ) , i = 1 n ( A | e σ i ) i I kd ( A | e i = A | e σ i ) + i I ˜ kd ( A | e i A | e σ i ) n · Ent * ( A ) 1 · q ( e ¯ ) q ( e ¯ ) 1 ,
where A = B denotes the isomorphism coupling of two naturally isomorphic diagrams, while A B denotes the “independence” coupling. □
Lemma 2. 
Let A be a G -diagram of probability spaces and E be a space in A . Then,
E n k ( A n , A n | e ¯ ) d p ( e ¯ ) 2 n · [ [ G ] ] · Ent ( E ) + o ( n ) .
Proof. 
First, we apply Proposition 1 slicing the first argument:
E n k ( A n , A n | e ¯ ) d p ( e ¯ ) E n E n k ( A n | e ¯ , A n | e ¯ ) d p ( e ¯ ) d p ( e ¯ ) + 2 n · [ [ G ] ] · Ent ( E ) .
We will argue now that the double integral on the right-hand side grows sub-linearly with n. We estimate the double integral by applying Lemma 1 to the integrand
E n E n k ( A n | e ¯ , A n | e ¯ ) d p ( e ¯ ) d p ( e ¯ ) E n E n n · [ [ G ] ] · | Ent * ( A ) | 1 · | q ( e ¯ ) q ( e ¯ ) | 1 d p ( e ¯ ) d p ( e ¯ ) = n · [ [ G ] ] · | Ent * ( A ) | 1 · Δ E Δ E | π π | 1 d τ n ( π ) d τ n ( π ) = o ( n ) ,
where the convergence to zero of the last double integral follows from Sanov’s theorem. □
Corollary 2. 
Let A be a G -diagram and E a probability space included in A . Then,
κ A , [ A | E ] 2 [ [ G ] ] · Ent ( E ) .
Proof. 
Let n N . Then,
κ A , [ A | E ] = 1 n κ A n , [ A n | E n ] = 1 n κ A n , E n A n | e ¯ d p ( e ¯ ) 1 n E n κ A n , A n | e ¯ d p ( e ¯ ) = 1 n E n κ ( A n , A n | e ¯ ) d p ( e ¯ ) 2 · [ [ G ] ] · Ent ( E ) + o ( n 0 ) ,
where we used Lemma 2 and the fact that κ k in the last line. We finish the proof by taking the limit n . □
Proof of Proposition 2.
We start with a note on general terminology: a reduction f : A B of probability spaces can also be considered as a fan F : = ( A = A f B ) . Then, the entropy distance of f is
kd ( f ) : = kd ( F ) = Ent A Ent B .
If the reduction f is a part of a bigger diagram containing also space U, then the following inequality holds:
U kd ( f | u ) d p ( u ) kd ( f ) .
Let K Prob G , Λ 2
K = ( X f Z g Y ) Prob G , Λ 2 = Prob Λ 2 , G
be an optimal coupling between X and Y . It can also be viewed as a G -diagram of two fans, K = K i i G , each of which is a minimal coupling between X i and Y i . Among them is the minimal fan W : = K ι = ( U f ι W g ι V ) .
We use the triangle inequality to bound the distance κ [ X | U ] , [ Y | V ] by four summands as follows:
κ [ X | U ] , [ Y | V ] κ [ X | U ] , [ Z | U ] + κ [ Z | U ] , [ Z | W ] + κ [ Z | W ] , [ Z | V ] + κ [ Z | V ] , [ Y | V ] .
We will estimate each of the four summands separately. The bound for the first one is as follows:
κ ( [ X | U ] , [ Z | U ] ) = κ U X | u d p ( u ) , U Z | u d p ( u ) U κ X | u , Z | u d p ( u ) = U κ X | u , Z | u d p ( u ) U k X | u , Z | u d p ( u ) U kd ( f | u ) d p ( u ) i G U kd ( f i | u ) d p ( u ) = i G kd ( f i ) = kd ( f ) .
An analogous calculation shows that
κ [ Z | V ] , [ Y | V ] kd ( g ) .
To bound the second summand, we will use Corollary 2:
κ [ Z | U ] , [ Z | W ] = κ U Z | u d p ( u ) , W Z | w d p ( w ) = κ U Z | u d p ( u ) , U W | u Z | w d p ( w | u ) d p ( u ) U κ Z | u , W | u Z | w d p ( w | u ) d p ( u ) .
We will now use Corollary 2 with A = Z | u and E = W | u to estimate the integrand. Then,
κ [ Z | U ] , [ Z | W ] = U κ Z | u , W | u Z | w d p ( w | u ) d p ( u ) 2 [ [ G ] ] · U Ent ( W | u ) d p ( u ) 2 [ [ G ] ] · Ent ( W | U ) 2 [ [ G ] ] · kd ( f ) .
Similarly,
κ [ Z | W ] , [ Z | V ] 2 [ [ G ] ] · kd ( g ) .
Combining the estimates, we obtain
κ [ X | U ] , [ Y | V ] ( 2 [ [ G ] ] + 1 ) · ( kd ( f ) + kd ( g ) ) = ( 2 [ [ G ] ] + 1 ) · k ( X , Y ) .

3.5. Tropical Conditioning

Let [ X ] be a tropical G -diagram and [ U ] = [ X ι ] for some ι G . Choose a representative X ( n ) n N 0 and denote u ( n ) : = X ι ( n ) . We define now a conditioned diagram [ X | U ] by the following limit:
[ X | U ] : = lim n 1 n [ X ( n ) | U ( n ) ] .
Proposition 1 guarantees that the limit exists and is independent of the choice of representative. For a fixed ι G , the conditioning is a linear Lipschitz map of
[ · | · ι ] : Prob [ G ] Prob [ G ] .

Author Contributions

Investigation, R.M. and J.W.P. All authors have read and agreed to the published version of the manuscript.

Funding

Open Access funding was provided by the Max Planck Society.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Matveev, R.; Portegies, J.W. Asymptotic dependency structure of multiple signals. Inf. Geom. 2018, 1, 237–285. [Google Scholar] [CrossRef]
  2. Matveev, R.; Portegies, J.W. Tropical diagrams of probability spaces. Inf. Geom. 2020, 3, 61–88. [Google Scholar] [CrossRef]
  3. Ay, N.; Bertschinger, N.; Der, R.; Güttler, F.; Olbrich, E. Predictive information and explorative behavior of autonomous robots. Eur. Phys. J. B 2008, 63, 329–339. [Google Scholar] [CrossRef]
  4. Bertschinger, N.; Rauh, J.; Olbrich, E.; Jost, J.; Ay, N. Quantifying unique information. Entropy 2014, 16, 2161–2183. [Google Scholar] [CrossRef]
  5. Friston, K. The free-energy principle: A rough guide to the brain? Trends Cogn. Sci. 2009, 13, 293–301. [Google Scholar] [CrossRef] [PubMed]
  6. Kingma, D.P.; Welling, M. Auto-encoding variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  7. Steudel, B.; Ay, N. Information-theoretic inference of common ancestors. Entropy 2015, 17, 2304–2327. [Google Scholar] [CrossRef]
  8. Dijk, S.G.V.; Polani, D. Informational constraints-driven organization in goal-directed behavior. Adv. Complex Syst. 2013, 16, 1350016. [Google Scholar] [CrossRef]
  9. Matveev, R.; Portegies, J.W. Tropical probability theory and an application to the entropic cone. Kybernetika 2020, 56, 1133–1153. [Google Scholar] [CrossRef]
  10. Kovačević, M.; Stanojević, I.; Šenk, V. On the hardness of entropy minimization and related problems. In Proceedings of the 2012 IEEE Information Theory Workshop, Lausanne, Switzerland, 3–7 September 2012; pp. 512–516. [Google Scholar]
  11. Vidyasagar, M. A metric between probability distributions on finite sets of different cardinalities and applications to order reduction. IEEE Trans. Autom. Control 2012, 57, 2464–2477. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Matveev, R.; Portegies, J.W. Conditioning in Tropical Probability Theory. Entropy 2023, 25, 1641. https://doi.org/10.3390/e25121641

AMA Style

Matveev R, Portegies JW. Conditioning in Tropical Probability Theory. Entropy. 2023; 25(12):1641. https://doi.org/10.3390/e25121641

Chicago/Turabian Style

Matveev, Rostislav, and Jacobus W. Portegies. 2023. "Conditioning in Tropical Probability Theory" Entropy 25, no. 12: 1641. https://doi.org/10.3390/e25121641

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop