Consider a scenario in which a message, encoded as
, must be decoded at the primary receiver
while it is also seen at the unintended/secondary receiver
for which it is interference, as shown in
Figure 6a. The transmitter wishes to maximize its communication rate, while subject to a constraint on the disturbance it inflicts on the secondary receiver, and where the disturbance is measured by some function
. It is common to refer to such a scenario as communication with a disturbance constraint. The choice of
depends on the application one has in mind. For example, a common application is to limit the interference that the primary user inflicts on the secondary. In this case, two possible choices of
are the mutual information
and the MMSE
, considered in [
69,
70], respectively. In what follows we review these two possible measures of disturbance, so as to explain the advantages of the MMSE as a measure of disturbance that best models the interference.
5.1. Max-I Problem
Consider a Gaussian noise channel and take the disturbance to be measured in terms of the MMSE (i.e.,
), as shown on
Figure 6b. Intuitively, the MMSE disturbance constraint quantifies the remaining interference after partial interference cancellation or soft-decoding have been performed [
47,
70]. Formally, the following problem was considered in [
50]:
Definition 6. (Max-I problem.)
For some The subscript
n in
emphasizes that we consider length
n inputs
. Clearly
is a non-decreasing function of
n. The scenario depicted in
Figure 6b is captured when
in the Max-I problem, in which case the objective function has a meaning of reliable achievable rate.
The scenario modeled by the Max-I problem is motivated by the two-user Gaussian interference channel (G-IC), whose capacity is known only for some special cases. The following strategies are commonly used to manage interference in the G-IC:
Interference is treated as Gaussian noise: in this approach the interference is not explicitly decoded. Treating interference as noise with Gaussian codebooks has been shown to be sum-capacity optimal in the so called very-weak interference regime [
71,
72,
73].
Partial interference cancellation: by using the Han-Kobayashi (HK) achievable scheme [
74], part of the interfering message is jointly decoded with part of the desired signal. Then the decoded part of the interference is subtracted from the received signal, and the remaining part of the desired signal is decoded while the remaining part of the interference is treated as Gaussian noise. With Gaussian codebooks, this approach has been shown to be capacity achieving in the strong interference regime [
75] and optimal within 1/2 bit per channel per user otherwise [
76].
Soft-decoding/estimation: the unintended receiver employs soft-decoding of part of the interference. This is enabled by using non-Gaussian inputs and designing the decoders that treat interference as noise by taking into account the correct (non-Gaussian) distribution of the interference. Such scenarios were considered in [
44,
46,
49], and shown to be optimal to within either a constant or a
gap for all regimes in [
45].
Even though the Max-I problem is somewhat simplified, compared to that of determining the capacity of the G-IC, as it ignores the existence of the second transmission, it can serve as an important building block towards characterizing the capacity of the G-IC [
47,
70], especially in light of the known (but currently uncomputable) limiting expression for the capacity region [
77]:
where
denotes the convex closure operation. Moreover, observe that for any finite
n we have that the capacity region can be inner bounded by
where
The inner bound
will be referred to as the
treating interference as noise (TIN) inner bound. Finding the input distributions
that exhaust the achievable region in
is an important open problem. In
Section 8, for a special case of
, we will demonstrate that
is within a constant or
from the capacity
. Therefore, the Max-I problem, denoted by
in (48), can serve as an important step in characterizing the structure of optimal input distributions for
. We also note that in [
47,
70] it was conjectured that the optimal input for
is discrete. For other recent works on optimizing the TIN region in (51), we refer the reader to [
43,
46,
49,
78,
79] and the references therein.
The importance of studying models of communication systems with disturbance constraints has been recognized previously. For example, in [
69] Bandemer et al. studied the following problem related to the Max-I problem in (48).
Definition 7. (Bandemer et al. problem [69]) For some In [
69] it was shown that the optimal solution for
, for any
n, is attained by
where
; here
is such that the most stringent constraint between (52b) and (52c) is satisfied with equality. In other words, the optimal input is independent and identically distributed (i.i.d.) Gaussian with power reduced such that the disturbance constraint in (52c) is not violated.
Theorem 7 ([
69])
. The rate-disturbance region of the problem in (52) is given bywith equality if and only if where . Measuring the disturbance with the mutual information as in (52), in contrast to the MMSE as in (48), suggests that it is always optimal to use Gaussian codebooks with reduced power without any rate splitting. Moreover, while the mutual information constraint in (52) limits the amount of information transmitted to the unintended receiver, it may not be the best choice for measuring the interference, since any information that can be reliably decoded by the unintended receiver is not really interference. For this reason, it has been argued in [
47,
70] that the Max-I problem in (48) with the MMSE disturbance constraint is a more suitable building block to study the G-IC, since the MMSE constraint accounts for the interference, and captures the key role of rate splitting.
We also refer the reader to [
80] where, in the context of discrete memoryless channels, the disturbance constraint was modeled by controlling the type (i.e., empirical distribution) of the interference at the secondary user. Moreover, the authors of [
80] were able to characterize the tradeoff between the rate and the type of the induced interference by exactly characterizing the capacity region of the problem at hand.
We first consider a case of the Max-I problem when .
5.2. Characterization of as
For the practically relevant case of
, which has an operational meaning,
has been characterized in [
70] and is given by the following theorem.
Theorem 8 ([
70])
. For any and which is achieved by using superposition coding with Gaussian codebooks. The proof of the achievability part of Theorem 8 is by using superposition coding and is outside of the scope of this work. The interested reader is referred to [
63,
70,
81] for a detailed treatment of MMSE properties of superposition codes.
Next, we show a converse proof of Theorem 8. In addition, to the already familiar use of the LMMSE bound technique, as in the wiretap channel in
Section 4.1, we also show an application of the SCPP bound. The proof for the case of
follows by ignoring the MMSE constraint at
and using the LMMSE upper bound
Next, we focus on the case of
where the last inequality follows by upper bounding the integral over
by the LMMSE bound in (11) and by upper bounding the integral over
using the SCPP bound in (21).
Figure 7 shows a plot of
in (54) normalized by the capacity of the point-to-point channel
. The region
(flat part of the curve) is where the MMSE constraint is inactive since the channel with
can decode the interference and guarantee zero MMSE. The regime
(curvy part of the curve) is where the receiver with
can no-longer decode the interference and the MMSE constraint becomes active, which in practice is the more interesting regime because the secondary receiver experiences “weak interference” that cannot be fully decoded (recall that in this regime superposition coding appears to be the best achievable strategy for the two-user Gaussian interference channel, but it is unknown whether it achieves capacity [
76]).
5.3. Proof of the Disturbance Constraint Problem with a Mutual Information Constraint
In this section we show that the mutual information disturbance constraint problem in (52) can also be solved via an estimation theoretic approach.
An Alternative Proof of the Converse Part of Theorem 7. Observe that, similarly to the Max-I problem, the interesting case of the
is the “weak interference” regime (i.e.,
). This, follows since for the “strong interference” regime (i.e.,
) the result follows trivially by the data processing inequality
and maximizing (55) under the power constraint. To show Theorem 7, for the case of
, observe that
where the inequality on the right is due to the power constraint on
. Therefore, there exists some
such that
Using the I-MMSE, (57) can be written as
From (58) and SCPP property we conclude that
and
are either equal for all
t, or cross each other once in the region
. In both cases, by the SCPP, we have
We are now in the position to bound the main term of the disturbance constrained problem. By using the I-MMSE relationship the mutual information can be bounded as follows:
where the bound in (60) follows by the inequality in (59). The proof of the converse is concluded by establishing that the maximum value of
in (61) is given by
which is a consequence of the bound
.
This concludes the proof of the converse. ☐
The achievability proof of Theorem 7 follows by using an i.i.d. Gaussian input with power . This concludes the proof of Theorem 7.
In contrast to the proof in [
69] which appeals to the EPI, the proof outlined here only uses the SCPP and the I-MMSE. Note, that unlike the proof of the converse of the Max-I problem, which also requires the LMMSE bound, the only ingredient in the proof of the converse for
is a clever use of the SCPP bound. In
Section 6, we will make use of this technique and show a converse proof for the scalar Gaussian broadcast channel.
Another observation is that the achievability proof of the
holds for an arbitrary finite
n while the achievability proof of the Max-I problem holds only as
. In the next section, we demonstrate techniques for how to extend the achievability of the Max-I problem to the case of finite
n. These techniques will ultimately be used to show an approximate optimality of the TIN inner bound for the two-user G-IC in
Section 8.
5.4. Max-MMSE Problem
The Max-I problem in (48) is closely related to the following optimization problem.
Definition 8. (Max-MMSE problem [50,82]) For some The authors of [
63,
70] proved that
achieved by superposition coding with Gaussian codebooks. Clearly there is a discontinuity in (63) at
for
. This fact is a well known property of the MMSE, and it is referred to as a
phase transition [
63].
The LMMSE bound provides the converse solution for in (63) in the regime . An interesting observation is that in this regime the knowledge of the MMSE at is not used. The SCPP bound provides the converse in the regime and, unlike the LMMSE bound, does use the knowledge of the value of MMSE at .
The solution of the Max-MMSE problem provides an upper bound on the Max-I problem (for every
n including in the limit as
), through the I-MMSE relationship
The reason is that in the Max-MMSE problem one maximizes the integrand in the I-MMSE relationship for every , and the maximizing input may have a different distribution for each . The surprising result is that in the limit as we have equality, meaning that in the limit there exists an input that attains the maximum Max-MMSE solution for every . In other words, the integration of over results in . In view of the relationship in (64) we focus on the problem.
Note that SCPP gives a solution to the Max-MMSE problem in (62) for
and any
as follows:
achieved by
.
However, for
, where the LMMSE bound (11) is used without taking the constraint into account, it is no longer tight for every
. Therefore, the emphasis in the treatment of the Max-MMSE problem is on the regime
. In other words, the phase transition phenomenon can only be observed as
, and for any finite
n the LMMSE bound on the MMSE at
must be sharpened, as the MMSE constraint at
must restrict the input in such a way that would effect the MMSE performance at
. We refer to the upper bounds in the regime
as complementary SCPP bounds. Also, for any finite
n,
is a continuous function of
[
30]. Putting these two facts together we have that, for any finite
n, the objective function
must be continuous in
and converge to a function with a jump-discontinuity at
as
. Therefore,
must be of the following form:
for some
. The goal is to characterize
in (66) and the continuous function
such that
and give scaling bounds on the width of the phase transition region defined as
In other words, the objective is to understand the behavior of the MMSE phase transitions for arbitrary finite n by obtaining complementary upper bounds on the SCPP. We first focus on upper bounds on .
Theorem 9. (D-Bound [50]) For any and , we have The proof of Theorem 9 can be found in [
50] and relies on developing bounds on the derivative of the MMSE with respect to the SNR.
Theorem 10. (M-Bound [25]) For ,where and The bounds in (69a) and in (70a) are shown in
Figure 8. The key observation is that the bounds in (69a) and in (70a) are sharper versions of the LMMSE bound that take into account the value of the MMSE at
. It is interesting to observe how the bounds converge with
n going to
∞.
The bound in (70a) is asymptotically tighter than the one in (69a). It can be shown that the phase transition region shrinks as
for (70a), and as
for the bound in (69a). It is not possible in general to assert that (70a) is tighter than (69a). In fact, for small values of
n, the bound in (69a) can offer advantages, as seen for the case
shown in
Figure 8b. Another advantage of the bound in (69a) is its analytical simplicity.
With the bounds in (69a) and in (70a) at our disposal we can repeat the converse proof outline in (61).
5.5. Mixed Inputs
Another question that arises, in the context of finite n, is how to mimic the achievability of superposition codes? Specifically, how to select an input that will maximize when .
We propose to use the following input, which in [
45] was termed a
mixed input:
where
and
are independent. The parameter
and the distribution of
are to be optimized over.
The behavior of the input in (71) exibits many properties of superposition codes and we will see that the discrete part will behave as the common message and the Gaussian part will behave as the private message.
The input exhibits a decomposition property via which the MMSE and the mutual information can be written as the sum of the MMSE and the mutual information of the and components, albeit at different SNR values.
Proposition 7 ([
50])
. For defined in (71) we have that Observe that Proposition 7 implies that, in order for mixed inputs (with
) to comply with the MMSE constraint in (48c) and (62c), the MMSE of
must satisfy
Proposition 7 is particularly useful because it allows us to design the Gaussian and discrete components of the mixed input independently.
Next, we evaluate the performance of
in
for the important special case of
.
Figure 9 shows upper and lower bounds on
where we show the following:
The upper bound in (63) (solid red line);
The upper D-bound (69a) (dashed cyan line) and upper M-bound (dashed red line) (70a);
The Gaussian-only input (solid green line), with , where the power has been reduced to meet the MMSE constraint;
The mixed input (blue dashed line), with the input in (71). We used Proposition 7 where we optimized over
for
. The choice of
is motivated by the scaling property of the MMSE, that is,
, and the constraint on the discrete component in (74). That is, we chose
such that the power of
is approximately
while the MMSE constraint on
in (74) is not equal to zero. The input
used in
Figure 9 was found by a local search algorithm on the space of distributions with
, and resulted in
with
, which we do not claim to be optimal;
The discrete-only input (Discrete 1 brown dashed-dotted line), with
with , that is, the same discrete part of the above mentioned mixed input. This is done for completeness, and to compare the performance of the MMSE of the discrete component of the mixed input with and without the Gaussian component; and
The discrete-only input (Discrete 2 dotted magenta line), with
with , which was found by using a local search algorithm on the space of discrete-only distributions with points.
The choice of is motivated by the fact that it requires roughly points for the PAM input to approximately achieve capacity of the point-to-point channel with SNR value .
On the one hand,
Figure 9 shows that, for
, a Gaussian-only input with power reduced to
maximizes
in agreement with the SCPP bound (green line). On the other hand, for
, we see that discrete-only inputs (brown dashed-dotted line and magenta dotted line) achieve higher MMSE than a Gaussian-only input with reduced power. Interestingly, unlike Gaussian-only inputs, discrete-only inputs do not have to reduce power in order to meet the MMSE constraint. The reason discrete-only inputs can use full power, as per the power constraint only, is because their MMSE decreases fast enough (exponentially in SNR) to comply with the MMSE constraint. However, for
, the behavior of the MMSE of discrete-only inputs, as opposed to mixed inputs, prevents it from being optimal; this is due to their exponential tail behavior. The mixed input (blue dashed line) gets the best of both (Gaussian-only and discrete-only) worlds: it has the behavior of Gaussian-only inputs for
(without any reduction in power) and the behavior of discrete-only inputs for
. This behavior of mixed inputs turns out to be important for the Max-I problem, where we need to choose an input that has the largest area under the MMSE curve.
Finally,
Figure 9 shows the achievable MMSE with another discrete-only input (Discrete 2, dotted magenta line) that achieves higher MMSE than the mixed input for
but lower than the mixed input for
. This is again due to the tail behavior of the MMSE of discrete inputs. The reason this second discrete input is not used as a component of the mixed inputs is because this choice would violate the MMSE constraint on
in (74). Note that the difference between Discrete 1 and 2 is that, Discrete 1 was found as an optimal discrete component of a mixed input (i.e.,
), while Discrete 2 was found as an optimal discrete input without a Gaussian component (i.e.,
).
We conclude this section by demonstrating that an inner bound on with the mixed input in (71) is to within an additive gap of the outer bound.
Theorem 11 ([
50])
. A lower bound on with the mixed input in (71)
, with and with input parameters as specified in Table 1, is to within . We refer the reader to [
50] for the details of the proof and extension of Theorem 11 to arbitrary
n.
Please note that the gap result in Proposition 11 is constant in
(i.e., independent of
) but not in
.
Figure 10 compares the inner bounds on
, normalized by the point-to-point capacity
, with mixed inputs (dashed magenta line) in Proposition 11 to:
The upper bound in (54) (solid red line);
The upper bound from integration of the bound in (69a) (dashed blue line);
The upper bound from integration of the bound in (70a) (dashed red line); and
The inner bound with , where the reduction in power is necessary to satisfy the MMSE constraint (dotted green line).
Figure 10 shows that Gaussian inputs are sub-optimal and that mixed inputs achieve large degrees of freedom compared to Gaussian inputs. Interestingly, in the regime
, it is approximately optimal to set
, that is, only the discrete part of the mixed input is used. This in particular supports the conjecture in [
70] that discrete inputs may be optimal for
and
. For the case
our results partially refute the conjecture by excluding the possibility of discrete inputs with finitely many points from being optimal.
The key intuition developed in this section about the mixed input and its close resemblance to superposition coding will be used in
Section 8 to show approximate optimality of TIN for the two-user G-IC.