Next Article in Journal
Decision, Inference, and Information: Formal Equivalences Under Active Inference
Previous Article in Journal
GAME-YOLO: Global Attention and Multi-Scale Enhancement for Low-Visibility UAV Detection with Sub-Pixel Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Capacity Gains of Gaussian Channels with Unstable Versus Stable Autoregressive Noise

by
Charalambos D. Charalambous
1,*,
Christos Kourtellaris
1,
Stelios Louka
1 and
Sergey Loyka
2
1
Department of Electrical and Computer Engineering, University of Cyprus, P.O. Box 20537, CY-1678 Nicosia, Cyprus
2
School of Electrical Engineering and Computer Science, University of Ottawa, Ottawa, ON K1N 6N5, Canada
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(12), 1264; https://doi.org/10.3390/e27121264
Submission received: 16 September 2025 / Revised: 7 December 2025 / Accepted: 8 December 2025 / Published: 18 December 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

In this paper, we consider Cover’s and Pombra’s formulation of feedback capacity of additive Gaussian noise (AGN) channels, with jointly Gaussian nonstationary and nonergodic noise. We derive closed-form feedback capacity formulas, using Karush–Kuhn–Tucker (KKT) conditions and convergence properties of difference Riccati equations to limiting algebraic Riccati equations of filtering theory, for unstable and stable autoregressive (AR) noise. Surprisingly, the capacity formulas depend on the parameters of the AR noise, its pole c ( , ) and noise variance K W ( 0 , ) , and the total transmit power κ [ 0 , ) , indicating substantial gains for the unstable noise region c 2 ( 1 , ) , κ > κ m i n = K W 1 + 4 c 2 3 2 c 2 1 2 compared to its complement region. In particular, feedback capacity is distinguished by three regimes, as follows. Regime 1, c 2 ( 1 , ) , κ > κ m i n : the optimal channel input includes an innovations part, the capacity increases as | c | > 1 increases, while κ m i n and the allocated transmit power decrease. Regime 2, c 2 ( 1 , ) , κ κ m i n , Regime 3, c [ 1 , 1 ] , κ [ 0 , ) (complement of Regime 1): the innovations part of the optimal channel is asymptotically zero and the capacity is fundamentally different compared to Regime 1. The differences of capacity formulas for Regimes 1, 2 and 3 are directly related to their operational meaning: (i) Regime 1 is an ergodic capacity while Regimes 2 and 3 are nonergodic capacities; (ii) Regime 1 is achieved by an asymptotically stationary channel input with a non-zero innovations part, while Regimes 2 and 3 are achieved by an asymptotically zero innovations part. The gains of capacity for Regime 1 are attributed to the high correlation of noise samples compared to stable noise and the use of an informative innovations part by the optimal channel input, which make possible the prediction of future noise samples from past samples, unlike memoryless noise. Our results provide answers to certain open questions regarding the validity of capacity formulas of stable noise that appeared in the literature.

1. Introduction

The main objective of this paper is to show fundamental gains of feedback and nonfeedback capacity of additive Gaussian noise channels, driven by nonstationary and nonergodic noise compared to stationary noise. A secondary objective is to provide answers to the open question of the recent comment paper [1] [Abstract, Conclusion] concerning the correctness of the feedback capacity characterization [2] [Theorem 6.1], by identifying conditions of its validity, which are consistent with the technical analysis of our application example, and indicate its implications on the operational meaning of capacity.

1.1. The AGN Channel with Nonstationary and Nonergodic Noise

We consider the additive Gaussian noise (AGN) channel depicted in Figure 1, driven by nonstationary and nonergodic noise as in Cover and Pombra [3],
Y t = X t + V t , t = 1 , , n , 1 n E t = 1 n ( X t ) 2 κ , κ [ 0 , )
where X n = ( X 1 , X 2 , , X n ) T and Y n = ( Y 1 , Y 2 , , Y n ) T are vectors of channel input and output random variables (RVs), X t : Ω R , t = 1 , , n and Y t : Ω R , t = 1 , , n , respectively, and V n = ( V 1 , , V n ) T is a vector of nonstationary and nonergodic, jointly Gaussian noise, and V t : Ω R , t = 1 , , n , with mean μ V n , and covariance K V n , denoted by V n N ( μ V n , K V n ) . Here, V 1 N ( 0 , K V 1 ) is the initial noise RV, with induced distribution P V 1 , called the initial state of the channel. Without loss of generality we assume μ V n = 0 .
Feedback (resp. Nonfeedback) Code. We wish to communicate to anyone the uniformly distributed messages, W : Ω M ( n ) = 1 , 2 , , M n , M n = 2 n R , by using an encoder with feedback X 1 = e 1 ( W ) , X 2 = e 2 ( W , X 1 , Y 1 ) , , X n = e n ( W , X n 1 , Y n 1 ) (resp., without feedback X 1 = e 1 ( W ) , X 2 = e 2 ( W , X 1 ) e ¯ 2 ( W ) , , X n = e n ( W , X n 1 ) e ¯ n ( W ) ), and a decoder with arbitrary small average error probability P e r r o r ( n ) ( P V 1 ) = P d n ( Y n ) W = 1 M n w = 1 M n P d n ( Y n ) w | W = w , after n sufficiently large transmissions, under the assumption of independence of W and V n (the dependence of P e r r o r ( n ) ( P V 1 ) on P V 1 emphasizes the fact that the error probability, in general, depends on the RV V 1 N ( 0 , K V 1 ) ). A feedback rate R is called an achievable rate, if there exists a sequence of encoders and decoders satisfying lim n P e r r o r ( n ) ( P V 1 ) = 0 and lim inf n 1 n log M n R . The operational definition of the feedback capacity is C F B ( κ , P V 1 ) sup R | R i s a c h i e v a b l e , i.e., in general, the capacity depends on P V 1 . If C F B ( κ , P V 1 ) = C F B ( κ ) , P V 1 , then it is independent of the choice of the initial distribution P V 1 and P Y 1 . Similarly, the nonfeedback capacity is denoted by C ( κ , P V 1 ) .
In 1989, Cover and Pombra [3], derived a time-domain (TD) information theoretic characterization of feedback capacity C ( κ , P V 1 ) (for arbitrary nonstationary and nonergodic noise V n , not necessarily finite-dimensional), expressed as an asymptotic limit, C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) , where C n , F B ( κ , P V 1 ) is called the n - finite transmission or block length feedback information ( n - FTFI) capacity, given as follows.
C n , F B ( κ , P V 1 ) = sup P X t | X t 1 , Y t 1 , t = 1 , , n , 1 n E t = 1 n ( X t ) 2 κ 1 n H ( Y n ) H ( V n )
= sup ( B n , K Z ¯ n ) | 1 n T r a c e B n K V n ( B n ) T + K Z ¯ n κ 1 2 n log | B n + I n K V n B n + I n T + K Z ¯ n | | K V n |
where B n R n × n is a lower diagonal matrix, I n is an n × n identity matrix, and K Z ¯ n is a covariance matrix of a jointly Gaussian process Z ¯ n N ( 0 , K Z ¯ n ) , and H ( X ) denotes (differential) entropy of RV X.
In [3], it is shown that C n , F B ( κ , P V 1 ) is achieved by a channel input process X n with time-varying coefficients given by
X t = j = 1 t 1 b t , j V j + Z ¯ t , Y t = j = 1 t 1 b t , j V j + Z ¯ t + V t , t = 1 , , n ,
X n = B n V n + Z ¯ n , Y n = B n + I n V n + Z ¯ n ,
Z ¯ n N ( 0 , K Z ¯ n ) , K Z ¯ n 0
where Z ¯ n is independent of the noise V n . The reason [3] expressed X t in terms of V t 1 and not ( X t 1 , Y t 1 ) is because ( X t 1 , Y t 1 ) uniquely defines ( V t 1 , Y t 1 ) and vice-versa, t , and hence ( X t 1 , Y t 1 ) can be expressed in terms of V t 1 and the jointly Gaussian process Z ¯ t , t = 1 , , n (this should not be confused by an innovations process, which in the standard terminology is an orthogonal process).
Cover and Pombra [3] also derived a TD information theoretic characterization for nonfeedback capacity C ( κ , P V 1 ) , for the same nonstationary and nonergodic noise V n , also expressed as an asymptotic limit, C ( κ , P V 1 ) = lim n C n ( κ , P V 1 ) , where C n ( κ , P V 1 ) is called the n - finite transmission or block length without feedback information ( n - FTwFI) capacity, given as follows.
C n ( κ , P V 1 ) = sup P X t | X t 1 , t = 1 , , n , 1 n E t = 1 n ( X t ) 2 κ 1 n H ( Y n ) H ( V n )
= sup K Z ¯ n | 1 n T r a c e K Z ¯ n κ 1 2 n log | K Z ¯ n + K V n | | K V n | .
The optimal channel input process X n that achieves C n ( κ , P V 1 ) is obtained directly from (4)–(6), by letting B n = 0 , Ref. [3] [Equations (14)–(16)], i.e., X n = Z ¯ n , Y n = Z ¯ n + V n .
The main objective of this paper is to understand the implications of nonstationary and nonergodic Gaussian noise on feedback and nonfeedback capacity compared to the corresponding capacity of stationary or asymptotically stationary and ergodic noise.
The main result of the paper are the gains on feedback and nonfeedback capacity of nonstationary and nonergodic autoregressive noise compared to the corresponding capacity of stationary or asymptotically stationary and ergodic noise.
Our derivations also provide insight and answers to the recent comment paper by Derpich and Ostergaard [1], which identified gaps in the proofs of the frequency-domain (FD) and TD characterizations of feedback capacity [2] [Theorem 4.1] and [2] [Theorem 6.1] for stationary or asymptotically stationary noise V n . In fact, from our analysis it follows that the TD characterization of feedback capacity [2] [Theorem 6.1] is correct, provided certain conditions hold. We identify these conditions.

1.2. Motivation

The current study is motivated by the fundamental limitations of capacity analysis of AGN channels driven by stationary or asymptotically stationary Gaussian noise (stable) V t , t = 1 , 2 , . To make our point precise, we recall Shannon’s FD information theoretic characterization of nonfeedback capacity of AGN channels driven by stationary or asymptotically stationary (stable) Gaussian noise V t , t = 1 , 2 , with power spectral density (PSD), S V ( e j θ ) , θ [ π , π ] .
The nonfeedback capacity is described by the well-known water-filling solution of the optimal PSD of the channel input S X ( e j θ ) , θ [ π , π ] , as follows [4,5,6].
C = C S h a n n o n = 1 4 π π π log max 1 , λ S V ( e j θ ) d θ 1 2 π π π S X ( e j θ ) d θ = 1 2 π π π max 0 , λ S V ( e j θ ) d θ = κ , λ 0 , S X ( e j θ ) = max 0 , λ S V ( e j θ )
where λ is the Lagrange multiplier associated with the transmit average power constraint. If the noise PSD is white, i.e., S V ( e j θ ) = K V , θ [ π , π ] , then
C = C S h a n n o n = 1 2 log max 1 , λ K V , max 0 , λ K V = κ , S X ( e j θ ) = K X = max 0 , λ K V .
The above water-filling formula is restricted to stationary or asymptotically stationary noise, i.e., the pole c of an autoregressive unit memory noise needs to satisfy | c | < 1 . As | c | approaches 1 the noise power approaches + , and this imposes a limitation on the use of formulae C S h a n n o n . In this paper, we show that we can use the time-domain formulae for C to compute simple lower bounds on C for nonstationary and nonergodic channels, especially channels with unstable noise (i.e., their poles lie outside the unit disk in the space of complex numbers).

1.3. Literature Review

The literature on feedback and nonfeedback capacity and bounds for AGN channels with stationary or asymptotically stationary noise is very extensive. However, corresponding results for nonstationary and nonergodic noise are not sufficiently developed.
Information Theoretic Bounds on Feedback Capacity. In the early 1970s, Butman, Tienan and Schalkwijk, and Wolfowitz [7,8,9,10] derived TD information theoretic bounds on feedback capacity C F B ( κ , P V 1 ) of the AGN channel (1), when the channel noise V n is a stable autoregressive (AR) Gaussian process. These bounds are derived by invoking linear feedback coding schemes of communicating Gaussian random variables (RVs), Θ : Ω R and digital messages W : Ω { 1 , 2 , , 2 n R } , under the assumption V 1 = v 1 is known to the encoder and the decoder (see [7,8,9,10]).
Information Theoretic Characterization of Feedback and Nonfeedback Capacity for Nonstationary and Nonergodic Noise. In 1989, Cover and Pombra [3] derived a TD information theoretic characterization of feedback and nonfeedback capacity for the AGN channel (1), when the noise V n is arbitrary nonstationary and nonergodic, as introduced earlier. Although, C n , F B ( κ , P Y 1 ) and C n ( κ , P Y 1 ) are compactly represented, to this date their computation for any finite n, even for the stationary (i.e., stable) AR noise, is very challenging (see [11] for discussion). However, Ref. [3] influenced many subsequent studies [1,2,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25].
Sequential Information Theoretic Characterization of Feedback for Stationary and Asymptotically Stationary Noise. In 2007, Yang, Kavcic and Tatikonda [20] considered the AGN channel (1), when the noise is stationary or asymptotically stationary described by a power spectral density (PSD) with stable poles and zeros. In [20] [Section Section II.C, Equations (19) and (20)], the authors introduced specific stable state space realizations of the noise, with state S t : Ω R n s , t = 1 , , n , which satisfies two important conditions [20] [Section II.C, I–III], specifically,
Condition 1.  The initial state of the noise or channel S 1 = s is known to the encoder and decoder.
Condition 2.  Given a fixed initial state S 1 = s , known to the encoder and decoder, at each t, the state space realization of the noise is such that the noise V t 1 uniquely defines the state of the noise S t and vice-versa, t .
Under Conditions 1 and 2, Ref. [20] [Theorem 1] presents an equivalent TD sequential characterization of Cover’s and Pombra’s n - FTFI capacity, C n , F B ( κ , P Y 1 ) , which is much simpler. The simplification is attributed to Conditions 1 and 2, which imply the optimal channel input conditional distribution that achieves C n , F B ( κ , P Y 1 ) can be expressed causally in terms of the state of the noise and the channel output, i.e., ( S n , Y n ) and an innovations process Z n . This gave rise to [20] [Theorem 5], which is a sequential characterization of the n - FTFI capacity, C n , F B ( κ , P Y 1 ) , expressed with respect to (w.r.t.) a single matrix difference Riccati equation (DRE) of Kalman-filtering, parametrized by two variables which need to be optimized. In [20] [Section IV], the asymptotic limit, C F B ( κ , P Y 1 ) = lim n C n , F B ( κ , P Y 1 ) , is also investigated, called “maximal information rate”, Ref. [20] [Theorem 6, I m a x ]; this is expressed w.r.t. a single matrix algebraic Riccati equation (ARE) of Kalman-filtering, parametrized by the two variables that need to be optimized.
In 2010, Kim [2] re-visited the AGN channel (1), when the noise is stationary or asymptotically stationary noise described by a power spectral density (PSD) with stable poles and zeros (as in [20]), and by a state space realization [2] [Equation (58)] (the realization is equivalent to the one considered in [20], i.e., both correspond to the same PSD of the noise). The main results of [2] are two intriguing characterizations of feedback capacity, C F B ( κ , P Y 1 ) , the FD characterization [2] [Theorem 4.1] and the TD characterization [2] [Theorem 6.1] for state space [2] [Equation (58)]. Similarly to [20] [Theorem 6], the TD characterization of feedback capacity C F B ( κ , P Y 1 ) in [2] [Theorem 6.1] is also expressed w.r.t. a single matrix (ARE of Kalman-filtering, but unlike [20] [Theorem 6], it is parametrized by only a single variable that needs to be optimized. The simplification of C F B ( κ , P Y 1 ) in [2] [Theorem 6.1] by a one-parameter optimization problem is based on [2] [Lemma 6.1], which states the optimal covariance of the innovations part of the channel input is asymptotically zero. The simplification of the FD characterization, Ref. [2] [Theorem 4.1], is based on the statement that the Gaussian process Z ¯ n N ( 0 , K Z ¯ n ) in (4) has a zero optimal PSD.
The simplified FD characterization of feedback capacity [2] [Theorem 4.1] is further analyzed in subsequent papers, [21,22,23].
In 2019, Gattami [24] re-visited Kim’s [2] [Theorem 4.1 and Theorem 6.1], i.e., the simplified FD and TD characterizations of feedback capacity, and presented alternative characterizations in slightly different form [24] [Theorems 1–4]. In particular, Ref. [24] [Theorem 1] presented an FD characterization of feedback capacity, in which the optimal PSD of the Gaussian process Z ¯ n N ( 0 , K Z ¯ n ) in (4) corresponds to white noise, while [24] [Theorem 2] presented a TD characterization of C n , F B ( κ , P V 1 ) , expressed w.r.t. a single matrix DRE of Kalman-filtering. Moreover, Ref. [24] [Theorem 3] presented a TD characterization of feedback capacity C F B ( κ , P V 1 ) , expressed w.r.t. a single matrix ARE of Kalman-filtering, but unlike [2] [Theorem 6.1], the covariance of the innovations part of the optimal input is strictly positive. Moreover, Ref. [24] [Theorem 4] presented a semi-definite programming (SDP) formulation of the optimization problem of C F B ( κ , P V 1 ) of [24] [Theorem 3], using linear matrix inequalities (LMI).
The reader may easily verify that the TD and FD characterizations of C F B ( κ , P V 1 ) in Kim [2] and Gattami [24], are fundamentally different, although both considered the same AGN channel and models of noise. Moreover, the TD characterizations of [2,24] do not impose Conditions 1 and 2 (at least explicitly), contrary to Yang, Kavcic and Tatikonda [20], who stated explicitly that their TD characterization of feedback capacity is based on Conditions 1 and 2. The confusion and misinterpretation of the TD characterizations of feedback capacity of [2,24] compared to [20] are first clarified in the arXiv paper [26,27] (we will discuss this shortly).
Recently, the comment paper by Derpich and Ostergaard [1] identified gaps in proofs of the TD and FD characterizations [2] [Theorem 4.1 and Theorem 6.1], and many other results in [2], which include Corollary 4.4, Theorems 4.6 and 5.3, Propositions 4.7 and 5.1, as well as Remarks 4.5 and 5.2, and Lemma 6.1. However, as stated in the Abstract and Conclusion in [1], the above results of [2] may be valid, but a proof is currently missing.
Sequential Information Theoretic Characterization of Feedback and Nonfeedback Capacity for Nonstationary and Nonergodic Noise. Equivalent TD sequential characterizations of the Cover and Pombra characterization C F B , n ( κ , P V 1 ) are presented in [27] [Theorem 1] (and earlier in [26]). Moreover, Ref. [27] [Theorem 3] presented a TD sequential characterization of C F B , n ( κ , P V 1 ) , when the noise V n is described by a nonstationary and nonergodic state space realization. A fundamental observation is that C F B , n ( κ , P V 1 ) in [27] [Theorem 3], is expressed in terms of two matrix DREs of Kalman-filtering theory instead of one, as previously believed in [2,20,24], irrespective of whether the noise is restricted to stationary or asymptotically stationary. This generalization is attributed to the fact that Conditions 1 and 2 are not imposed; hence, the encoder cannot determine the state of the noise from past channel inputs and output sequences or past noise and output sequences. However, when Conditions 1 and 2 are introduced, it is shown in [27] [Corollary 7, Section 2.5], that the characterizations of C F B , n simplify to the TD characterizations previously derived in [20] [Theorem 1]. Similarly, Ref. [27] [Theorem 7], showed that the asymptotic limit C F B ( κ , P Y 1 ) = lim n C n , F B ( κ , P Y 1 ) is expressed in terms of two matrix AREs of Kalman-filtering theory instead of one, as previously obtained in [2,20,24]. However, in [27] [Corollary 12] it is shown that, if Conditions 1 and 2 are imposed, then under the conditions known as [28,29], (i) detectability and (ii) stabilizability (resp., (iii) unit circle controllability), C F B ( κ , P Y 1 ) reduces to the characterizations previously obtained in [2,20,24] (which did not impose (i) and (ii) (resp., (iii)). Another major observation is that (i) detectability and (ii) stabilizability (resp., (iii) unit circle controllability) are necessary and sufficient conditions for the convergence of matrix DREs to corresponding matrix AREs [28] and the existence of the asymptotic limit C F B ( κ , P Y 1 ) = lim n C n , F B ( κ , P Y 1 ) . Characterizations of C F B , n and C F B for multiple-input multiple-output (MIMO) Gaussian channels are presented in [30], and follow directly from [26]. Preliminary results on an initial study of unstable autoregressive noise are announced in [31], when the initial state is known to the encoder and the decoder and [32]. For nonfeedback capacity we suggest [33].

1.4. Main Contributions

The main contributions of this paper are the following.
(1) The feedback capacity C F B ( κ , P V 1 ) for unstable noise, i.e., nonstationary and nonergodic, is higher compared to the feedback capacity of stable, i.e., stationary or asymptotically stationary noise.
(2) The feedback capacity C F B ( κ , P V 1 ) monotonically increases with respect to (w.r.t.) the unstable poles of the noise, while the total transmit power κ monotonically decreases, contrary to the water-filling nonfeedback capacity of stationary or asymptotically stationary noise.
(3) There are achievable lower bounds on nonfeedback capacity of unstable noise, which exhibit the properties of feedback capacity, i.e., the lower bounds monotonically increase w.r.t. the unstable pole of the noise, without increasing the total transmit power.
(4) We provide a proof that the feedback capacity results for stable noise previously presented in [2,20,24] do not hold for asymptotically stationary noise. On the contrary, the capacity formulas in [2,20,24] correspond to stationary noise, and it is not an ergodic capacity (i.e., it is not independent of the channel initial RVs).
This is due to subtle technical issues related to the asymptotic convergence of information theoretic characterizations of capacity, which are not sufficiently explained in [2,20,24], and often lead to misinterpretation of the results therein. We explain these issues throughout the paper, by clarifying the technical conditions and their various implications (i.e., Remark 2, Remark 9 and Remark 10).
(5) With respect to the comment paper [1], we show that the TD feedback capacity [2] [Theorem 6.1] is correct, provided Conditions 1 and 2 hold, and the optimal innovations part of the channel input is asymptotically zero. This is attributed to the technical issue of existence of feasible solution to the optimization problem that corresponds to the maximal and stabilizing solution of the algebraic Riccati equation of mean-square estimation. However, Ref. [2] [Theorem 6.1] is not equivalent to Cover and Pombra feedback capacity [3] for general stationary noise. Moreover, the claims in [2,24] that their asymptotic formulas hold for asymptotically stationary noise is not accurate; for their capacity formulas to be correct, the noise should be stationary (due to a technical condition related to asymptotic limits, via the maximal and stabilizing solution of the algebraic Riccati equation). Although in the paper we show using Karush–Kuhn–Tucker (KKT) conditions that for stable AR noise the optimal innovations part of the channel input is asymptotically zero, this does not mean the same property holds for [2] [Theorem 6.1] that corresponds to the general stable state space realization of the noise [2] [Equation (58)]. According to our analysis, Ref. [2] [Theorem 6.1] holds provided the optimal covariance of the innovations part of the channel input is asymptotically zero. This can be performed by addressing the optimization problem, i.e., whether a feasible solution exists and under what conditions, for example, by applying KKT conditions.
The rest of the paper is organized as follows.
In Section 2, we present the closed-form expressions of C F B ( κ , P V 1 ) for stable and unstable AR noise, numerical simulations and comparisons to expressions of feedback and nonfeedback capacity that appeared in the literature. We provide an extensive intuitive discussion on many of the technical aspects and their implications on the operational meaning of the capacity.
In Section 3, we present the derivations of asymptotic capacity C F B ( κ , P V 1 ) for stable and unstable AR noise, and lower bounds on asymptotic characterizations of nonfeedback capacity. In this section, we also review properties of convergence of difference Riccati equations to corresponding algebraic equations, which are essential in the analysis of asymptotic capacity C F B ( κ , P V 1 ) . These properties verify our discussion and simulations of Section 2. In this section, we also provide corrections to the TD feedback capacity characterization of [2] [Theorem 6.1].

2. Main Results, Discussion, Simulations and Relations to the Literature

In this section we present the contributions listed in Section 1.4. The derivations are presented in Section 3.

2.1. Notation and Definitions

Throughout the paper, we use the following notation.
Z = { , 1 , 0 , 1 , } , Z + = { 1 , } , Z + n = { 1 , 2 , , n } where n is a finite positive integer.
R = ( , ) , and R m , m Z + is the vector space of tuples of the real numbers.
R n × m , ( n , m ) Z + × Z + is the set of n by m matrices with entries from the set of real numbers.
For a matrix A R n × m , ( n , m ) Z + × Z + , its transpose is denoted by A T ; for n = m the trace of A is denoted by Trace ( A ) , and A 0 (resp. A 0 ) denotes a symmetric positive semi-definite (resp., positive definite) matrix.
Ω , F , P denotes a probability space. Given a random variable (RV) X : Ω R n x , n x Z + n , its induced distribution on R n x is denoted by P X . H ( X ) denotes the differential entropy of the RV X, and I ( X ; Y ) denotes the mutual information between RVs X and Y. X N ( μ X , K X ) , K X 0 denotes a Gaussian RV X, with, mean value μ X = E { X } and covariance matrix K X = c o v ( X , X ) 0 , defined by K X = c o v ( X , X ) = E X E X X E X T . Given another Gaussian random variable Y : Ω R n y , n y Z + n , which is jointly Gaussian distributed with X, i.e., the joint distribution P X , Y is Gaussian, the conditional covariance of X given Y, K X | Y = c o v ( X , X | Y ) , is defined by
K X | Y = c o v ( X , X | Y ) = E X E X | Y X E X | Y T | Y = ( a ) E X E X | Y X E X | Y T
where ( a ) is due to a property of jointly Gaussian RVs.
Next, we introduce standard definitions of nonstationary, asymptotically stationary and stationary autoregressive Gaussian noise [28].
Definition 1. 
(Time-varying, time-invariant stable/unstable AR unit memory Gaussian noise models)
(a) A time-varying (nonstationary and nonergodic) Gaussian unit memory stable/unstable AR noise model, denoted by AR ( c t ) , c t ( , ) , t , is defined by
A R ( c t ) : V t = c t V t 1 + W t , t = 2 , , n , W t N ( 0 , K W t ) , K W t 0 , t = 2 , , n , i n d e p .   G a u s s i a n , W t , t = 2 , , n i n d e p .   o f   V 1 N ( 0 , K V 1 ) , K V 1 0 , c t ( , ) , t = 2 , , n .
The time-varying AR ( c t ) , c t ( , ) , t noise model is called asymptotically time-invariant if lim n c n = c ( , ) and lim n K W n = K W ( 0 , ) .
(b) A time-invariant stable/unstable AR noise model denoted by AR ( c ) , c ( , ) is the restriction of AR ( c t ) to K W t = K W > 0 , t = 2 , , c t = c , t = 2 , .
(c) An unstable (resp., stable) time-invariant AR noise model is the restriction of AR ( c ) to c ( 1 , 1 ) (resp., c ( 1 , 1 ) ).
(d) The stable AR ( c ) , c ( 1 , 1 ) noise model is called asymptotically stationary if the sequence K V n = E V n 2 , n = 1 , 2 , is such that lim n K V n = K V ( 0 , ) , K W > 0 , K V 1 0 , and it is called stationary if K V n = K V 1 for all n = 1 , 2 , .
Remark 1. 
We recall (see [28]) some known facts.
Given an AR ( c t ) , c t ( , ) , t noise the value c n is a measure of correlation between the samples V n and V n 1 , i.e., E V n V n 1 = c n K V n 1 , K V n 1 = E ( V n 1 ) 2 , and the variance satisfies the recursion, K V n = c n 2 K V n 1 + K W n , K V 1 0 , n = 2 , 3 .
For an AR ( c ) noise, K V n = c 2 K V n 1 + K W , K W > 0 , K V 1 0 , n = 2 , , n . The AR ( c ) is asymptotically stationary, if and only if c ( 1 , 1 ) and lim n K V n = K W 1 c 2 , K W > 0 , K V 1 0 . The AR ( c ) is stationary, if and only if c ( 1 , 1 ) and K V t = K V 1 = K W 1 c 2 for all t = 1 , 2 , , c ( 1 , 1 ) , K W > 0 . Hence, the standard definition [28] of an asymptotically stationary noise does not require K V 1 > 0 ; K V 1 = 0 means the RV has zero variance, and hence V 1 = v 1 takes a specific nonrandom value.

2.2. Closed-Form Feedback Capacity and Nonfeedback Capacity Lower Bounds for Stable/Unstable AR Noise

In Section 3, we show that the characterization of feedback capacity, C F B ( κ , P V 1 ) for asymptotically time-invariant AR ( c t ) , c t ( , ) , lim n c n = c ( , ) , lim n K W n = K W > 0 , is given by the optimization problem,
C F B ( κ , P V 1 ) = sup Λ ( , ) , K Z 0 1 2 log Λ + c 2 K + K Z + K W K W
such   that   ( s . t . ) Λ 2 K + K Z κ , Λ c , c ( , ) , K W > 0 ,
K = c 2 K + K W K W + c K Λ + c 2 K Z + K W + Λ + c 2 K , ( Algebraic   Riccati   eqn ) ,
K 0   is   a   stabilizing   solution   equivalent   to   | F C L ( K , Λ , K Z ) | < 1 ,
F C L ( K , Λ , K Z ) = c M ( K , Λ , K Z ) Λ + c , M ( K , Λ , K Z ) = K W + c K Λ + c K Z + K W + Λ + c 2 K ,
K = lim n K n , K n = E V n E V n | Y n 2 , K n , n = 1 , 2 , satisfies   recursion ( 39 ) .
According to the theory of Riccati equations (presented in Section 3), the constraint | F C L ( K , Λ , K Z ) | < 1 is necessary and sufficient for existence of capacity C F B ( κ , P V 1 ) as an asymptotic limit; this is equivalent to the conditions of detectability and unit circle controllability [28,29]. However, these conditions do not imply uniqueness of non-negative solutions of Riccati equations K 0 ; uniqueness requires the stronger conditions of detectability and stabilizability (instead of unit circle controllability).
We show that C F B ( κ , P V 1 ) ( 0 , ) is achieved by a channel input parametrized by the optimal values ( Λ , K Z , K ) = ( Λ , K Z , K ) of the above optimization problem, as follows.
X n = Λ n V n 1 E V n 1 | Y n 1 + Z n , X 1 = Z 1 , Z n N ( 0 , K Z n ) , n = 1 , 2 , , s . t . lim n ( Λ n , K Z n ) = ( Λ , K Z ) , K Z 0 , ( Λ , K Z , K ) the   optimal   values   of   C F B ( P V 1 , κ ) = C F B ( κ ) , P V 1 , K V 1 0 .
Thus, condition | F C L ( K , Λ , K Z ) | < 1 is necessary and sufficient for the asymptotic convergence of the estimation error E n = V n E V n | Y n , n = 1 , 2 , in a mean-square sense.
Remark 2. 
As pointed out in comment paper [1], the feedback capacity C F B ( κ , P V 1 ) in [2] [Theorem 6.1] for stable noise realization, claimed without a valid proof that K Z = K Z = 0 is optimal. Thus, for the stable AR noise, i.e., c ( 1 , 1 ) , Ref. [2] [Theorem 6.1] corresponds to C F B ( κ , P V 1 ) with K Z = 0 (but a proof is missing [1]). If K Z = 0 is assumed then C F B ( κ , P V 1 ) is very easy to compute. Our analysis using KKT conditions and convergence properties of Riccati equations illustrates that for c ( 1 , 1 ) , the optimal is K Z = K Z = 0 because there is no feasible solution with K Z > 0 of the optimization problem C F B ( κ , P V 1 ) . Moreover, there are different Regimes of capacity. Further, we also identify conditions for the validity of [2] [Theorem 6.1] for general noise in state space form.
Upon applying the KKT conditions, we show the solution of the optimization C F B ( κ , P V 1 ) consists of multiple Regimes, which depend on the values ( c , K W , κ ) , as follows.

2.2.1. Regime 1: Ergodic Feedback Capacity for Unstable Noise (Theorem 4)

Regime 1.  c 2 ( 1 , ) , κ > κ m i n = K W 1 + 4 c 2 3 2 c 2 1 2 .
The optimal values ( Λ , K Z , K ) = ( Λ , K Z , K ) of optimization problem C F B ( κ , P V 1 ) in (10) are
Λ = c K W κ c 2 1 2 K W ( , ) , K Z = κ c 2 1 κ c 2 1 2 K W K W 2 c 2 1 κ c 2 1 2 K W ( 0 , ) ,
K = κ c 2 1 2 K W c 2 c 2 1 ( 0 , ) , Λ , 2 + K Z = κ .
The feedback capacity C F B ( κ , P V 1 ) ( 0 , ) is an ergodic capacity given by
C F B ( κ , P V 1 ) = C F B ( κ ) = 1 2 log Λ + c 2 K + K Z + K W K W , P V 1 , K V 1 0
= 1 2 log c 2 c 2 1 κ + K W c 2 1 K W
= log | c | + 1 2 log 1 c 2 1 + κ K W log | c | ,
C F B ( κ ) log | c | + 1 2 log γ , f o r   l a r g e   γ = κ K W , 1 < c 2 < .
We also emphasize that for ergodic capacity C F B ( κ , P V 1 ) = C F B ( κ ) , P V 1 , i.e., it does not depend on the specific distribution of the RV V 1 (see Section 3.3).
Moreover, C F B ( κ ) ( 0 , ) is achieved by the asymptotically stationary and ergodic optimal input,
X n = Λ n V n 1 E V n 1 | Y n 1 + Z n , X 1 = Z 1 , Z n N ( 0 , K Z n ) , n = 1 , 2 , ,
s . t . lim n ( Λ n , K Z n , K n ) = ( Λ , K Z , K ) , K Z > 0 , K > 0 .
The optimal parameters of the input ( Λ , K Z ) are such that K Z > 0 , which imply there is one and only one limiting solution K = K = lim n K n 0 (where K n satisfies recursion (39)) of the algebraic Riccati equation (ARE) (12) which is positive and stabilizing, i.e., K > 0 and | F C L ( K , Λ , K Z ) | < 1 . Moreover, using the theory of Riccati equations (Theorem 2), we deduce that K Z > 0 is a necessary and sufficient condition for the feedback capacity to correspond to an ergodic capacity, i.e., lim n C n , F B ( κ , P V 1 ) = C F B ( κ , P V 1 ) = C F B ( κ ) , P V 1 , K V 1 0 , that is, the limit does not depend on P V 1 , i.e., the convergence is uniform P V 1 .
It is well-known (see, Gallager [4]) that for channels with memory there are fundamental differences between ergodic capacity in which the decoding error probability is independent of initial channel states (i.e., V 1 in our case and the value of K V 1 ) and nonergodic capacity in which the decoding error probability depends on initial channel states (i.e., the value of K V 1 in our case).
Comparison to Memoryless Noise. The condition for Regime 1 is alternatively expressed as | c | > 1 , γ > γ m i n 1 + 4 c 2 3 2 ( c 2 1 ) 2 where γ κ / K W is the signal-to-noise ration (SNR) of the memoryless noise (i.e., corresponding to c = 0 ), and the approximation holds γ m i n 1 | c | 3 for | c | 1 . Note that γ m i n approaches 0 as | c | increases. The feedback capacity in this regime is also expressed as
C F B ( κ ) = log | c | + 1 2 log 1 c 2 1 + γ > 1 2 log ( 1 + γ ) C m e m o r y l e s s
where C m e m o r y l e s s is the capacity of the memoryless channel, i.e., corresponding to c = 0 . Note that for strong memory | c | 1 ; by (20) the capacity can be approximated as
C F B ( κ ) 1 2 log 1 + c 2 γ f o r | c | 1
i.e., the combined effect of memory and feedback is to introduce gain of c 2 in the memoryless SNR, γ κ / K W , and this gain can be substantial for large c 2 . This shows significant benefits for strong memory noise combined with feedback communication.
By the above expressions the capacity C F B ( κ ) is increased by strong memory combined with feedback and this increase can also be characterized as follows:
Δ C ( κ ) C F B ( κ ) C m e m o r y l e s s ( κ ) = log | c | + 1 2 log 1 c 2 1 + γ 1 + γ > 0
Further note that for sufficiently large γ max { 1 , ( c 2 1 ) 1 } , this becomes
Δ C ( κ ) log | c |
i.e., Δ C ( κ ) increases with memory as measured by | c | > 1 , and Δ C ( κ ) can be large for large | c | . That is, we observe again that strong memory combined with feedback brings significant benefits.
Finally, we note that for unstable noise ( | c | > 1 ) with K W > 0 or/and K V 1 > 0 , the average noise power K ¯ V is (asymptotically) infinite, K ¯ V lim n 1 n i = 1 n K V n = . Yet, C F B ( κ ) > 0 , i.e., reliable communication with non-zero rate is possible. This should be contrasted with the memoryless channel, where infinite average noise power implies zero capacity, i.e. reliable communication with non-zero rate is impossible.

2.2.2. Regimes 2 and 3: Nonergodic Feedback Capacity for Stable/Unstable Noise-Complement of Regime 1 (Theorem 5)

For the complement of Regime 1, we deduce from the KKT conditions that C F B ( κ , P V 1 ) ( 0 , ) if and only if the variance K Z n , n = 1 , 2 , is asymptotically zero, i.e., lim n K Z n = K Z = 0 , giving rise to the feedback capacity of Regimes 2, 3 below.
Regime 2.  c 2 ( 1 , ) , κ κ m i n = K W 1 + 4 c 2 3 2 c 2 1 2 , and
Regime 3.  c [ 1 , 1 ] , κ [ 0 , ) .
The feedback capacity C F B ( κ , P V 1 ) ( 0 , ) and the corresponding optimal values ( Λ , K Z , K ) = ( Λ , K Z , K ) of the optimization problem in (10) are
C F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) = log max { 1 , | Λ | } , K V 1 > 0 ,
K Z = 0 , | Λ | > 1 , K = K + ( Λ ) = K W Λ 2 1 Λ + c 2 ( 0 , ) ,
where   K + ( Λ ) is   the   maximal   stab .   sol . ,   | F C L ( K , Λ , K Z = 0 ) = 1 Λ | < 1   for   | Λ | > 1 ,
Λ the   maximal   root   of   quartic   eqn , K W Λ 4 K W + κ Λ 2 2 c κ Λ c 2 κ = 0 .
We note that from the ARE (12), when K Z = 0 there is also a zero solution K ( Λ , ) = 0 such that | Λ | < 1 , but this is discarded because it implies C F B ( κ , P V 1 ) = 0 .
Moreover, (using properties of Riccati equations of Theorem 2), C F B m a x ( κ , P V 1 ) ( 0 , ) is achieved by the optimal input,
X n = Λ n V n 1 E V n 1 | Y n 1 , n = 2 , 3 , , X 1 = Z 1 , Z 1 N ( 0 , K Z 1 ) , K Z 1 > 0 ,
s . t . lim n ( Λ n , K Z n , K n ) = ( Λ , K Z , K ) , K Z = 0 , K = K + > 0 the   maximal   sol .   of ( 12 ) .

2.2.3. Discussion: Comparison of Regimes 1, 2, 3

Now, we make some observations. We clarify the gaps in the proof of the TD characterizations of feedback capacity in [2] [Theorem 6.1], which are identified in the comment paper [1] but not resolved. We also explain the capacity gains of Regime 1 compared to Regimes 2, 3 (Figure 2).
(1) For stable noise, i.e., c ( 1 , 1 ) , Ref. [2] [Theorem 6.1] does not give the capacity for asymptotically stationary noise and is not the ergodic capacity, because K V 1 = 0 should be excluded, unlike Regime 1. This is contrary to previous beliefs in [2,24], who interpreted [2] [Theorem 6.1] as the feedback capacity of asymptotically stationary noise.
This follows directly from expression C F B m a x ( κ , P V 1 ) = (25) of Regime 3 with c ( 1 , 1 ) , because this formula is correct provided the condition K V 1 > 0 holds. However, the condition K V 1 > 0 contradicts the standard definition of asymptotically stationary noise, i.e., Definition 1.(d). Moreover, since C F B m a x ( κ , P V 1 ) = (25) requires K V 1 > 0 , and it follows from our derivations that K V 1 = 0 implies the asymptotic limit is zero, i.e., lim n C n , F B ( κ , P V 1 ) = 0 , then it depends on the value of K V 1 (unlike Regime 1); hence, it is not an ergodic capacity. In fact, the proof of Theorem 5 shows that, when channel input (29) is used and K V 1 = 0 , then the limiting mean-square error converges to zero, i.e., K = K n = lim n E V n E V n | Y n 2 = 0 , ( K n satisfies recursion (39)) and necessarily implies | Λ | < 1 and C F B ( κ , P V 1 ) = 0 ; hence, we require K V 1 > 0 as stated in (25).
(2) For the stable state space realization of Gaussian noise V n , which is considered in [2,24], i.e., S t + 1 = A S t + G W t , V t = C S t + W t , t = 1 , , n , with stable A, the TD characterizations of feedback capacity in [2] [Theorem 6.1] should have explicitly stated that Conditions 1 and 2 of Yang, Kavcic and Tatikoda [20] [Section II.C, I–III] hold. Moreover, Ref. [2] [Theorem 6.1] did not provide a valid proof (see also comment paper [1]) to support the claim that asymptotically the optimal value is K Z = K Z = 0 . A necessary and sufficient condition for such K Z = K Z = 0 is that the matrix DRE converges to the maximal and stabilizing solution K = K + , corresponding to an asymptotic optimal value of K Z = 0 . This is precisely the reason that for Regimes 2 and 3, the optimal value is K Z = 0 . Moreover, K Z = 0 turned out to be optimal, because for Regimes 2 and 3, which include stable c ( 1 , 1 ) , there is no feasible unique and stabilizing solution of the corresponding matrix ARE, as in Regime 1 with K Z > 0 , that will ensure positive asymptotic value of capacity. Therefore, Ref. [2] [Theorem 6.1] should be corrected by including K Z 0 in all equations. Then we apply KKT conditions to determine whether the optimal value is K Z = K Z = 0 ([2] [Theorem 6.1] is restricted to A with stable eigenvalues).
(3) The fundamental differences between the feedback capacities of Regime 1 and Regimes 2 and 3, are the following.
For Regime 1, the limit C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = C F B ( κ ) ( 0 , ) is independent of P V 1 , K V 1 0 , i.e., the convergence is uniform P V 1 , K V 1 0 and the asymptotic variance of the innovations part of the channel input is strictly positive, i.e., K Z > 0 . The optimal value K Z > 0 implies there is a unique optimal value of the asymptotic mean-square error K > 0 . This is attributed to the optimal solution ( Λ , K Z ) , which implies that the conditions of detectability and stabilizability hold. Stabilizability implies K Z > 0 .
For Regimes 2 and 3, the limit C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) ( 0 , ) presupposes K V 1 > 0 and the asymptotic variance of the innovations part of the channel input is zero, i.e., K Z = 0 . This is attributed to the fact there is no feasible solution ( Λ , K Z ) to the optimization problem with positive capacity when K Z > 0 . To ensure positive capacity it is necessary that K Z = 0 ; this implies the condition of unit circle controllability condition holds but the condition of Stabilizability does not hold (this requires K Z > 0 ).
The optimal value K Z = 0 implies (as easily verified by the quadratic Equation (12)) there are two limiting optimal values of the asymptotic mean-square error K 0 , where K = K + > 0 is the maximal limiting value corresponding to C F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) ( 0 , ) , K V 1 > 0 , and K = 0 is the minimum non-negative limiting value corresponding to C F B ( κ , P V 1 ) = 0 , K V 1 0 .
The main point of the proof of Theorem 5 is that, if K V 1 = 0 , then channel input (29) and (30) implies C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = 0 .
(4) The capacity gains of Regime 1 compared to Regimes 2 and 3 are attributed to the following two factors.
(a) Unstable noise, i.e., | c | > 1 exhibits strong memory dependence, compared to stable noise, i.e., | c | < 1 . This follows from Remark 1; the value c is a measure of correlation between the samples V n and V n 1 , i.e., E V n V n 1 = c K V n 1 , K V = E ( V n 1 ) 2 .
As | c | becomes large, E V n V n 1 = c K V n 1 becomes large in absolute value. On the other hand, as | c | goes to zero, E V n V n 1 = c K V n 1 goes to zero, and the channel behaves as a memoryless channel Y n X n + W n . It is known channels with memory have higher capacity compared to memoryless channels.
(b) Optimal channel inputs should generate new information at each time instant; inputs that make use of an innovations process generate new information.
For Regime 1, the optimal input includes an innovations part, i.e., Z n , and it is not asymptotically zero, i.e., K Z > 0 , while for Regimes 2 and 3, the innovations part of the optimal channel input is asymptotically zero, i.e., K Z = 0 .
For Regime 1, we have | c | > 1 and K Z > 0 , and this explains the reason that the capacity of Regime 1 is much higher than the capacity of Regimes 2, 3, as | c | becomes large.
In fact, our simulations (see Figure 2) show that as | c | increases in the respective regions of Regimes 1, 2, 3, “capacity of Regime 1 ≥ capacity of Regime 2 ≥ capacity of Regime 3”.
In Section 2.3, we present numerical comparisons of C F B ( κ ) , C F B m a x ( κ , P V 1 ) .
In Section 2.5, we confirm the values of the closed-form expressions of C F B ( κ , P V 1 ) for Regimes 1–3, by solving the optimization problem (10) numerically, using a semi-definite programming (SDP) formulation as in [24].
In Section 3.2, we confirm the behavior of the asymptotic limits from the well-known properties of difference Riccati equations of filtering theory, for the convergence of the sequence K n 0 , n = 1 , 2 , to K 0 . It follows that the convergence of K Z n , n = 1 , 2 , to K Z > 0 is a necessary and sufficient condition for convergence of sequence K n 0 , n = 1 , 2 , for all initial conditions K 1 0 , K V 1 0 to a unique limit K 0 , and for lim n C n , F B ( κ , P V 1 ) = C F B ( κ , P V 1 ) = C F B ( κ ) , P V 1 , K V 1 0 , i.e., the limit is independent of P V 1 . This is valid only for feasible Region 1.

2.2.4. Lower Bounds: Lower Bounds on Ergodic Nonfeedback Capacity for Stable/Unstable Noise

In Theorem 7.(2), for Regimes 1–3, i.e., c ( , ) , κ [ 0 , ) , we derive the non-trivial lower bound on nonfeedback capacity corresponding to an independent input, X 1 , X 2 , given by
C F B ( κ , P V 1 ) C ( κ , P V 1 ) C L B ( κ ) , C L B ( κ ) = 1 2 log κ 1 + c 2 + K W + κ 1 c 2 + K W 2 + 4 c 2 K W κ 2 K W , P V 1 , K V 1 0 ,
and achieved by (23) with
lim n Λ n = 0 , i . e . , X n = Z n N ( 0 , K Z n ) , lim n K Z n = κ
where C L B ( κ ) is independent of the distribution of V 1 , i.e., it holds K V 1 0 , and hence it is an ergodic achievable rate.
Numerical simulations show that the lower bound on nonfeedback capacity C L B ( κ ) for c ( 1 , 1 ) is higher than the feedback capacity C F B m a x ( κ , P V 1 ) for c ( 1 , 1 ) . This is consistent with our discussion in Section 2.2.3 under (4). For unstable c ( 1 , 1 ) , we are not aware of any literature that computed nonfeedback capacity or bounds.

2.3. Numerical Comparisons of C F B ( κ ) , C F B m a x ( κ , P V 1 ) and Lower Bound C L B ( κ ) for Stable/Unstable Noise

Figure 2 and Figure 3 illustrate the performance of Regime 1, C F B ( κ , P V 1 ) = C F B ( κ ) , K V 1 0 , Regimes 2, 3, C F B ( κ , P V 1 ) = C F B m a x ( P V 1 , κ ) = log max { 1 , | Λ | } , K V 1 > 0 , and lower bound on nonfeedback capacity for Regimes 1–3, C ( κ , P V 1 ) C L B ( κ ) , K V 1 0 , for values c = 0.5 , c = 1.5 , c = 2 and K W = 1 , as a function of transmit power κ .
Some of the conclusions are the following.
(1) Regime 1, C F B ( κ ) , K V 1 0 is always higher than Regimes 2, 3,
C F B m a x ( P V 1 , κ ) = log max { 1 , | Λ | } , K V 1 > 0 , and the lower bound C L B ( κ ) . This surprising feature is fundamentally different than the behavior of the Shannon capacity C S h a n n o n that is limited by the signal-to-noise ratio.
(2) Regime 1, C F B ( κ ) is an increasing function of the parameter | c | ( 1 , ) , that is, the more unstable the AR ( c ) noise, the higher the value of C F B ( κ ) , while κ m a x is reduced, i.e., as | c | ( 1 , ) increases, higher rates are achieved at reduced transmit power κ .
(3) The lower bound C L B ( κ ) approaches Regime 1, C F B ( κ ) , for larger values of | c | ( 1 , ) . This is also a surprising result of unstable, not encountered for stable noise.
(4) Regimes 2 and 3, C F B m a x ( κ , P V 1 ) , cannot be optimal for Regime 1 because the broken lines of Figure 3 are below Regime 1, C F B ( κ ) .
Figure 4 illustrates the tightness of the lower bound on nonfeedback capacity, C L B L B ( κ ) given by (31), and incurred by an IID channel input Z t N ( 0 , κ ) , K Z = κ , to the optimal water-filling solution of nonfeedback capacity (recall that the water-filling characterization is based on the use of time-invariant or asymptotically time-invariant channel input strategies); Equation (5.514) in Ihara [5] (see also Equation (6) in Butman [10]), given by
C ( κ ) = 1 2 log 1 + κ + c 2 1 c 2 , κ > 1 ( 1 | c | ) 2 1 ( 1 c 2 ) , c ( 1 , 1 ) , K W = 1 .
Contrary to the nonfeedback lower bound Formula (31), which holds for all stable and unstable AR ( c ) , c ( , ) noise and κ [ 0 , ) , the closed form nonfeedback Formula (33) based on water-filling, is restricted to κ > 1 ( 1 | c | ) 2 1 ( 1 c 2 ) , and to the stable AR ( c ) , c ( 1 , 1 ) . The maximum difference C ( κ ) C L B ( κ ) , when K W = 1 occurs at c = 0.75 , and is less than 1.5 × 10 2 bits per channel use.

2.4. Time-Sharing Increases Achievable Rates

Figure 5 considers the values c = 1.5 , K W = 1 to illustrate the rate of a time-sharing scheme between Regime 1, C F B ( κ ) , κ > κ m i n = K W 1 + 4 c 2 3 2 c 2 1 2 , and the lower bound C L B ( κ ) , κ [ 0 , ) on the asymptotic characterization of nonfeedback capacity. This scheme results in higher rates, because it employs a time-varying channel input strategy, that is, two different strategies—one without feedback and one with feedback are applied. The fundamental conclusion is that, for unstable noise, time-varying channel input strategies incur higher rates compared to asymptotically time-invariant channel input strategies.

2.5. Comparison of Closed-Form Formulas C F B ( κ ) and C F B m a x ( κ , P V 1 ) with Numerical Solutions Produced by the Semi-Definite Program of [24]

Table 1 and Table 2 present comparisons of the closed-form expressions of feedback capacity C F B ( κ ) = (20) for Regime 1, C F B m a x ( κ , P V 1 ) = (25) for Regimes 2 and 3, and lower bound C L B ( κ ) = (31) for Regimes 1–3, to the numerical solutions produced by the semi-definite program (SDP) formulation [24] of the optimization problem C F B ( κ , P V 1 ) of (10).
Regime 1: c 2 ( 1 , ) , κ > κ m i n . The Table 1 second block illustrates that the numerical solution of C F B ( κ , P V 1 ) produced by SDP coincides with the closed-form solution of C F B ( κ ) of Regime 1 given by (17)–(20), while C F B m a x ( κ , P V 1 ) of Regimes 2 and 3 is sub-optimal. The Table 2 second block illustrates that the optimal parameters ( Λ , K Z ) produced by SDP of the optimization problem C F B ( κ , P V 1 ) coincide with the optimal parameters of the closed-form solution of C F B ( κ ) of Regime 1.
Regime 2, c 2 ( 1 , ) , κ κ m i n . The Table 1 first block illustrates that the numerical solution of C F B ( κ , P V 1 ) produced by SDP coincides with the closed-form solution of C F B m a x ( κ , P V 1 ) of Regime 2 given by (25), which further implies that C F B ( κ ) of Regimes 1 cannot be optimal (actually there is no feasible solution under the constraint K Z > 0 ). The Table 2 first block illustrates that the optimal parameters ( Λ , K Z ) produced by SDP of the optimization problem C F B ( κ , P V 1 ) coincide with the optimal parameters of the closed-form solution of C F B m a x ( κ , P V 1 ) of Regime 2.
Regime 3, c [ 1 , 1 ] , κ [ 0 , ) . The numerical solution of C F B ( κ , P V 1 ) produced by SDP coincides with the closed-form solution of C F B m a x ( κ , P V 1 ) of Regime 3 given by (25), hence we omit the details.
The above numerical solution is consistent with Theorem 4, where we show that the optimization problem (10)–(12) for Regimes 2 and 3 does not have a nontrivial feasible solution that corresponds to an optimal value K Z > 0 .

3. Asymptotic Characterizations of Capacity for Stable and Unstable AR Noise

In this Section, we present the mathematical tools and derivations of the expressions presented in Section 2. Our derivations are generic and applicable to other more general noise models. In fact, it will become apparent that the analysis of the AR unit memory noise model is rich enough and brings out all technical issues that one will be encounter, if more general noise models are considered.

3.1. Sequential Characterizations of Feedback Capacity and Lower Bounds on Achievable Nonfeedback Capacity for Gaussian Channels Driven by Time-Varying AR Noise

Theorem 1 is the starting point prior to the analysis of the asymptotic feedback capacity, lim n C n , F B ( κ , P V 1 ) . It was first presented in [27] (and earlier in [26]).
Theorem 1. 
[26,32] Sequential characterization of C n , F B ( κ , P V 1 ) for Gaussian channels driven by AR ( c t ) Noise
Consider the feedback code of [3] and the AR ( c t ) noise of Definition 1. Define the conditional means and covariances,
X ^ t = E X t | Y t , V ^ t = E V t | Y t , t = 1 , , n ,
K t = E X t X ^ t 2 | Y t = E V t V ^ t 2 | Y t , t = 1 , , n .
(1) The characterization of the n - FTFI capacity C n , F B ( κ , P V 1 ) is given by
C n , F B ( κ , P V 1 ) = sup P X t | X t 1 , Y t 1 , t = 1 , , n , 1 n E t = 1 n X t 2 κ 1 n H ( Y n ) H ( V n ) [ 0 , ]
= sup P n , F B ( κ ) 1 2 n log K Z 1 + K V 1 K V 1 + 1 2 n t = 2 n log Λ t + c t 2 K t 1 + K Z t + K W t K W t P n , F B ( κ ) = { Λ t , K Z t ( , ) × [ 0 , ) , t = 1 , , n | K t 0 , t = 1 , , n s a t i s f i e s ( 39 ) ,
1 n K Z 1 + 1 n t = 2 n { Λ t 2 K t 1 + K Z t } κ } ,
K t = c t 2 K t 1 + K W t K W t + c t K t 1 Λ t + c t 2 K Z t + K W t + Λ t + c t 2 K t 1 , K t 0 , K 1 = K V 1 K Z 1 K V 1 + K Z 1 0
where H ( X ) denotes the entropy (differential) of RV X.
(2) C n ( κ , P V 1 ) in part (1) is achieved by the time-varying jointly Gaussian channel input process X n , with a representation
X t = Λ ¯ t X t 1 X ^ t 1 + Z t , X 1 = Z 1 , t = 2 , , n ,
= Λ t V t 1 V ^ t 1 + Z t , Λ ¯ t = Λ t ,
Z t N ( 0 , K Z t ) , K Z t 0 , t = 1 , , n i n d e p .   G a u s s i a n   s e q u e n c e ,
Z t i n d e p e n d e n t   o f ( V t 1 , X t 1 , Y t 1 ) , t = 1 , , n , Z n i n d e p e n d e n t   o f V n ,
V t = c t V t 1 + W t , , c t ( , ) , t = 2 , , n ,
Y t = X t + V t = Λ ¯ t X t 1 X ^ t 1 + c t Y t 1 X t 1 + W t + Z t , t = 2 , , n
= Λ t V t 1 V ^ t 1 + c t V t 1 + W t + Z t ,
Y 1 = Z 1 + V 1 ,
1 n E t = 1 n X t 2 = 1 n K Z 1 + 1 n t = 2 n Λ t 2 K t 1 + K Z t κ ,
( Λ t , K Z t ) ( , ) × [ 0 , ) s c a l a r v a l u e d ,   n o n r a n d o m ,
V ^ 1 = E V 1 | Y 1 = K V 1 K V 1 + K Z 1 Y 1 , K V 1 0 .
and its distribution satisfies, P X t | X t 1 , Y t 1 = P X t | V t 1 , Y t 1 , t = 1 , , n .
Further, ( V ^ t , K t ) , t = 1 , , n are determined by the generalized (unlike [20], we use the term generalized because the conditions for the asymptotic analysis to hold are fundamentally different from those of asymptotic analysis of classical Kalman-filter equations) time-varying Kalman-filter and generalized time-varying difference Riccati equation (DRE) of estimating V n from Y n , given below.
Generalized Kalman-filter Recursion for (44)–(47) (see application of Kalman filter in [28,29]):
V ^ t = c t V ^ t 1 + M t ( K t 1 , Λ t , K Z t ) I t , V ^ 1 = ( 50 ) ,
= F t C L ( K t 1 , Λ t , K Z t ) V ^ t 1 + M t ( K t 1 , Λ t , K Z t ) Y t ,
I t = Y t E Y t | Y t 1 = Y t c t V ^ t 1 , t = 2 , , n , I 1 = Z 1 + V 1 ,
= Λ t + c t V t 1 V ^ t 1 + Z t + W t ,
M t ( K t 1 , Λ t , K Z t ) = K W t + c t K t 1 Λ t + c t K Z t + K W t + Λ t + c t 2 K t 1 1 ,
F t C L ( K t 1 , Λ t , K Z t ) = c t M t ( K t 1 , Λ t , K Z t ) Λ t + c t ,
I t , t = 1 , , n , a n   o r t h o g o n a l   i n n o v a t i o n s   p r o c e s s .
Error Recursion of the Generalized Kalman filter, E t = V t V ^ t , t = 1 , , n :
E t = F t C L ( K t 1 , Λ t , K Z t ) E t 1 M t ( K t 1 , Λ t , K Z t ) Z t + W t + W t , E 1 = V 1 V ^ 1 .
Generalized Time-Varying Difference Riccati Equation: K t = E E t 2 0 , t = 1 , , n satisfies (39).
Theorem 1 part (2) is presented because it is this part that gives rise to part (1) and provides fundamental insight.
Remark 3. 
The special case of Theorem 1 with ( c t , K W t = ( c , K W ) , c ( 1 , 1 ) , t , and initial state V 1 = v 1 fixed and known to the encoder and the decoder can be obtained from [2,24], while bounds on C n , F B ( κ , P V 1 ) are given in [8,9,10]. However, there is a subtle difference; if V 1 = v 1 fixed and known to the encoder, then C n , F B ( κ , P V 1 ) = C n , F B ( κ , v 1 ) = , i.e., depends on the initial state V 1 = v 1 not its distribution P V 1 , because the following adjustments are needed in Theorem 1, K 1 = 0 , V ^ 1 = v 1 (because V 1 = v 1 is fixed). This corresponds to Theorem 1, by replacing K V 1 0 by K V 1 = 0 . The subtle issue of dependence of C n , F B ( κ , P V 1 ) on P V 1 will emerge often in the subsequent analysis of asymptotic limit C F B ( κ , P V 1 ) .
Now, we discuss a subtle issue of K V 1 = 0 . By (37), if K V 1 = 0 then 1 2 log K Z 1 + K V 1 K V 1 | K V 1 = 0 = + . To exclude + values of mutual information, we use Gelfand and Yaglom [34] (see Equation (2.8’) and Chapter II), as stated in Remark 4.
Remark 4. 
By Theorem 1 if K V 1 = 0 and K Z 1 > 0 , since X 1 = Z 1 and Y 1 = Z 1 , we have at time t = 1 a value of mutual information I ( X 1 ; Y 1 ) = I ( Z 1 ; Z 1 ) = + , which is well-known and allowed, for the mutual information between two jointly Gaussian RVs; a detailed discussion on this point is given in Gelfand and Yaglom [34] (see Equation (2.8’) and Chapter II). The value I ( X 1 ; Y 1 ) = + is precisely the value of 1 2 log K Z 1 + K V 1 K V 1 | K V 1 = 0 = + , i.e., the first right hand side term in (37). As discussed in Gelfand and Yaglom [34] (see Equation (2.8’) and Chapter II), to avoid values of + , it is a standard practice to remove such RVs from the evaluation of rates. A transmission policy which avoids values of I ( X 1 ; Y 1 ) = + is the following.
Step # 1. Initialize K n , n = 2 , 3 , at K 1 = 0 , i.e., corresponding to K V 1 = 0 .
Step # 2. Omit transmission at time t = 1 , hence, the term 1 2 log K Z 1 + K V 1 K V 1 | K V 1 = 0 = + in (37) is excluded.
Step # 3. Apply Step #2 and evaluate 1 n C n ( κ , P V 1 ) and its limit, with K V 1 = 0 , i.e., K 1 = 0 .
By Steps #1–#3, we can treat the case when the noise initial variance is K V 1 = 0 , and hence we can also consider asymptotically stationary noise, since such noise is defied for all K V 1 0 .
We should emphasize the following subtle issue: Steps #1–#3 do not apply to the channel input of Regimes 2, 3, (29) and (30), because by the proof of Theorem 5, if K V 1 = 0 then K 1 = 0 , and C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = 0 .
Next, we give a lower bound on the characterization of n FTwFI Capacity for nonstationary and nonergodic noise, i.e., for the nonfeedback code [3], which is based on a sequential characterization similar to Theorem 1. In the paper, we evaluate this lower bound against the feedback capacity C F B ( κ , P V 1 ) .
Corollary 1. 
(Lower bound on sequential characterization of C n ( κ , P V 1 ) for Gaussian channels driven by AR Noise)
Consider the code without feedback in [3], and the AR ( c t ) noise of Definition 1. Define
K X t = E X t 2 , X ^ t = E X t | Y t , V ^ t = E V t | Y t , E t n f b = V t V ^ t ,
K t = E X t X ^ t 2 | Y t = E V t V ^ t 2 | Y t , t = 1 , , n .
(1) A lower bound on the characterization of the n FTwFI capacity C n ( κ , P V 1 ) is given by
C n ( κ , P V 1 ) = sup P X t | X t 1 , t = 1 , , n , 1 n E t = 1 n X t 2 κ 1 n H ( Y n ) H ( V n ) [ 0 , ] C n L B ( κ , P V 1 ) = sup P n L B ( κ ) { 1 2 n log K Z 1 + K V 1 K V 1
+ 1 2 n t = 2 n log Λ ¯ t c t 2 K t 1 + K Z t + K W t K W t } , P n L B ( κ ) = { Λ ¯ t , K Z t ( , ) × [ 0 , ) , t = 1 , , n | K t 0 , K X t 0 , t = 1 , , n
s a t i s f y ( 64 ) a n d ( 65 ) , 1 n K Z 1 + 1 n t = 2 n Λ ¯ t 2 K X t 1 + K Z t κ } ,
K t = Λ ¯ t 2 K t 1 + K Z t K Z t + Λ ¯ t K t 1 Λ ¯ t c t 2 K Z t + K W t + Λ ¯ t c t 2 K t 1 , K 1 = K V 1 K Z 1 K V 1 + K Z 1 0 , K t 0 ,
K X t = Λ ¯ t 2 K X t 1 + K Z t , K X t 0 , K X 1 = K Z 1 , t = 2 , , n .
(2) C n L B ( κ , P V 1 ) of part (1) is achieved by the time-varying jointly Gaussian channel input process X n , with a representation
X t = Λ ¯ t X t 1 + Z t , X 1 = Z 1 , t = 2 , , n ,
Z t N ( 0 , K Z t ) , K Z t 0 , t = 1 , , n i n d e p .   G a u s s i a n   s e q u e n c e ,
Z t i n d e p e n d e n t   o f ( V t 1 , X t 1 , Y t 1 ) , t = 1 , , n , Z n i n d e p e n d e n t   o f V n ,
V t = c t V t 1 + W t , , c t ( , ) , t = 2 , , n ,
Y t = X t + V t = Λ ¯ t c t X t 1 + c t Y t 1 + W t + Z t , t = 2 , , n ,
Y 1 = Z 1 + V 1 ,
1 n E t = 1 n X t 2 = 1 n K Z 1 + 1 n t = 2 n Λ ¯ t 2 K X t 1 + K Z t κ ,
( Λ ¯ t , K Z t ) ( , ) × [ 0 , ) s c a l a r v a l u e d , n o n r a n d o m .
and its distribution satisfies P X t | X t 1 = P X t | X t 1 , t = 1 , , n .
Further, ( X ^ t , K t ) , t = 1 , , n are determined by the generalized time-varying Kalman-filter and generalized time-varying difference Riccati equation (DRE) of estimating X n from Y n , and K X t , t = 1 , , n is determined by the time-varying Lyapunov difference equation, given below. Generalized Kalman-filter Recursion for (66)–(71):
X ^ t = Λ ¯ t X ^ t 1 + M t n f b ( K t 1 , Λ ¯ t , K Z t ) I t , X ^ 1 = ( 50 ) , t = 2 , , n
= F t n f b ( K t 1 , Λ ¯ t , K Z t ) X ^ t 1 + M t n f b ( K t 1 , Λ ¯ t , K Z t ) Y t c t Y t 1 ,
I t = Y t Λ ¯ t c t X ^ t 1 c t Y t 1 , I 1 = Z 1 + W 1 , t = 2 , , n ,
= Λ ¯ t c t X t 1 X ^ t 1 + Z t + W t ,
M t n f b ( K t 1 , Λ ¯ t , K Z t ) = K Z t + Λ ¯ t K t 1 Λ ¯ t c t K Z t + K W t + Λ ¯ t c 2 K t 1 1 ,
F t C L , n f b ( K t 1 , Λ ¯ t , K Z t ) = Λ ¯ t M t n f b ( K t 1 , Λ ¯ t , K Z t ) Λ ¯ t c t
I t , t = 1 , , n , a n   o r t h o g o n a l   i n n o v a t i o n s   p r o c e s s .
Error Recursion of the Generalized Kalman-filter, E t n f b = X t X ^ t , t = 1 , , n :
E t n f b = F t C L , n f b ( K t 1 , Λ ¯ t , K Z t ) E t 1 M t n f b ( K t 1 , Λ ¯ t , K Z t ) Z t + W t + Z t , E 1 n f b = V 1 V ^ 1 .
Generalized Time-Varying Difference Riccati Equation: K t = E E t n f b 2 0 , t = 1 , , n satisfies (64).
Time-Varying Difference Lyapunov Equation: K X t 0 , t = 1 , , n satisfies (65).
Proof. 
From Cover and Pombra [3], we know that a jointly Gaussian X n achieved the supremum in C n ( κ , P V 1 ) , i.e., its distribution is P X t | X t 1 , t = 1 , , n . Hence, a Markovian input, i.e., P X t | X t 1 , t = 1 , , n gives a lower bound. A realization of such Markovian input is X t = Λ ¯ t X t 1 + Z t , X 1 = Z 1 , t = 2 , , n . This implies the recursion (65). The rest of the statements are easily obtained by calculating the n FTwFI capacity of Cover and Pombra [3] for this specific input, using the generalized Kalman-filter recursions following [28,29]. Some of the details are presented in [33], hence we omit them. □
Remark 5. 
The solution to the optimization problem C n , F B ( κ , P V 1 ) of (37) for arbitrary finite n remains to this date a challenging open problem because the optimization needs to be carried out over the time-varying parameters ( Λ n , K Z n ) , K Z n 0 , n = 1 , 2 . For the rest of the paper we consider the asymptotic feedback capacity problem lim n C n , F B ( κ , P V 1 ) for asymptotically time-invariant stable and unstable AR ( c ) noise.
In Section 3.2 (to be introduced shortly) we present necessary and sufficient conditions to address the limit C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) and the convergence properties of the error recursion (58).
Although recursion (58) is linear, lim n K n is not expected to exist, for arbitrary ( F t C L ( K t 1 , Λ t , K Z t ) , M t ( K t 1 , Λ t , K Z t ) ) , t = 2 , . Previous papers [2,20,24] did not address this point. Indeed, by Section 3.2, the convergence properties of the sequence K 2 , , K n with K 1 0 , as n , are characterized by the detectability and stabilizability or unit circle controllability conditions [28,29]. The detectability and stabilizability (resp., unit circle controllability) conditions ensure existence of the limit, lim n K n = K , such that K 0 is the unique solution (resp., maximal solution) of a generalized ARE and satisfies the stability property, lim n F n C L ( K n 1 , Λ n , K Z n ) = F C L ( K , Λ , K Z ) ( 1 , 1 ) , for all K V 1 0 (resp., for some K V 1 > 0 ). We show that the solution of the asymptotic feedback capacity optimization problem is such that detectability and stabilizability conditions hold, but this is only for Regime 1. For the complement of Regime 1, we show there exists an optimal solution, provided the stabilizability condition is replaced by the more relaxed unit circle controllability condition, and this gives rise to the feedback capacity of Regimes 2, 3. In the sequel, these detectability and stabilizability or unit circle controllability conditions are encountered in the asymptotic feedback capacity optimization problem.
Remark 6. 
In Section 3.2 we introduce the properties of generalized DREs and AREs for time-invariant strategies ( Λ n , K Z n ) = ( Λ , K Z ) n ; we note that these also hold for asymptotically time-invariant strategies, lim n ( Λ n , K Z n ) = ( Λ , K Z ) , due to the continuity properties of DREs with respect to their coefficients.

3.2. Convergence Properties of Generalized DREs

First, we emphasize that all materials of this section have natural generalizations to the finite-dimensional state space realizations of the noise V n , which are considered in [2,20,24] (these are found in [27] and earlier arXiv paper [26]). Unfortunately, [2,20,24] did not relate their asymptotic feedback capacity, C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) , to the material of this section, and this omission leads to various confusions, including the statements found in the comment paper [1] regarding the TD feedback capacity given in [2] [Theorem 6.1]. The complexity will become apparent below.
Consider the generalized DREs and AREs for time-invariant parameters and strategies ( c n , K W n ) = ( c , K W ) , ( Λ n , K Z n ) = ( Λ , K Z ) n . Let { K t , t = 1 , 2 , , n } denote a sequence that satisfies the time-invariant generalized DRE with arbitrary initial condition
K t = c 2 K t 1 + K W K W + c K t 1 Λ + c 2 K Z + K W + Λ + c 2 K t 1 , K 1 = g i v e n , t = 2 , , n .
The generalized algebraic Riccati equation (ARE) corresponding to (82) is
K = c 2 K + K W K W + c K Λ + c 2 K Z + K W + Λ + c 2 K , K 0 .
We note that a solution of (82) is a functional of the parameters of the right hand side, that is, K t K t ( c , K W , Λ , K Z , K 1 ) , t = 2 , , n . To discuss the properties of the generalized DRE (82), we introduce, as is often performed in the analysis of generalized DREs [28,29] [Section 14.7, page 540], the following standard definitions.
A = c , C = Λ + c , A = c K W R 1 C , B , 1 2 = K W 1 2 B 1 2
R = K Z + K W > 0 , B = 1 K W K Z + K W 1 ,
K W > 0 , c ( , ) , Λ ( , ) , K Z 0 .
By (55) and (57), we also define
M ( K , Λ , K Z ) = K W + A K C R + C 2 K 1 , F C L ( K , Λ , K Z ) = A M ( K , Λ , K Z ) C .
Next, we introduce the definition of asymptotic stability of the error recursion (58) for ( c n , K W n , Λ n , K Z n ) = ( c , K W , Λ , K Z ) n .
Definition 2. 
Asymptotic stability [28,29]
A solution K 0 to the generalized ARE (83), assuming it exists, is called
(a) stabilizing if | F C L ( K , Λ , K Z ) | < 1 ,
(b) maximal and stabilizing solution, denoted by K + , if K + is stabilizing, and in addition K + K 0 , where K is any other solution.
For such solutions, K or K = K + , we say F C L ( K , Λ , K Z ) is asymptotically stable, i.e., | F C L ( K , Λ , K Z ) | < 1 ,
Next, we define the important notions of detectability, unit circle controllability, and stabilizability, which are directly related to asymptotic stability of the error recursion (58), in a mean-square sense.
Definition 3. 
Detectability, Stabilizability, Unit Circle controllability [28,29]
(a) The pair A , C is called detectable if there exists a G R such that | A G C | < 1 (stable).
(b) The pair A , B , 1 2 is called unit circle controllable if there exists a G R such that | A B , 1 2 G | 1 .
(c) The pair A , B , 1 2 is called stabilizable if there exists a G R such that | A B , 1 2 G | < 1 .
The next lemma characterizes detectability, unit circle controllability, and stabilizability [29,35].
Lemma 1. 
[29,35] Necessary and sufficient conditions for detectability, unit circle controllability, and stabilizability
(a) The pair A , C is detectable if and only if there exists no eigenvalue, eigenvector { λ , x } , of A, i.e., A x = λ x such that | λ | 1 , and such that C x = 0
(b) The pair A , B , 1 2 is unit circle controllable if and only if there exists no eigenvalue, eigenvector { λ , x } , x A = x λ , such that | λ | = 1 , and such that that x B , 1 2 = 0 .
(c) The pair A , B , 1 2 is stabilizable if and only if there exists no eigenvalue, eigenvector { λ , x } , x A = x λ such that | λ | 1 , and such that x B , 1 2 = 0 .
(d) If C 0 then the pair A , C is detectable. If B , 1 2 0 then the pair A , B , 1 2 is stabilizable.
In the next theorem we summarize known results on sufficient and necessary conditions for convergence of solutions { K t , t = 1 , 2 , , n } of the generalized time-invariant DRE, as n , to a non-negative K, which is stabilizing (not necessarily unique, i.e., it may be a maximal solution K = K + ), and unique and stabilizing. First, we recall that the pair A , C is detectable—is a necessary condition but not sufficient condition for convergence of the sequence { K t , t = 1 , 2 , , n } generated by the generalized DRE, as n , to a non-negative K 0 which is a stabilizing solution of a corresponding generalized ARE. The sufficient condition is that the pair A , B , 1 2 is unit circle controllable. However, under the conditions where the pair A , C is detectable and the pair A , B , 1 2 is unit circle controllable, the limiting K is not necessarily unique, and there may be multiple stabilizing solutions (i.e, such as K + , where K + K 0 , and K 0 is any other solution) depending on the initial condition K 1 0 . Uniqueness of K 0 is ensured by replacing the pair A , B , 1 2 that is unit circle controllable by the pair A , B , 1 2 is stabilizable.
Theorem 2. 
[28,29] Convergence of time-invariant generalized DRE to stabilizing solutions
Let { K t , t = 2 , , n } denote a sequence that satisfies the time-invariant generalized DRE (82) with arbitrary initial condition, K 1 and ( A , C , A , B , 1 2 ) defined by (84) and (85), with R = K Z + K W > 0 . Then the following hold.
(1) Consider the generalized RDE (82) with zero initial condition, i.e., K 1 = 0 , and assume the pair A , C is detectable and the pair A , B , 1 2 is unit circle controllable.
Then the sequence { K t : t = 2 , , n } that satisfies the generalized DRE (82), with zero initial condition K 1 = 0 , converges to K, i.e., lim n K n = K , where K satisfies the ARE (83) if and only if the pair A , B , 1 2 is stabilizable.
(2) Assume the pair A , C is detectable and the pair A , B , 1 2 is unit circle controllable. Then there exists a unique stabilizing solution K 0 to the generalized ARE (83), i.e., such that | F C L ( K , Λ , K Z ) | < 1 , if and only if { A , B , 1 2 } is stabilizable.
(3) If { A , C } is detectable and { A , B , 1 2 } is stabilizable, then any solution K t , t = 2 , , n to the generalized DRE (82) with arbitrary initial condition, K 1 0 is such that lim n K n = K , where K 0 is the unique solution of the generalized ARE (83) with | F C L ( K , Λ , K Z ) | < 1 , i.e., it is stabilizing.
(4) { A , C } is detectable and { A , B , 1 2 } unit circle controllable are necessary and sufficient conditions for any solution K t , t = 2 , , n to the generalized DRE (82) to converge, lim n K n = K , from some initial condition, K 1 0 where K 0 is a stabilizing solution of the generalized ARE (83), but it may not be unique, i.e., Ref. (83) may have multiple solutions K 0 , where one of the solutions is the maximal and stabilizing solution, denoted by K + 0 , such that K + K 0 .
Theorem 2.(1) follows by combining [29] [Lemma 14.2.1, page 507] of classical DREs and AREs with [29] [Section 14.7] of generalized DREs and AREs. Theorem 2.(2) is given in [29] [Theorem E.6.1, page 784]. Theorem 2.(3) is obtained from [28] [Theorem 4.2, page 164], and also [29], etc.
From Theorem 2, follows the next lemma.
Lemma 2. 
Properties of Solutions of time-invariant DREs and AREs for different cases
Let ( A , C , A , B , 1 2 ) be defined by (84) and (85), with R = K Z + K W > 0 , i.e., K W > 0 .
(1.a) Suppose c ( 1 , 1 ) . Then the pair { A , C } is detectable.
(1.b) Suppose c ( 1 , 1 ) . Then the pair { A , C } is detectable if C = Λ + c 0 .
(1.c) Suppose c ( , ) . Then the pair { A , B , 1 2 } is stabilizable if K Z > 0 .
(2) Suppose K Z = 0 . Then the pair { A , B , 1 2 } is unit circle controllable if and only if | Λ | 1 .
(3) Suppose K Z = 0 . Then the pair { A , B , 1 2 } is stabilizable if and only if | Λ | < 1 .
(4) Suppose c ( 1 , 1 ) , K Z = 0 . The sequence { K t , t = 1 , 2 , , n } that satisfies the generalized DRE with zero initial condition K 1 = 0 (reps. non-zero initial conditions K 1 > 0 )
K t = c 2 K t 1 + K W K W + c K t 1 Λ + c 2 K W + Λ + c 2 K t 1 , K 1 = 0 ( r e s p . K 1 > 0 ) , t = 2 , , n
converges to K 0 , i.e., lim n K n = K , where K satisfies the generalized ARE (83) with K Z = 0 if and only if the { A , B , 1 2 } is stabilizable, equivalently, | Λ | < 1 (resp., if and only if the { A , B , 1 2 } is unit circle controllable, equivalently, | Λ | 1 ).
(5) Suppose K Z = 0 , and | Λ | 1 , with the corresponding generalized ARE (83),
K = c 2 K + K W K W + c K Λ + c 2 K W + Λ + c 2 K .
Then
F C L ( K , Λ , K Z = 0 ) = c M ( K , Λ , K Z = 0 ) Λ + c ,
M ( K , Λ , K Z ) = K W + c K Λ + c K W + Λ + c 2 K 1 .
Moreover, the solutions K 0 of the ARE (89) are
K ( Λ ) = 0 , u n i q u e , s t a b i l i z i n g , | F C L ( K , Λ , K Z = 0 ) = Λ | < 1 , i f f | Λ | < 1 , K + ( Λ ) = K W Λ 2 1 Λ + c 2 ( 0 , ) , m a x i m a l , s t a b i l . , | F C L ( K , Λ , K Z = 0 ) = 1 Λ | < 1 , i f f | Λ | > 1 , c Λ .
Proof. 
See Appendix A.1. □
Proposition 1. 
Properties of solutions of asymptotically time-invariant DREs and AREs
Consider the generalized DRE (82) coefficients of ( c , K W , Λ , K Z ) replaced by time-varying ( c t , K W t , Λ t , K Z t ) , t = 2 , 3 , , n ,
K t = c t 2 K t 1 + K W t K W t + c t K t 1 Λ t + c t 2 K Z t + K W t + Λ t + c t 2 K t 1 , K 1 = g i v e n , t = 2 , , n ,
s u c h   t h a t lim n ( c n , K W n , Λ n , K Z n ) = ( c , K W , Λ , K Z ) , K W > 0 , K Z 0 .
Then Theorem 2 and Lemma 2 remain valid.
Proof. 
This is due to the well-known continuity properties of DREs with respect to their coefficients, i.e., the convergence properties are characterized by the limiting pairs, lim n A n , C n = A , C and lim n A n , B n , 1 2 = A , B , 1 2 , where A n = c n ,   C n = Λ n + c n ,   A n = c n K W n R n 1 C n ,   B n , 1 2 = K W n 1 2 B 1 2 ,   R n = K Z n + K W n > 0 ,   B n = 1 K W n K Z n + K W n 1 ,   K W n > 0 ,   c n ( , ) , Λ n ( , ) ,   K Z n 0 . □

3.3. Asymptotic Characterizations of Feedback Capacity Under Detectability and Stabilizability Versus Unit Circle Controllability

Now, we are ready to address C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) , by making use of the properties of generalized DREs and AREs of Section 3.2, under detectability and stabilizability or unit circle controllability. By invoking Theorem 2, we distinguish the following two cases.
(1) The sufficient and necessary conditions for existence of the limit such that C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = C F B ( κ ) , P V 1 are detectability and stabilizability, provided there exists a feasible solution to the optimization problem C F B ( κ ) , P V 1 . This leads to the solution of Regime 1.
(2) The sufficient and necessary conditions for existence of a limit such that C F B ( κ , P V 1 ) = lim n C n , F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) for some P V 1 are detectability and unit circle controllability (relaxation of stabilizability). This leads to the solutions of Regimes 2, 3.
First, we state the main theorem for (1).
Theorem 3. 
(Uniform asymptotic limit of feedback capacity for asymptotically time-invariant AR noise-detectability and stabilizability)
Consider Theorem 1 with asymptotically time-varying parameters of the stable and unstable noise, AR ( c n ) , c n ( , ) , n of Definition 1, i.e., lim n ( c n , K W n ) = ( c , K W ) , K W > 0 .
Define the set
P F B = { ( Λ , K Z ) ( , ) × [ 0 , ) | ( i )   t h e   p a i r   { A , C } { A , C ( Λ ) }   i s   d e t e c t a b l e , ( i i )   t h e   p a i r { A , B , 1 2 } { A ( K Z ) , B , 1 2 ( K Z ) } i s   s t a b i l i z a b l e , A , C , A , B , 1 2 g i v e n   b y ( 84 ) } .
Then
C F B ( κ , P V 1 ) = C F B ( κ ) = sup P F B ( κ ) 1 2 log Λ + c 2 K + K Z + K W K W , P V 1 , K V 1 0 ,
P F B ( κ ) = ( Λ , K Z ) P F B | K Z 0 , Λ 2 K + K Z κ s . t . K 0   i s   t h e   u n i q u e ,   s t a b i l i z a b l e   s o l . ,   i . e . ,   | F C L ( K , Λ , K Z ) | < 1   o f   t h e   g e n e r a l i z e d   A R E ,
K = c 2 K + K W K W + c K Λ + c 2 K Z + K W + Λ + c 2 K ,
F C L ( K , Λ , K Z ) = c M ( K , Λ , K Z ) Λ + c ,
M ( K , Λ , K Z ) = K W + c K Λ + c K Z + K W + Λ + c 2 K 1
provided there exists κ [ 0 , ) such that the set P F B ( κ ) is non-empty (if the set P F B ( κ ) is empty for some κ [ 0 , ) we should replace the pair { A , B , 1 2 } that is stabilizable by the pair that is unit circle controllable; this is treated in Corollary 2.
Moreover, C F B ( κ ) is achieved by the asymptotically time-invariant channel input,
X t = Λ t V t 1 V ^ t 1 + Z t , X 1 = Z 1 , t = 2 , , Z t N ( 0 , K Z t ) , K Z t 0 ,
lim n ( c n , K W n ) = ( c , K W ) , K W > 0 , lim n ( Λ n , K Z n ) = ( Λ , K Z ) , K Z 0 .
and the optimal element ( Λ , K Z ) = ( Λ , K Z ) P F B ( κ ) of C F B ( κ ) is such that
(i) if the noise is stable, i.e., c ( 1 , 1 ) then the input and the output processes ( X t , Y t ) , t = 1 , are asymptotic stationary, and
(ii) if the noise is unstable i.e., c ( 1 , 1 ) then the input and the innovations processes ( X t , I t ) , t = 1 , are asymptotic stationary.
Proof. 
This follows from Theorem 2, Proposition 1. □
Remark 7. 
Clearly, Theorem 3 imposes detectability and stabilizability, and as in all optimization problems with constraints, the set P F B ( κ ) must be non-empty. However, by Lemma 2, a necessary and sufficient condition for stabilizability is K Z > 0 . As we show shortly, for the parameters ( c , K W , κ ) of Regime 1 a feasible feedback strategy ( Λ , K Z ) , Λ 0 , K Z > 0 exists in the set P F B ( κ ) P F B . However, for the parameters ( c , K W , κ ) of Regimes 2 and 3, a feasible feedback strategy ( Λ , K Z ) , Λ 0 , K Z > 0 does not exist in the set P F B ( κ ) P F B . Therefore, for Regimes 2 and 3, we relax the stabilizability to a unit circle controllability, and then show there is a feasible feedback strategy ( Λ , K Z ) under this relaxation. However, in this case it is necessary that K V 1 > 0 .
Corollary 2 is a relaxation of Theorem 3 by replacing stabilizability by unit circle controllability.
Corollary 2. 
(Variations of Theorem 3—detectability and unit circle Ccontrollability)
Consider the statement of Theorem 3, and suppose the set P F B ( κ ) P F B is empty, i.e., there do not exist parameters ( c , K W , κ ) and ( Λ , K Z ) , Λ 0 , such that K Z > 0 .
Replace the set P F B in (95) (i.e., the detectability and the stabilizability conditions) by the larger set P F B , + P F B of detectability and unit circle controllability conditions, defined by
P F B , + = { ( Λ , K Z ) ( , ) × [ 0 , ) | ( i )   t h e   p a i r { A , C } { A , C ( Λ ) }   i s   d e t e c t a b l e , ( i i )   t h e   p a i r { A , B , 1 2 } { A ( K Z ) , B , 1 2 ( K Z ) }   i s   u n i t   c i r c l e   c o n t r o l l a b l e , A , C , A , B , 1 2 g i v e n b y ( 84 ) } .
Then feedback capacity is given by
C F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) = sup P F B , + ( κ ) 1 2 log Λ + c 2 K + K Z + K W K W , f o r   s o m e K 1 0
P F B , + ( κ ) = ( Λ , K Z ) P F B , + | K Z 0 , Λ 2 K + K Z κ ,
s . t .   K 0   i s   a   s t a b i l i z a b l e   s o l u t i o n ,   i . e . ,   | F C L ( K , Λ , K Z ) | < 1   o f   A R E ( 98 ) ( 100 )
provided there exists κ [ 0 , ) such that the set P F B , + ( κ ) is non-empty.
Moreover, C F B m a x ( κ , P V 1 ) is generally neither the capacity of asymptotically stationary noise nor the ergodic capacity.
If the optimal value is K Z = K Z = 0 , then C F B m a x ( κ , P V 1 ) = C F B ( κ , P V 1 ) , K V 1 > 0 and is achieved by the asymptotically time-invariant channel input (101) such that lim n K Z n = K Z = 0 .
Proof. 
Suppose in Theorem 3, the set P F B ( κ ) P F B is empty for some parameter values. Then by Theorem 2.(4), we can relax the conditions of detectability and stabilizability, and replace the set P F B by the larger set P F B , + specified by detectability and unit circle controllability. Under this relaxation, by Theorem 2.(4) the limit exists lim n K n = K for some initial condition, K 1 0 (not for all K 1 , i.e., K 1 = 0 may not be included), where K 0 is a stabilizing solution of the generalized ARE (83), but it may not be unique, i.e., Ref. (83) may have multiple solutions K 0 , where one of the solutions is the maximal and stabilizing solution, denoted by K + 0 , such that K + K 0 w.r.t. any other solution K 0 . Since in general, under detectability and unit circle cotrolllability conditions, the limit depends on K 1 0 and by (39), K 1 = K V 1 K Z 1 K V 1 + K Z 1 0 depends on K V 1 , then C F B m a x ( κ , P V 1 ) is neither the capacity of asymptotically stationary noise nor the ergodic capacity. The last part is obtained as follows. Suppose K Z = K Z = 0 is optimal for C F B m a x ( κ , P V 1 ) . By Lemma 2.(5), when K Z = K Z = 0 there are two solutions of the ARE (98)–(100), where one is the zero solution K = 0 and the other the maximal solution K = K + = K W Λ 2 1 Λ + c 2 ( 0 , ) , which is stabilizing, | F C L ( K , Λ , K Z = 0 ) = 1 Λ | < 1 , if | Λ | > 1 , c Λ . It is also verified in the proof of Theorem 5. Since the zero solution K = 0 , K Z = 0 gives zero value of C F B m a x ( κ , P V 1 ) , we pick the maximal solution K = K + > 0 which gives a positive value of C F B m a x ( κ , P V 1 ) , K 1 > 0 . By Theorem 2.(4) and K 1 = K V 1 K Z 1 K V 1 + K Z 1 0 , we require K V 1 > 0 , K Z 1 > 0 ; otherwise, the limit of K 1 , K 2 , , K n as n would correspond to the zero solution K = 0 . This implies C F B m a x ( κ , P V 1 ) = C F B ( κ , P V 1 ) , K V 1 > 0 and is achieved by the asymptotically time-invariant channel input (101) such that lim n K Z n = K Z = 0 . □
Remark 8. 
The point to be made is that there are fundamental differences between the optimization problems of Theorem 3 under the constraint of detectability and stabilizability, and Corollary 2, under the relaxed constraint of detectability and unit circle controllability. This will become even more clear in the next sections, where we derive the optimal values of capacity for Regimes 1–3.

3.4. Closed-Form Expressions of Asymptotic Feedback Capacity and Lower Bounds on Nonfeedback Capacity

In this section we apply KKT conditions to derive the closed-form expressions of feedback capacity for Regimes 1–3, i.e., C F B ( κ ) P V 1 , K V 1 0 and C F B m a x ( κ , P V 1 ) , P V 1 such that K V 1 > 0 .

3.4.1. Closed-Form Expressions of Ergodic Capacity C F B ( κ ) Under Detectability/Stabilizability

First, we give the KKT conditions for C F B ( κ ) defined by (96).
Lemma 3. 
(Optimality conditions for ergodic capacity of optimization problem C F B ( κ ) —detectability/stabilizability)
(a) Consider the optimization problem of feedback capacity C F B ( κ ) of (96), which presupposes that the optimal policy ( Λ , K Z ) P F B ( κ ) satisfies the detectability and stabilizability conditions.
Define the Lagrangian by
L ( Λ , K Z , K , λ ) = Λ + c 2 K + K Z + K W λ 1 { K c 2 K K W K Z + K W + Λ + c 2 K + K W + c K Λ + c 2 } λ 2 ( Λ ) 2 K + K Z κ λ 3 K λ 4 K Z , λ = ( λ 1 , λ 2 , λ 3 , λ 4 ) R 4 .
(i) The following hold.
Stationarity:
K Z L ( Λ , K Z , K , λ ) | Λ = Λ , K Z = K Z , K = K , λ = λ = 0 ,
Λ L ( Λ , K Z , K , λ ) | Λ = Λ , K Z = K Z , K = K , λ = λ = 0 ,
K L ( Λ , K Z , K , λ ) | Λ = Λ , K Z = K Z , K = K , λ = λ = 0 .
Complementary Slackness:
λ 2 ( Λ ) 2 K + K Z κ = 0 , λ 3 K = 0 , λ 4 K Z = 0 ,
λ 1 K c 2 K K W K Z + K W + Λ + c 2 K + K W + c K Λ + c 2 = 0 .
Primal Feasibility:
( Λ ) 2 K + K Z κ , K Z 0 , K 0 ,
K c 2 K K W K Z + K W + Λ + c 2 K + K W + c K Λ + c 2 0 .
Dual Feasibility:
λ 1 0 , λ 2 0 , λ 3 0 , λ 4 0 .
  • (ii) If K Z = 0 then K = 0 and C F B ( κ ) = 0 , κ [ 0 , ) .
  • (iii) A necessary condition for existence of κ ( 0 , ) such that C F B ( κ ) > 0 is λ 1 > 0 , λ 2 > 0 , λ 3 = 0 , λ 4 = 0 , and (114) holds with equality. (b) Suppose the optimization problem of feedback capacity C F B ( κ ) of (96) of part (a) does not produce a policy ( Λ , K Z ) P F B ( κ ) such that C F B ( κ ) > 0 , i.e., the KKT conditions of part (a) do not produce an optimal K > 0 and K Z > 0 .
Then feedback capacity is given by the optimization problem C F B m a x ( κ , P V 1 ) of (104), where the optimization is over the larger set P F B , + ( κ ) .
Proof. 
See Appendix A.2. □
Theorem 4 gives the complete closed-form expressions of the optimization problem C F B ( κ ) .
Theorem 4. 
(Closed-form expressions for ergodic capacity C F B ( κ ) —detectability/stabilizability)
(1) The non-zero value of C F B ( κ ) defined by (96), for asymptotically time-invariant parameters of the stable and unstable AR ( c ) noise, i.e, c ( , ) , with c 0 , c 1 , and strategies of the channel input, is given as follows.
C F B ( κ ) = 1 2 log Λ + c 2 K + K Z + K W K W , κ K 1 ( c , K W ) ,
= 1 2 log c 2 c 2 1 κ + K W c 2 1 K W
where
Λ = κ 1 c 2 + K W + c 2 K c c 2 2 K ( , ) , K Z + Λ 2 K = κ ,
K 1 ( c , K W ) = κ [ 0 , ) : K > 0 , K Z > 0 , c ( , ) , c 0 , c 1
and where K is the unique positive and stabilizing solution, i.e., | F C L ( K , Λ , K Z ) | < 1 , of the quadratic equation
c 4 c 2 1 K 2 + c 4 1 c 2 κ + K W K + 1 c 2 κ + K W 2 + 4 c 2 1 K W κ c 4 κ K W = 0 .
Further, for any κ K 1 ( c , K W ) , then
K = κ c 2 1 2 K W c 2 c 2 1 ( 0 , ) ,
Λ = c K W κ c 2 1 2 K W ( , ) ,
K Z = κ c 2 1 κ c 2 1 2 K W K W 2 c 2 1 κ c 2 1 2 K W ( 0 , ) .
(2) The non-zero value of C F B ( κ ) , κ K 1 ( c , K W ) of part (1), is restricted to the region:
(a) Regime 1. K 1 ( c , K W ) = κ [ 0 , ) | 1 < c 2 < , κ > K W + K W 4 c 2 3 2 c 2 1 2 .
Moreover, C F B ( κ ) , κ K 1 ( c , K W ) is achieved by the optimal channel input,
X n = Λ n V n 1 E V n 1 | Y n 1 + Z n , X 1 = Z 1 , Z n N ( 0 , K Z n ) , n = 1 , 2 , ,
s u c h   t h a t lim n ( Λ n , K Z n ) = ( Λ , K Z ) , K Z > 0 , t h e   u n i q u e   s o l u t i o n   o f   p a r t   ( 1 ) .
(3) A non-zero value of feedback capacity C F B ( κ ) , i.e., with Λ 0 , does not exist for the two regions:
(a) Regime 2. κ K 2 ( c , K W ) = κ [ 0 , ) | 1 < c 2 < , κ K W + K W 4 c 2 3 2 c 2 1 2 .
(b) Regime 3. κ K 3 ( c ) = κ [ 0 , ) | 0 c 2 1 .
Moreover, for Regions 2 and 3, a non-zero value exists for the feedback capacity optimization problem C F B m a x ( κ , P V 1 ) for some P V 1 such that K V 1 > 0 (with Λ 0 ).
Proof. 
See Appendix A.3. □

3.4.2. Closed-Form Expression of Capacity C F B ( κ , P V 1 ) Under Detectability/Unit Circle Controllability

Since for the complement of Regime 1 there does not exist a non-zero value solution C F B ( κ , P V 1 ) = C F B ( κ ) ( 0 , ) for K V 1 = 0 , we have have to consider Corollary 2 to derive the closed-form expression of C F B ( κ , P V 1 ) = C F B m a x ( κ , P V 1 ) , K V 1 > 0 , where the optimization is over the set P F B , + in (95), for Regimes 2 and 3.
Theorem 5. 
(Closed-form expression of capacity C F B m a x ( κ , P V 1 ) —detectability/unit circle controllability)
The nonergodic capacity C F B m a x ( κ , P V 1 ) of Corollary 2, where optimization is over the larger set P F B , + , is given by
C F B m a x ( κ , P V 1 ) = sup Λ ( , ) , ( Λ ) 2 K κ 1 2 log Λ + c 2 K + K W K W , K V 1 > 0 ,
s . t .   | Λ | 1 , K = K + ( Λ ) 0   i s   t h e   m a x i m a l   a n d   s t a b i l i z i n g   s o l .   o f   t h e   A R E ,
K = c 2 K + K W K W + c K Λ + c 2 K W + Λ + c 2 K , K 0 ,
K = lim n K n , K n = E V n E V n | Y n 2 .
Moreover, a non-zero value of feedback capacity C F B m a x ( κ , P V 1 ) exists for all K V 1 > 0 , for Regimes 2 and 3, and it given by
C F B m a x ( κ , P V 1 ) = log max { 1 , | Λ | } , K V 1 > 0 , κ K 2 ( c , K W ) K 3 ( c ) ,
where Λ is the maximal root of the fourth order equation [10],
K W Λ 4 K W + κ Λ 2 2 c κ Λ c 2 κ = 0 ,
K = K + = K W Λ 2 1 Λ + c 2 , | Λ | > 1
and C F B m a x ( κ , P V 1 ) is achieved by the optimal channel input,
X n = Λ n V n 1 E V n 1 | Y n 1 , n = 2 , 3 , , X 1 = Z 1 , Z 1 N ( 0 , K Z 1 ) , K Z 1 > 0 ,
s u c h   t h a t lim n ( Λ n , K Z n ) = ( Λ , K Z ) , K Z = 0 ,
( Λ , K Z , K ) = ( Λ , K Z , K + )   i s   t h e   m a x i m a l   s o l u t i o n   o f   C F B ( P V 1 , κ ) , K V 1 > 0 .
Proof. 
See Appendix A.4. □
In the next remark, we clarify and correct some of the misinterpretations of results in [2,24],
Remark 9. 
Previous papers [2,24] presented feedback capacity formula C F B = log max { 1 , | Λ | } for stable AR ( c ) noise, i.e., c ( 1 , 1 ) , which is included in Regime 3, κ K 3 ( c ) = κ [ 0 , ) | 0 c 2 1 . However, papers [2,24] claimed that their feedback capacity is C F B = log max { 1 , | Λ | } , and this is the capacity of asymptotically stationary noise V n . This statement should be read with caution because K V 1 > 0 should be imposed, and this excludes asymptotically stationary noise V n . However, C F B ( κ , P V 1 ) = C F B , K V 1 > 0 corresponds to the capacity of stationary noise V n . This means, the probability of error P e r r o r ( n ) ( P V 1 ) = P e r r o r ( n ) ( K V 1 ) , i.e., depends on K V 1 , and if K V 1 = 0 then lim n P e r r o r ( n ) ( P V 1 ) = 1 (see Gallager [4] for extensive discussion).
The clarifications of Remark 9 also apply to the TD feedback capacity formulae given in [2] [Theorem 6.1], for general stable noise described by the state space realization [2] [Equation (58)] and also to [24].
First, we state a theorem that first appeared in [27] (and earlier in an arXiv paper [26]).
Theorem 6. 
([26,27] Asymptotic feedback capacity with initial state)
Consider the Gaussian noise V n , n = 1 , 2 , described by the stable state space realization (see above and below [2]) [Equation (70), Equation (43)],
S t + 1 = A S t + B W t , V t = C S t + W t , t = 1 , 2 ,
where S t : Ω R n s is the state of the noise ( n s a finite positive integer), A , B , C are specific matrices of dimensions n s × n s , n s × 1 , 1 × n s , respectively, W t N ( 0 , K W ) , K W > 0 , t = 1 , , n is the independent Gaussian process, independent of S 1 N ( 0 , K S 1 ) , K S 1 0 , and the eigenvalues of A lie inside the unit disk in the space of complex numbers, i.e., A is a stable matrix, meaning its poles and zeros inside the unit disk ([2], Theorem 6.1 consider the value K W = 1 ).
Suppose Conditions 1 and 2 hold.
Condition 1.  The initial state of the noise or channel S 1 = s is known to the encoder and decoder.
Condition 2.  Given a fixed initial state S 1 = s , known to the encoder and the decoder, at each t, the state space realization of the noise is such that the noise V t 1 uniquely defines the state of the noise S t and vice-versa, t .
(1) An equivalent sequential characterization of Cover and Pombra [3] feedback capacity C n , F B ( κ , P V 1 ) is
C n , F B ( κ , P V 1 ) = C n , F B A S t a t N o i s e ( κ , s ) = sup { Λ t , K Z t } t = 1 n , 1 n t = 1 n Λ t K t Λ t T + K Z t κ 1 2 n t = 1 n log Λ t + C K t Λ t + C T + K Z t + K W K W ,
s . t . K Z t 0 , t , K t = E S t E S t | Y t 1 , S 1 = s S t E S t | Y t 1 , S 1 = s T
a sol. of the matrix DRE K t 0 of (139),
K t + 1 = A K t A T + B K W B T B K W + A K t Λ t + C T K W + Λ + C K Λ t + C T + K Z t 1 . B K W + A K Λ t + C T T , K 1 = 0 .
where K t R n s × n s and K t 0 , t means a symmetric positive semidefinite matrix.
Moreover, the optimal channel input corresponding to C n , F B A S t a t N o i s e ( κ , P V 1 ) is given by
X t = Λ t S t E S t | Y t 1 , S 1 = s + Z t , t = 1 , 2 , , n .
Z t N ( 0 , K Z t ) , K Z t 0 , t = 1 , 2 , , n o r t h o g o a l   p r o c e s s   a n d   i n d p .   o f   ( S t , V t , Y t 1 ) .
(2) Suppose the (i) detectability condition and (ii) unit circle controllability condition of the matrix algebraic Riccati equation (ARE) (144) hold.
The asymptotic feedback capacity is
C F B ( κ , P V 1 ) = C F B A S t a N o i s e ( κ , s ) = lim n C n , F B A S t a t N o i s e ( κ , s ) = sup Λ R 1 × n s , K Z , Λ K Λ T + K Z κ 1 2 log Λ + C K Λ + C T + K Z + K W K W ,
s . t . K Z 0 , K i s   a   p o s i t i v e   s e m i - d e f i n i t e   s o l u t i o n   K 0 o f   t h e   A R E ( 144 ) ,
K = A K A T + B K W B T B K W + A K Λ + C T K W + Λ + C K Λ + C T + K Z 1
. B K W + A K Λ + C T T
w h e r e   ( K Z , Λ , K )   a r e   t h e a s y m p t o t i c   l i m i t s ( K Z , Λ , K ) = lim n ( K Z n , Λ n , K n ) .
Moreover, the following hold.
(2.1) If the optimal pair ( Λ , K Z ) is such that K Z > 0 then C F B A S t a t N o i s e ( κ , s ) is independent of the initial state S 1 = s and is achieved by the optimal channel input,
X t = Λ S t E S t | Y t 1 , S 1 = s + Z t , t = 1 , 2 , , n .
Z t N ( 0 , K Z ) , K Z > 0 , t = 1 , 2 , , n o r t h o g o a l   p r o c e s s   a n d   i n d p .   o f   ( S t , V t , Y t 1 ) .
(2.2) If the optimal pair ( Λ , K Z ) is such that K Z = 0 then C F B A S t a t N o i s e ( κ , s ) depends on the initial state S 1 = s and is achieved by
X 1 = Z 1 N ( 0 , K Z 1 ) , K Z 1 > 0 , X t = Λ S t E S t | Y t 1 , S 1 = s , t = 2 , , n .
(3) The asymptotic characterization of feedback capacity [2] [Theorem 6.1] corresponds to C F B ( κ , P V 1 ) = C F B A S t a N o i s e ( κ , s ) , and optimal pair ( Λ , K Z ) such that K Z = 0 .
Proof. 
Statement (1) and (2) are shown in [26,27]. Statement (3) is also discussed in [26,27]. It also follows from the simple observation that according to Cover and Pombra [3], the n - FTFI capacity C n , F B ( κ , P V 1 ) corresponds to an optimal channel input, X t = j = 1 t 1 b t , j V j + Z ¯ t , with corresponding output Y t = j = 1 t 1 b t , j V j + Z ¯ t + V t , t = 1 , , n , where for some jointly Gaussian process Z ¯ n N ( 0 , K Z ¯ n ) and nonrandom coefficients b t , j . Clearly, Theorem 6 is inconsistent with the Cover and Pombra [3], the n - FTFI capacity C n , F B ( κ , P V 1 ) , unless the optimal input is expressed as a function of the state S n of the noise instead of the noise S n . It is now easy to verify that given a state space realization, Conditions 1 and 2 are necessary and sufficient for the validity of Theorem 6. To verify Conditions 1 and 2, note the following. If Condition 1 holds, i.e., S 1 = s is known to the encoder, then from the state space realization, (136), V 1 = C s 1 + B W 1 , hence W 1 = V 1 C s 1 is uniquely determined from ( V 1 , s 1 ) , and this specifies S 2 = A s 1 + B W 1 = A s 1 + B V 1 B C s 1 , etc., therefore Condition 2 holds. From [2] [Theorem 6.1], it is clear that optimal pair ( Λ , K Z ) is taken such that K Z = 0 (without a valid proof). □
Remark 10. 
We emphasize that C n , F B A S t a t N o i s e ( κ , P V 1 ) is specific to the state space realization, [2]) [Equation (70), Equation (43)], i.e., (136), and does not hold for the alternative realization, S t + 1 = A S t + B W t 1 , V t = C S t + W t 2 , t = 1 , , n , where W t 1 , t = 1 , , n and W t 2 , t = 1 , , n are two independent Gaussian processes.

3.4.3. Nonfeedback Ergodic Achievable Rates for Stable and Unstable AR Noise

By Theorem 3, for the value Λ = 0 , the pair { A , C } is detectable and the pair { A , B , 1 2 } is stabilizable, hence C F B ( κ ) | Λ = 0 is a nonfeedback achievable rate, i.e., by (101), for Λ = 0 , the channel input is an independent innovations process, X t = Z t , t = 1 , , n , and hence the code does not use feedback. In the next theorem we calculate C F B ( κ ) | Λ = 0 .
Theorem 7. 
Ergodic achievable rates without feedback for stable and unstable AR ( c ) noise
Consider the optimization problem of feedback capacity C F B ( κ ) of Theorem 3, with Λ = 0 , and corresponding channel input and output processes,
X t = Z t , t = 1 , , n ,
V t = c t V t 1 + W t , t = 2 , , , n ,
Y t = X t + V t = c t V t 1 + W t + Z t , Y 1 = Z 1 + W 1 , t = 2 , , n .
(1) A lower bound on the asymptotic characterization of nonfeedback capacity, C ( κ , P V 1 ) = lim n C n ( κ , P V 1 ) , is
C ( κ , P V 1 ) C F B ( κ ) | Λ = 0 = C L B ( κ ) = max P 0 , n f b ( κ ) 1 2 log c 2 K + K Z + K W K W , P V 1 , K V 1 0
where
P 0 , n f b ( κ ) = K Z P 0 , n f b | K Z κ , P 0 , n f b = { K Z [ 0 , ) | ( i )   t h e   p a i r { A , C } | Λ = 0 { A , C ( Λ ) } | Λ = 0   i s   d e t e c t a b l e ,
( i i )   t h e   p a i r { A , B , 1 2 } | Λ = 0 { A ( K Z ) , B , 1 2 ( K Z ) } | Λ = 0 i s   s t a b i l i z a b l e } ,
K 0   i s   t h e   u n i q u e   a n d   s t a b i l i z i n g   o f ( 156 ) , i . e . ,   | F n f b ( K , K Z ) | < 1 ,
K = c 2 K + K W K W + c 2 K 2 K Z + K W + c 2 K ,
F C L , n f b ( K , K Z ) = c M n f b ( K , K Z ) c , M n f b ( K , K Z ) = K W + c 2 K K Z + K W + c 2 K 1
provided there exists κ [ 0 , ) such that the set P 0 , n f b ( κ ) is non-empty.
Moreover, C L B ( κ ) is an achievable rate without feedback, independent of the initial state K 1 0 and
(i) if the noise is stable, i.e., c ( 1 , 1 ) then the input and the output processes ( X t , Y t ) , t = 1 , are asymptotically stationary;
(ii) if the noise is unstable i.e., c ( 1 , 1 ) then the input and the innovations processes ( X t , I t = Y t E { Y t | Y t 1 } ) , t = 1 , are asymptotic stationary.
(2) The lower bound on C ( κ , P Y 1 ) of (1) is given by
C L B ( κ ) = 1 2 log c 2 K + κ + K W K W , κ K , n f b ( c , K w ) ,
K , n f b ( c , K W ) = κ [ 0 , ) | K 0 = [ 0 , )
where K 0 is unique and stabilizing, and K Z , are given by
K = κ 1 c 2 K W + κ 1 c 2 + K W 2 + 4 c 2 K W κ 2 c 2 , c 0 , κ K , n f b ( c , K W ) , κ K W κ + K W , c = 0 , κ [ 0 , ) ,
K Z = κ .
Proof. 
(1) The statements follow from Theorem 3 as a special case. (2) This follows from (1), since the optimal K Z occurs on the boundary, i.e., K Z = κ . Then by substituting ( Λ , K Z ) = ( 0 , κ ) into the generalized ARE (156), we obtain
c 2 K 2 + K κ 1 c 2 + K W K W κ = 0 , c 0 .
From (162), the unique and non-negative stabilizing solution of the generalized ARE is given by (160), for c 0 . For c = 0 , the generalized ARE reduces to K κ + K W K W κ = 0 , and hence all equations under (2) are obtained. The validity of the stability condition, i.e., | F C L , n f b ( K , K Z ) | < 1 , although it is ensured by the conditions, can be easily verified. □
Remark 11. 
By Theorem 7.(2) the lower bound C L B ( κ ) , κ [ 0 , ) holds for asymptotically time-invariant stable and unstable AR(c) noise.
(1) For c = 0 , we recover, as expected, the capacity of the AGN channel with memoryless noise.
(2) For | c | = 1 , we obtain the lower bound C L B ( κ ) on the nonfeedback capacity given by
C L B ( κ ) = 1 2 log c 2 K + κ + K W K W , K = K W + K W 2 + 4 K W κ 2 , K Z = κ [ 0 , ) .
The above choice of a channel input strategy ensures the pair { A , C } | Λ = 0 = { c , c } is detectable, and the pair { A , B , 1 2 } | K Z = κ = { c K W κ + K W , K W 1 2 ( 1 K W κ + K W 1 } is stabilizable, for any κ ( 0 , ) (since K W > 0 ), and hence this is an ergodic achievable rate. We note that the nonfeedback capacity, when | c | ( 0 , 1 ) , cannot be handled by the frequency-domain water-filling formula [4] because V n is neither stationary nor asymptotically stationary.

4. Conclusions

This paper considers additive Gaussian noise (AGN) channels when the noise is nonstationary and nonergodic. It presents closed-form feedback capacity expressions when the noise is an autoregressive (AR) unstable process and an AR stable process (i.e., asymptotically stationary or stationary). These are described by three Regimes which depend on the parameters of the AR noise. It is shown that the feedback capacity for unstable AR noise is much higher compared to the feedback capacity for stable AR noise (i.e., asymptotically stationary or stationary). Similarly, it is shown that achievable lower bounds on nonfeedback rates for unstable AR noise are much higher than the nonfeedback capacity for stable AR noise. The derivations make use of the asymptotic properties of difference Riccati equations to their corresponding limits described by unique solutions of algebraic Riccati equations and KKT conditions. From the closed-form expressions, it follows that for unstable noise, the feedback capacity is an ergodic capacity, i.e., it does not depend on the initial distribution of the AR noise, while for stable noise it depends on the initial distribution noise, and hence it does not correspond to the feedback capacity for an asymptotically stationary noise. This observation is also shown to hold for more general noise described by the state space realization in [2]. This implies that the TD characterization of feedback capacity for stable noise stated in [2] [Theorem 4.1] is correct, under certain assumptions which are identified in this paper (these are precisely the assumptions of the earlier paper [20]), and the additional condition that the optimal covariance of the innovations part of the input is asymptotically zero. Our closed-form feedback capacity expressions provide fundamental insights and answers to the comment paper [1] that identified gaps in the proof of [2] [Theorem 4.1].

Author Contributions

Conceptualization, C.D.C., C.K., S.L. (Stelios Louka) and S.L. (Sergey Loyka); Methodology, C.D.C., C.K., S.L. (Stelios Louka) and S.L. (Sergey Loyka); Writing—review & editing, C.D.C., C.K., S.L. (Stelios Louka) and S.L. (Sergey Loyka). All authors have read and agreed to the published version of the manuscript.

Funding

The work of C.D. Charalambous was co-funded by the European Regional Development Fund and the Republic of Cyprus through the Research and Innovation Foundation (Project: EXCELLENCE/1216/0296).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of this study; in the collection, analyses, or interpretation of data; in the writing of this manuscript; or in the decision to publish the results.

Appendix A

Appendix A.1. Proof of Lemma 2

Recall (84) and (85). (1.a) By definition A = c , C = Λ + c . By c ( 1 , 1 ) , there always exists a G R such that | A G C | = | c G ( Λ + c ) | < 1 , i.e., take G = 0 . (1.b) Since Λ c , we can easily verify the existence of G R such that | c G ( Λ + c ) | < 1 . Of course, it follows from Lemma 1. (1.c) Since K Z > 0 then B , 1 2 0 and the statement follows from Lemma 1. This shows (1).
(2) Since K Z = 0 , then B = 1 K W K Z + K W 1 = 0 , B , 1 2 = K W 1 2 B 1 2 = 0 , A = c K W K W + K Z 1 C = c Λ c = Λ , and hence, there exists a G R such that | A B , 1 2 G | = | Λ | 1 if and only if | Λ | 1 . This shows (2).
(3) Since K Z = 0 , similar to the proof in (2), there exists a G R such that | A B , 1 2 G | = | Λ | < 1 if and only if | Λ | < 1 . This shows (3).
(4) Since c ( 1 , 1 ) , K Z = 0 , then, by (1), the pair { A , C } is detectable, by (2) the pair { A , B , 1 2 } is unit circle controllable if and only if | Λ | 1 , and by (3) the pair { A , B , 1 2 } is stabilizable if and only if | Λ | < 1 . By Theorem 2.(1), (4), we deduce the claim. This shows (4).
(5) Clearly, (89) is equivalent to the quadratic equation K 2 Λ + c 2 + K K W K W Λ 2 = 0 . Hence, the two solutions are (92). The last statement is also obtained by applying Theorem 2.(2), as follows. By (1), { A , C } is detectable, by (2), { A , B , 1 2 } is unit circle controllable if and only if | Λ | 1 , and by (3), { A , B , 1 2 } is stabilizable if and only if | Λ | < 1 . By invoking Theorem 2.(2), the last statement follows. On the other hand, it is easily verified from (92), that the uniqueness of solutions K 0 holds if and only if | Λ | < 1 , because for | Λ | > 1 there are two non-negative solutions. By (87), evaluated at K Z = 0 , K = 0 , we have F C L ( K , Λ , K Z ) | K Z = 0 , K = 0 = Λ . Hence, the non-negative solution K = 0 is unique and stabilizable if and only if | Λ | < 1 . Similarly, it can be shown that F C L ( K , Λ , K Z ) | K Z = 0 , K = K + is also stabilizable if and only if | Λ | > 1 .

Appendix A.2. Proof of Lemma 3

(a) (i) The conditions are a consequence of the optimization problem; Equation (114) are due to the standard relaxation of the generalized ARE to an inequality K c 2 K + K W K W + c K Λ + c 2 K Z + K W + Λ + c 2 , K 0 .
(ii) Suppose K Z = 0 . By (114), with K Z = 0 , then
K c 2 K K W K W + Λ + c 2 K + K W + c K Λ + c 2 0 ,
K K Λ + c 2 + K W 1 Λ 2 0 .
By Lemma 2, a necessary and sufficient condition for stabilizability of the pair { A , B , 1 2 } , when K Z = 0 , is | Λ | < 1 , and therefore (A2) is satisfied if and only if K lies in the region
K W Λ 2 1 Λ + c 2 K 0 , f o r | Λ | < 1 .
Since it must be that K 0 then necessarily K = 0 , which implies C F B ( κ ) = 0 , κ [ 0 , ) . (iii) By the stationarity conditions (108)–(110), with λ = ( λ 1 , λ 2 , λ 3 , λ 4 ) :
K Z L ( Λ , K Z , K , λ ) = 1 λ 1 K c 2 K K W λ 2 + λ 4 = 0 , Λ L ( Λ , K Z , K , λ ) = Λ + c K λ 1 { K c 2 K K W Λ + c K
+ K W + c K Λ + c c K } λ 2 Λ K = 0 , K L ( Λ , K Z , K , λ ) = Λ + c 2 λ 1 { 1 c 2 K Z + K W + Λ + c 2 K + K c 2 K K W Λ + c 2 + 2 c K W + c K Λ + c Λ + c }
λ 2 Λ 2 + λ 3 = 0 ,
Suppose λ 4 0 . Then, by complementary slackness (111), we have λ 4 K Z = 0 , which implies K Z = 0 , and hence by (ii), K = 0 . By complementary slackness (111), we also have λ 2 Λ 2 K + K Z κ = λ 2 0 κ = 0 , hence λ 2 = 0 . By (ii) it follows that C ( κ ) = 0 , κ [ 0 , ) , hence the rate is zero. Similarly, if λ 3 0 then K Z = 0 and K = 0 , which lead to a zero rate. However, it can be verified from Theorem 5 that in this case the feedback capacity is given by the relaxed optimization problem C F B m a x ( κ , P V 1 ) that imposes the relaxed conditions of detectability and unit circle controllability. Also, by Theorem 7, for Λ = 0 , K Z 0 , we exhibit a non-zero rate, which is a lower bound on the non-feedback rate. Next, consider the case λ 3 = 0 , λ 4 = 0 and λ 1 = 0 . Then, by (A4) λ 2 = 1 . However, when λ 2 = 1 equalities (A5) and (A6) hold only if c = 0 . Since, c 0 , otherwise the channel is memoryless, then the only choice is λ 1 > 0 . Moreover, since λ 1 > 0 , then in order to satisfy the complementary slackness condition (114), then inequality must hold with equality, i.e.,
K c 2 K K W K Z + K W + Λ + c 2 K + K W + c K Λ + c 2 = 0 .
Finally we consider the case λ 3 = 0 , λ 4 = 0 and λ 2 = 0 . Solving the system of Equations (111), (A4)–(A7), the following sets of solutions are obtained. The first solution is
K = K W + 1 1 c 2 , Λ = c 2 + K W c ( K W + 1 ) , K Z = K W ( c 2 + K W ) c 2 ( K W + 1 ) , λ 1 = 1 .
The second solution is
K = 0 , Λ = c 2 + 1 2 c , K Z = 0 , λ 1 = 1 K W .
The first solution is discarded since for K W > 0 , then, K Z < 0 , while the second solution is discarded due to (ii). Therefore λ 2 0 , which by the complementary slackness condition (111) implies that
( Λ , ) 2 K + K Z = κ
Thus, we have shown that a necessary condition for existence of κ ( 0 , ) such that C F B ( κ ) > 0 is λ 1 > 0 , λ 2 > 0 , λ 3 = 0 and λ 4 = 0 .
(b) If there does not exist a strategy in the set P F B ( κ ) such that C F B ( κ ) > 0 , then we show in Theorem 5 that we can always consider the relaxation C F B m a x ( κ , P V 1 ) , where the optimization problem is over the larger set P F B , + ( κ ) and exhibits a positive rate. This completes the proof.

Appendix A.3. Proof of Theorem 4

We prove the statements in several steps.
(i) First, we recall Lemma 3.(a).(ii) that states if K Z = 0 then K = 0 , and C F B ( κ ) = 0 , κ [ 0 , ) . In this case we shall consider the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) (see Theorem 5), to exhibit a non-zero rate.
(ii) Second, we recall Lemma 3.(a).(iii), which states a necessary condition for existence of a non-zero feedback rate for some κ ( 0 , ) is λ 1 > 0 , λ 2 > 0 , λ 3 = 0 and λ 4 = 0 . For the rest of the derivation we characterize the set of all values κ K 1 ( c , K W ) if such exist, and treat the case when K 1 ( c , K W ) is empty separately, using the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) (see Theorem 5).
(iii) Consider λ 1 > 0 , λ 2 > 0 , λ 3 = 0 and λ 4 = 0 . We solve the system of Equations (A4)–(A7) and (A10). First we solve the system of Equations (A4) and (A5) to obtain λ 1 and λ 2 as a function of { K , K Z , Λ } . By substituting λ 1 , λ 2 and K Z from (A10) in (A6), we obtain (118). Finally, by substituting K Z and Λ in (A7) we deduce the quadratic Equation (120). The two solutions of the the quadratic Equation (120) give rise to the following two solutions. The first solution is
K = K 1 = κ c 2 1 2 K W c 2 c 2 1 ,
Λ = Λ 1 = c K W κ c 2 1 2 K W ,
K Z = K Z 1 = κ κ c 2 1 2 K W c 2 1 K W 2 κ c 2 1 2 K W c 2 1 .
λ 1 = c 2 K W κ 1 c 2 , λ 2 = c 2 .
The second solution is
K = K 2 = κ K W c 2 , Λ = Λ 2 = c κ K W κ , K Z = K Z 2 = κ K W K W κ ,
λ 1 = c 2 κ K W 1 c 2 , λ 2 = 0 .
Solutions K = K 2 , Λ = Λ 2 , K Z = K Z 2 , ( λ 1 , λ 2 ) given by (A16) are not valid solutions, because if K = K 2 > 0 , then K Z = K Z 2 < 0 , and vice-versa. Thus, the only valid solution is Solution 1, from which all statements of (1) are obtained. It remains to show the statements under (2) and (3).
(iv) Consider c 2 < 1 and define the set
A 1 ( c , K W ) = κ [ 0 , ) | K = K 1 > 0 , K Z = K Z 1 > 0 , c 2 < 1 , c 0 , λ 1 > 0 , λ 2 > 0 .
Similarly, define the sets for c 2 > 1
A 2 ( c , K W ) = κ [ 0 , ) | K = K 1 > 0 , K Z = K Z 1 > 0 , c 2 > 1 , c 0 , λ 1 > 0 , λ 2 > 0
and the set for c 2 = 1
A 3 ( c , K W ) = κ [ 0 , ) | K = K 1 > 0 , K Z = K Z 1 > 0 , c 2 = 1 , c 0 , λ 1 > 0 , λ 2 > 0
The proof is then completed by determining the values of κ A 1 ( c , K W ) and κ A 2 ( c , K W ) , if such exist. If the set is empty then, we need to consider the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) (see Theorem 5).
(iv.1) For c 2 < 1 , we have the following
K = K 1 > 0 κ < K W ( c 2 1 ) 2 ,
λ 1 > 0 κ < K W ( 1 c 2 ) ,
K Z = K Z 1 > 0 κ 2 c 2 1 3 κ K W c 2 1 K W 2 > 0 ,
K W K W 4 c 2 3 2 1 c 2 2 < κ < K W + K W 4 c 2 3 2 1 c 2 2 , 4 c 2 3 .
Next, we show the set A 1 ( c , K W ) is empty. For 4 c 2 < 3 , κ 2 c 2 1 3 κ K W c 2 1 K W 2 < 0 , thus K Z < 0 . For 4 c 2 3 , and from (A21) and (A23), it suffices to show that
K W K W 4 c 2 3 2 1 c 2 2 > K W 1 c 2 , f o r 4 c 2 3 .
Clearly, after simple algebra, we can show that (A24) holds if c 2 1 2 > 0 , f o r 4 c 2 3 , thus it holds for all c 1 . Therefore, the set defined by A 1 ( c , K W ) is empty. In this case we consider the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) in Theorem 5.
(iv.2) For c 2 > 1 , we have the following
K = K 1 > 0 κ > K W ( c 2 1 ) 2 ,
λ 1 > 0 κ > K W ( 1 c 2 ) ,
K Z = K Z 1 > 0 κ 2 c 2 1 3 κ K W c 2 1 K W 2 > 0 ,
κ < K W K W 4 c 2 3 2 1 c 2 2 o r κ > K W + K W 4 c 2 3 2 1 c 2 2 .
Next, note that for c 2 > 1 the following inequalities hold
K W ( 1 c 2 ) < K W K W 4 c 2 3 2 1 c 2 2 < 0 < K W ( c 2 1 ) 2 < K W + K W 4 c 2 3 2 1 c 2 2 .
Then, from (A25)–(A29), we deduce that the set A 2 ( c , K W ) is non empty, only if the power κ satisfies κ > K W 1 + 4 c 2 3 2 1 c 2 2 . For values of κ K W 1 + 4 c 2 3 2 1 c 2 2 , since they do not belong in the set A 2 ( c , K W ) , we will consider the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) in Theorem 5.
(iv.3) For c 2 = 1 clearly the set A 3 ( c , K W ) is empty, thus we need to consider the relaxed expression of feedback capacity C F B m a x ( κ , P V 1 ) . Putting all the above together we obtain the statements under (2) and (3), where the last statement of (3) is shown by considering the feedback capacity optimization problem C F B m a x ( κ , P V 1 ) in Theorem 5. The proof is completed.

Appendix A.4. Proof of Theorem 5

By Theorem 4, under the constraint of detectability and stabilizability, there does not exist a non-zero value of the optimization problem C F B ( κ ) for Regimes 2 and 3. Further, stabilizability implies the optimal K Z > 0 for Regime 1. However, under the relaxations of detectability and unit circle controllability, the optimization problem C F B m a x ( κ , P V 1 ) is over the set P F B , + , as stated in Corollary 2, and unit circle controllability implies the relaxation K Z 0 . Since by the proof of Theorem 4 there is no non-zero value of C F B ( κ ) corresponding to Regimes 2 and 3 under the constraint K Z > 0 , then necessarily K Z = 0 , which implies the optimization problem is given by (126)–(128), where | Λ | 1 is due to unit circle controllability. Consequently, the solutions of the ARE (128) are
K = 0 u n i q u e   s t a b .   s o l . ,   | F C L ( K , Λ , K Z ) | K Z = 0 = Λ | < 1 , i f   | Λ | < 1 ,
K + ( Λ ) = K W Λ 2 1 Λ + c 2 > 0 m a x i m a l   s t a b .   s o l . ,   | F C L ( K , Λ , K Z ) | K Z = 0 1 Λ | < 1   i f   | Λ | > 1 .
where K + ( Λ ) > 0 , | Λ | > 1 is the maximal and stabilizing solution that gives a positive value of C F B m a x ( κ , P V 1 ) > 0 , while K ( Λ ) = 0 , | Λ | < 1 gives a zero value of C F B m a x ( κ , P V 1 ) = 0 . Finally, substituting K + ( Λ ) > 0 , | Λ | > 1 into the average power constraint and C F B m a x ( κ , P V 1 ) we obtain (130)–(132), where Λ = Λ denotes the optimal value. It remains to show (133)–(135) achieves C F B m a x ( κ , P V 1 ) = log max { 1 , | Λ | } > 0 , κ K 2 ( c , K W ) K 3 ( c ) , provided K V 1 > 0 , that is, it does not include K V 1 = 0 . To show the claim, we consider (133)–(135), and we determine the limit of K n = E V n E V n | Y n 2 , as n . By applying the Kalman-filter [28], K n , n = 1 , 2 , satisfy the recursions,
K n = c n 2 K n 1 + K W n K W n + c n K n 1 Λ n + c n 2 K W n + Λ n + c n 2 K n 1 , K n 0 , n = 2 , 3 , , K 1 = K V 1 K Z 1 K V 1 + K Z 1 ,
F n C L ( K n 1 , Λ ) = c n M n ( K n 1 , Λ n ) Λ n + c n ,
M n ( K n 1 , Λ n ) = K W n + c n K n 1 Λ n + c n K W n + Λ n + c n 2 K n 1 1 .
Fact 1. 
K V 1 = 0 implies K n = 0 , n = 1 , 2 , 3 , , hence lim n K n = K = 0 K + ( Λ ) , and by (126), C F B m a x = 0 ; consequently, C F B m a x cannot be the asymptotic feedback capacity.
Using Fact 1, the convergence lim n ( c n , K W n , Λ n ) = ( c , K W , Λ ) , and also when detectability and unit circle controllability are necessary and sufficient conditions for lim n K n = K , where K satisfies (128), then we have the following.
K = 0 , i f   | Λ | < 1 , K V 1 0 ,   t h e   u n i q u e ,   s t a b .   s o l . ,   | F C L ( K = 0 , Λ ) = Λ | < 1 ,
K + = K W ( Λ ) 2 1 Λ + c 2 > 0 , i f   | Λ | > 1 , K V 1 > 0 ,   t h e   m a x i m a l   a n d   s t a b .   s o l . ,   | F C L ( K , Λ ) = 1 Λ | < 1 ,
The limit of the average power is
lim n 1 n E t = 1 n X t 2 = lim n 1 n K Z 1 + 1 n t = 2 n Λ n 2 K t 1 = 0 i f | Λ | < 1 , K V 1 0 , corresponding   to   the   unique   and   stab .   sol . ( Λ ) 2 K W ( Λ ) 2 1 Λ + c 2 > 0 i f | Λ | > 1 , K V 1 > 0 , corresponding   to   the   max .   and   stab .   sol .
From the above equations we have the following. If K V 1 > 0 and | Λ | > 1 , the average power constraint is satisfied, lim n 1 n E t = 1 n X t 2 = ( Λ ) 2 K W ( Λ ) 2 1 Λ + c 2 = κ , which gives (131), and follows that C F B m a x = 1 2 log Λ + c 2 K + + K W K W = log | Λ | , | Λ | > 1 . If K V 1 = 0 , then lim n K n = K ( Λ ) = 0 and this is the unique and stabilizing solution, i.e., lim n F C L ( K n 1 , Λ n ) = F C L ( K , Λ ) = Λ , and | F C L ( K , Λ ) | < 1 if and only if | Λ | < 1 . If K V 1 = 0 the average power is lim n 1 n E t = 1 n X t 2 = 0 . Thus we have verified our statements.

References

  1. Derpich, M.S.; Ostergaard, J. Comments on “Feedback Capacity of Stationary Gaussian Channels”. IEEE Trans. Inf. Theory 2024, 70, 1848–1851. [Google Scholar] [CrossRef]
  2. Kim, Y.H. Feedback Capacity of Stationary Gaussian Channels. IEEE Trans. Inf. Theory 2010, 56, 57–85. [Google Scholar] [CrossRef]
  3. Cover, T.; Pombra, S. Gaussian Feedback Capacity. IEEE Trans. Inf. Theory 1989, 35, 37–43. [Google Scholar] [CrossRef]
  4. Gallager, R.T. Information Theory and Reliable Communication; John Wiley & Sons, Inc.: New York, NY, USA, 1968. [Google Scholar]
  5. Ihara, S. Information Theory for Continuous Systems; World Scientific: Singapore, 1993. [Google Scholar]
  6. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  7. Butman, S. A General Formulation of Linear Feedback Communications Systems with Solutions. IEEE Trans. Inf. Theory 1969, 15, 392–400. [Google Scholar] [CrossRef]
  8. Tienan, J.; Schalkwijk, J.P.M. An Upper Bound to the Capacity of Bandlimited Gaussian Autoregressive Channel with Noiseless Feedback. IEEE Trans. Inf. Theory 1974, 20, 311–316. [Google Scholar] [CrossRef]
  9. Wolfowitz, J. Signalling over a Gaussian Channel with Feedback and Autoregressive Noise. J. Appl. Probab. 1975, 12, 713–723. [Google Scholar] [CrossRef]
  10. Butman, S. Linear Feedback Rate Bounds for Regressive Channels. IEEE Trans. Inf. Theory 1976, 22, 363–366. [Google Scholar] [CrossRef]
  11. Ordentlich, E. A Class of Optimal Coding Schemes for Moving Average Additive Gaussian Noise Channels with Feedback. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), Trondheim, Norway, 27 June–1 July 1994; p. 467. [Google Scholar]
  12. Ozarow, L. Upper Bounds on the Capacity of Gaussian Channel with Feedback. IEEE Trans. Inf. Theory 1984, 36, 156–161. [Google Scholar] [CrossRef]
  13. Ozarow, L.H. Random Coding for Additive Gaussian Channels with Feedback. IEEE Trans. Inf. Theory 1984, 36, 17–22. [Google Scholar] [CrossRef]
  14. Dembo, A. On Gaussian Feedback Capacity. IEEE Trans. Inf. Theory 1989, 35, 1072–1076. [Google Scholar] [CrossRef]
  15. Ihara, S.; Yanaki, K. Capacity of Discrete-Time Gaussian Channels with and without Feedback—II. Jpn. J. Appl. Math. 1989, 6, 245–258. [Google Scholar] [CrossRef]
  16. Yanaki, K. Necessary and Sufficient Conditions for the Capacity of the Discrete-Time Gaussian Channel to Be Increased by Feedback. IEEE Trans. Inf. Theory 1992, 38, 1788–1791. [Google Scholar] [CrossRef]
  17. Yanaki, K. An Upper Bound on the Discrete-Time Gaussian Channel with Feedback—II. IEEE Trans. Inf. Theory 1994, 40, 588–593. [Google Scholar] [CrossRef][Green Version]
  18. Chen, H.W.; Yanaki, K. Refinements of the Half-Bit and Factor-of-Two Bounds for Capacity in Gaussian Channels with Feedback. IEEE Trans. Inf. Theory 1999, 45, 319–325. [Google Scholar] [CrossRef]
  19. Chen, H.W.; Yanaki, K. Upper Bounds on the Capacity of Discrete-Time Blockwise White Gaussian Channels with Feedback. IEEE Trans. Inf. Theory 2000, 43, 1125–1131. [Google Scholar] [CrossRef]
  20. Yang, S.; Kavcic, A.; Tatikonda, S. On Feedback Capacity of Power-Constrained Gaussian Noise Channels with Memory. IEEE Trans. Inf. Theory 2007, 53, 929–954. [Google Scholar] [CrossRef]
  21. Liu, T.; Han, G. The ARMAk Gaussian Feedback Capacity. In Proceedings of the IEEE International Symposium on Information Theory Proceedings (ISIT), Aachen, Germany, 25–30 June 2017; pp. 211–215. [Google Scholar]
  22. Liu, T.; Han, G. Feedback Capacity of Stationary Gaussian Channels Further Examined. IEEE Trans. Inf. Theory 2019, 64, 2494–2506. [Google Scholar] [CrossRef]
  23. Li, C.; Elia, N. Youla Coding and Computation of Gaussian Feedback Capacity. IEEE Trans. Inf. Theory 2019, 64, 3197–3215. [Google Scholar] [CrossRef]
  24. Gattami, A. Feedback Capacity of Gaussian Channels Revisited. IEEE Trans. Inf. Theory 2019, 65, 1948–1960. [Google Scholar] [CrossRef]
  25. Ihara, S. On the Feedback Capacity of the First-Order Moving Average Gaussian Channel. Jpn. J. Stat. Data Sci. 2019, 2, 491–506. [Google Scholar] [CrossRef]
  26. Charalambous, C.D.; Kourtellaris, C.; Louka, S. New Formulas of Feedback Capacity for AGN Channels with Memory: A Time-Domain Sufficient Statistic Approach. arXiv 2020, arXiv:2010.06226. [Google Scholar] [CrossRef] [PubMed]
  27. Charalambous, C.D.; Kourtellaris, C.; Louka, S. New Formulas of Feedback Capacity for AGN Channels with Memory: A Time-Domain Sufficient Statistic Approach. Entropy 2025, 27, 207. [Google Scholar] [CrossRef]
  28. Caines, P.E. Linear Stochastic Systems; Wiley Series in Probability and Statistics; John Wiley & Sons, Inc.: New York, NY, USA, 1988. [Google Scholar]
  29. Kailath, T.; Sayed, A.; Hassibi, B. Linear Estimation; Prentice Hall: Hoboken, NJ, USA, 2000. [Google Scholar]
  30. Charalambous, C.D.; Kourtellaris, C.; Louka, S. Sequential Characterization of Cover and Pombra Gaussian Feedback Capacity: Generalizations to MIMO Channels via a Sufficient Statistic. In Proceedings of the 2021 IEEE Information Theory Workshop (ITW), Kanazawa, Japan, 17–21 October 2021. [Google Scholar]
  31. Kourtellaris, C.; Charalambous, C.D.; Loyka, S. New Formulas for Ergodic Feedback Capacity of AGN Channels Driven by Stable and Unstable Autoregressive Noise. In Proceedings of the International Symposium on Information Theory (ISIT), Los Angeles, CA, USA, 21–26 June 2020; pp. 2073–2078. [Google Scholar]
  32. Charalambous, C.D.; Kourtellaris, C.; Loyka, S. Control-Coding Capacity of Decision Models Driven by Correlated Noise and Gaussian Application Examples. In Proceedings of the IEEE Conference on Decision and Control (CDC), Miami Beach, FL, USA, 17–19 December 2018; pp. 4967–4972. [Google Scholar]
  33. Kourtellaris, C.; Charalambous, C.D.; Loyka, S. From Feedback Capacity to Tight Achievable Rates without Feedback for AGN Channels with Stable and Unstable Autoregressive Noise. In Proceedings of the International Symposium on Information Theory (ISIT), Los Angeles, CA, USA, 21–26 June 2020; pp. 2091–2096. [Google Scholar]
  34. Gelfand, M.I.; Yaglom, M. Calculation of the Amount of Information about a Random Function Contained in Another Such Function. Am. Math. Soc. Transl. 1959, 2, 199–246. [Google Scholar]
  35. van Schuppen, J.H. Control and System Theory of Discrete-Time Stochastic Systems; Springer Nature AG: Cham, Switzerland, 2021. [Google Scholar]
Figure 1. Additive Gaussian noise channel with jointly Gaussian noise V n = { V 1 , V 2 , , V n } . If the switch is open the encoder does not use feedback; if the switch is closed the encoder uses feedback.
Figure 1. Additive Gaussian noise channel with jointly Gaussian noise V n = { V 1 , V 2 , , V n } . If the switch is open the encoder does not use feedback; if the switch is closed the encoder uses feedback.
Entropy 27 01264 g001
Figure 2. Comparisons of Regime 1, C F B ( κ ) , Regimes 2, 3, C F B m a x ( P V 1 , κ ) = log | Λ | , and lower bound on nonfeedback capacity C L B ( κ ) , for c = 0.5 , c = 1.5 , c = 2 and K W = 1 , as a function of transmit power κ .
Figure 2. Comparisons of Regime 1, C F B ( κ ) , Regimes 2, 3, C F B m a x ( P V 1 , κ ) = log | Λ | , and lower bound on nonfeedback capacity C L B ( κ ) , for c = 0.5 , c = 1.5 , c = 2 and K W = 1 , as a function of transmit power κ .
Entropy 27 01264 g002
Figure 3. Illustration that Regimes 2 and 3 C F B m a x ( κ , P V 1 ) = log | Λ | , | Λ | > 1 are not optimal compared to Regime 1.
Figure 3. Illustration that Regimes 2 and 3 C F B m a x ( κ , P V 1 ) = log | Λ | , | Λ | > 1 are not optimal compared to Regime 1.
Entropy 27 01264 g003
Figure 4. AR ( c ) , noise, c = 0.75 and K W = 1 . Comparison of nonfeedback capacity C ( κ ) of water-filling Formula (33) (see Equation (5.5.14) in [5] and Equation (6) in [10]), and lower bound C L B ( κ ) =  (31) of transmitting an IID channel input Z t N ( 0 , κ ) for κ [ κ m i n , ) , where κ m i n = 13.7 is the minimum threshold (the values correspond to the maximum difference).
Figure 4. AR ( c ) , noise, c = 0.75 and K W = 1 . Comparison of nonfeedback capacity C ( κ ) of water-filling Formula (33) (see Equation (5.5.14) in [5] and Equation (6) in [10]), and lower bound C L B ( κ ) =  (31) of transmitting an IID channel input Z t N ( 0 , κ ) for κ [ κ m i n , ) , where κ m i n = 13.7 is the minimum threshold (the values correspond to the maximum difference).
Entropy 27 01264 g004
Figure 5. AR ( c ) noise, with c = 1.5 and K W = 1 . Time-sharing between the feedback strategy of Regime 1, X t o = Λ , V t 1 V ^ t 1 o + Z t o , with values ( Λ , , K Z , ) given by (122) and (123) and the nonfeedback strategy X t o = Z t o , with value K Z , = κ of the achievable lower bound.
Figure 5. AR ( c ) noise, with c = 1.5 and K W = 1 . Time-sharing between the feedback strategy of Regime 1, X t o = Λ , V t 1 V ^ t 1 o + Z t o , with values ( Λ , , K Z , ) given by (122) and (123) and the nonfeedback strategy X t o = Z t o , with value K Z , = κ of the achievable lower bound.
Entropy 27 01264 g005
Table 1. c = 2 , K W = 1 , κ [ 0 , ) , κ m i n = 0.2558639597 . Comparisons of (i) Regime 1, feedback capacity C F B ( κ ) of (20), and Regime 2, lower bound on nonfeedback capacity C L B ( κ ) of (31) incurred by independent channel inputs, with (ii) solution of semi-definite program of [24], and (iii) C F B m a x = log | Λ | , | Λ | > 1 of (25). Numbers are rounded to the nearest decimal points.
Table 1. c = 2 , K W = 1 , κ [ 0 , ) , κ m i n = 0.2558639597 . Comparisons of (i) Regime 1, feedback capacity C F B ( κ ) of (20), and Regime 2, lower bound on nonfeedback capacity C L B ( κ ) of (31) incurred by independent channel inputs, with (ii) solution of semi-definite program of [24], and (iii) C F B m a x = log | Λ | , | Λ | > 1 of (25). Numbers are rounded to the nearest decimal points.
Comp. of C F B , C m a x and C L B with Numerical Sol. of Semi-Definite Programming (SDP) [24]
κ κ m i n R e g . 1 , C F B Reg. 1–3, C L B SDP [24] R e g . 2 , C F B m a x optimal
0.01no feasible set0.03560941220.06009255150.0600925523
0.03no feasible set0.10005260160.15395956040.1539595679
0.09no feasible set0.25894852460.34132680000.3413268001
0.19no feasible set0.45599304470.53193710430.5319371060
0.25no feasible set0.55474623430.61833303580.6183330413
κ > κ m i n Reg. 1, C F B opt.Reg. 1–3, C L B SDP [24]Reg. 2,  C F B m a x not optimal
0.300.67404065210.61932431030.67404064860.6726584307
0.330.70759088080.65758485700.70759087750.7041839537
1.511.44320471941.43610840331.44320471881.2924606576
3.131.89636837371.89426621001.89636837151.6247312241
6.762.41401541302.41350100322.41401541252.0077561174
Table 2. c = 2 , K W = 1 , κ [ 0 , ) , κ m i n = 0.2558639597 . Comparisons of optimal values ( Λ , K Z ) of (i) Regime 1, C F B ( κ ) and Regime 2, C L B ( κ ) incurred by independent channel inputs, i.e., with Λ = 0 with (ii) solution of semi-definite program of [24], and (iii) C F B m a x = log | Λ | , | Λ | > 1 of [2]. Numbers are rounded to the nearest decimal points.
Table 2. c = 2 , K W = 1 , κ [ 0 , ) , κ m i n = 0.2558639597 . Comparisons of optimal values ( Λ , K Z ) of (i) Regime 1, C F B ( κ ) and Regime 2, C L B ( κ ) incurred by independent channel inputs, i.e., with Λ = 0 with (ii) solution of semi-definite program of [24], and (iii) C F B m a x = log | Λ | , | Λ | > 1 of [2]. Numbers are rounded to the nearest decimal points.
Comparison of ( Λ , K Z ) of C F B , C m a x and C L B and Numerical Sol. of SDP [24]
Reg. 1–3, C L B SDP [24]Regimes 2, 3, C F B m a x opt.
κ κ m i n Λ K Z Λ K Z Λ K Z
0.0100.01021.04253.9172 × 10 10 1.04250
0.0300.03041.11265.5728 × 10 9 1.11260
0.0900.09101.26693.2046 × 10 10 1.26690
0.1900.19201.44581.9058 × 10 8 1.44590
0.2500.25581.53478.6363 × 10 5 1.53510
of Reg. 1, C F B opt.SDP [24]Reg. 2, 3, C F B m a x not opt.
κ > κ m i n Λ K Z Λ K Z Λ K Z
0.301.15750.11021.15740.11021.59400
0.330.99970.16680.99950.16681.62920
1.510.15831.48880.15821.48882.44940
3.130.07363.11910.07353.11913.08390
6.760.03346.76210.03336.76214.02160
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Charalambous, C.D.; Kourtellaris, C.; Louka, S.; Loyka, S. The Capacity Gains of Gaussian Channels with Unstable Versus Stable Autoregressive Noise. Entropy 2025, 27, 1264. https://doi.org/10.3390/e27121264

AMA Style

Charalambous CD, Kourtellaris C, Louka S, Loyka S. The Capacity Gains of Gaussian Channels with Unstable Versus Stable Autoregressive Noise. Entropy. 2025; 27(12):1264. https://doi.org/10.3390/e27121264

Chicago/Turabian Style

Charalambous, Charalambos D., Christos Kourtellaris, Stelios Louka, and Sergey Loyka. 2025. "The Capacity Gains of Gaussian Channels with Unstable Versus Stable Autoregressive Noise" Entropy 27, no. 12: 1264. https://doi.org/10.3390/e27121264

APA Style

Charalambous, C. D., Kourtellaris, C., Louka, S., & Loyka, S. (2025). The Capacity Gains of Gaussian Channels with Unstable Versus Stable Autoregressive Noise. Entropy, 27(12), 1264. https://doi.org/10.3390/e27121264

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop