Next Article in Journal
Effect of a Magnetic Quadrupole Field on Entropy Generation in Thermomagnetic Convection of Paramagnetic Fluid with and without a Gravitational Field
Previous Article in Journal
Numerical Study of the Magnetic Field Effects on the Heat Transfer and Entropy Generation Aspects of a Power Law Fluid over an Axisymmetric Stretching Plate Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Complexity Reduction of Coding WSS Vector Processes by Using a Sequence of Block Circulant Matrices

by
Jesús Gutiérrez-Gutiérrez
1,2,*,
Marta Zárraga-Rodríguez
2,
Xabier Insausti
1,2 and
Bjørn O. Hogstad
3
1
CEIT, Manuel Lardizábal 15, 20018 San Sebastián, Spain
2
TECNUN, University of Navarra, Manuel Lardizábal 13, 20018 San Sebastián, Spain
3
Department of Mathematical Sciences, Norwegian University of Science and Technology, Teknologiveien 22, 2815 Gjøvik, Norway
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(3), 95; https://doi.org/10.3390/e19030095
Submission received: 20 December 2016 / Revised: 23 February 2017 / Accepted: 27 February 2017 / Published: 2 March 2017
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
In the present paper, we obtain a result on the rate-distortion function (RDF) of wide sense stationary (WSS) vector processes that allows us to reduce the complexity of coding those processes. To achieve this result, we propose a sequence of block circulant matrices. In addition, we use the proposed sequence to reduce the complexity of filtering WSS vector processes.

1. Introduction

Applications where Toeplitz matrices arise commonly involve computation of inverses, products, and eigenvalues. Although inverses and products of Toeplitz matrices are not in general Toeplitz, inverses and products of circulant matrices are circulant. Moreover, unlike for Toeplitz matrices, an eigenvalue decomposition is known for any circulant matrix. Hence, the substitution of Toeplitz matrices by circulant matrices leads to a notable reduction of the computational complexity in those applications.
A sequence of Toeplitz matrices { T n ( f ) } generated by a function f can be found in any information-theoretic or statistical signal processing application in which a wide sense stationary (WSS) process appears. In those applications, { T n ( f ) } is the sequence of correlation matrices of the WSS process and the function f is its power spectral density (PSD).
The substitution of the sequence { T n ( f ) } by a sequence of circulant matrices can be done when both sequences are asymptotically equivalent. In [1], Gray gave an example of such a sequence of circulant matrices { C n ( f ) } , where each circulant matrix C n ( f ) depends on the entire sequence { T n ( f ) } . However, in practical situations, what we usually know is { T n ( f ) } n n 0 for certain natural number n 0 . For those practical situations, Pearl presented in [2] a different sequence of circulant matrices { C ^ n ( f ) } , which is also asymptotically equivalent to the sequence of Toeplitz matrices { T n ( f ) } , but this satisfies the condition that each circulant matrix C ^ n ( f ) only depends on the corresponding Toeplitz matrix T n ( f ) .
In the present paper, the sequence { C ^ n ( f ) } is generalized to block circulant matrices. This sequence of block circulant matrices is used to obtain a result on the rate-distortion function (RDF) of WSS vector processes that allows us to reduce the complexity of coding those processes. In addition, we use that sequence to reduce the complexity of filtering WSS vector processes.
The paper is organized as follows. In Section 2, we set up notation and review the mathematical definitions and results used in the rest of the paper. In Section 3, we define the sequence of block circulant matrices considered in this paper and show its main properties. Finally, in Section 4 and Section 5, we use this sequence of block circulant matrices to reduce the complexity of coding and filtering WSS vector processes, respectively.

2. Preliminaries

2.1. Notation

In this paper, N , Z , R , and C denote the set of natural numbers (i.e., the set of positive integers), the set of integer numbers, the set of (finite) real numbers, and the set of (finite) complex numbers, respectively. If m , n N , then C m × n , 0 m × n , and I n are the set of all m × n complex matrices, the m × n zero matrix, and the n × n identity matrix, respectively. The symbol ∗ denotes conjugate transpose, E stands for expectation, i is the imaginary unit, tr denotes trace, δ stands for the Kronecker delta, ⊗ is the Kronecker product, λ k ( A ) , k { 1 , , n } , are the eigenvalues of an n × n Hermitian matrix A arranged in decreasing order, and σ k ( B ) , k { 1 , , n } , are the singular values of an n × n matrix B arranged in decreasing order.
Let { x n : n N } be a random N-dimensional vector process, i.e., x n is a random (column) vector of dimension N for all n N . We denote by x n : 1 the random vector of dimension n N given by
x n : 1 : = x n x n 1 x n 2 x 1 , n N .

2.2. Block Toeplitz Matrices

We first review the concept of block Toeplitz matrix.
Definition 1.
An m × n block Toeplitz matrix with M × N blocks is an m M × n N matrix of the form
F 0 F 1 F 2 F 1 n F 1 F 0 F 1 F 2 n F 2 F 1 F 0 F 3 n F m 1 F m 2 F m 3 F 0 ,
where F k C M × N with k { 1 n , , m 1 } .
Consider a matrix-valued function of a real variable F : R C M × N , which is continuous and 2 π -periodic. For every n N , we denote by T n ( F ) the n × n block Toeplitz matrix with M × N blocks given by
T n ( F ) : = ( F j k ) j , k = 1 n ,
where { F k } k Z is the sequence of Fourier coefficients of F:
F k = 1 2 π 0 2 π e k ω i F ( ω ) d ω k Z .
It should be mentioned that T n ( F ) is Hermitian for all n N if and only if F ( ω ) is Hermitian for all ω R (see [3] (Theorem 4.4.1)). Furthermore, in this case, from [3] (Theorem 4.4.2) (it was previously given in [4] (p. 5674) but without a proof) and [5] (Corollary VI.1.6), we obtain
min ω [ 0 , 2 π ] λ N ( F ( ω ) ) = inf ( F ) λ n N ( T n ( F ) ) λ 1 ( T n ( F ) ) sup ( F ) = max ω [ 0 , 2 π ] λ 1 ( F ( ω ) ) n N .

2.3. Block Circulant Matrices

We first review the concept of a block circulant matrix.
Definition 2.
An n × n block circulant matrix with M × N blocks is an n × n block Toeplitz matrix with M × N blocks of the form
C 0 C 1 C 2 C 1 n C 1 n C 0 C 1 C 2 n C 2 n C 1 n C 0 C 3 n C 1 C 2 C 3 C 0 ,
where C k C M × N with k { 1 n , , 0 } .
The next result [6] (Lemma 3) characterizes block circulant matrices.
Lemma 1.
Let C C n M × n N ; then, the following statements are equivalent:
1.
C is an n × n block circulant matrix with M × N blocks.
2.
There exist A 1 , , A n C M × N such that
C = ( V n I M ) diag ( A 1 , , A n ) ( V n I N ) * ,
where diag ( A 1 , , A n ) = diag 1 k n ( A k ) : = ( δ j , k A j ) j , k = 1 n and V n is the n × n Fourier unitary matrix
[ V n ] j , k : = 1 n e 2 π ( j 1 ) ( k 1 ) n i , j , k { 1 , , n } .

2.4. Asymptotically Equivalent Sequences of Matrices

We now review the concept of asymptotically equivalent sequences of matrices introduced in [6] (Definition 2), which is an extension of the original concept given by Gray in [7] to sequences of non-square matrices.
Definition 3.
Consider two strictly increasing sequences of natural numbers { d n ( 1 ) } and { d n ( 2 ) } . Let A n and B n be d n ( 1 ) × d n ( 2 ) matrices for all n N . We say that the sequences { A n } and { B n } are asymptotically equivalent, and write { A n } { B n } , if { A n 2 } and { B n 2 } are bounded, and
lim n A n B n F n = 0 ,
where · 2 and · F are the spectral norm and the Frobenius norm, respectively (The definition and the main properties of these two matrix norms can be found, e.g., in [3] (Section 2.1)).
We finish this section by reviewing [6] (Lemma 4).
Lemma 2.
Let F : R C M × N be continuous and 2 π -periodic. Then,
{ T n ( F ) } { C n ( F ) } ,
where C n ( F ) is the n × n block circulant matrix with M × N blocks given by
C n ( F ) : = ( V n I M ) diag F ( 0 ) , F 2 π n , , F 2 π ( n 1 ) n ( V n I N ) * n N .
When M = N = 1 and F is in the Wiener class (We recall that a function F : R C M × N is said to be in the Wiener class if it is continuous and 2 π -periodic, and it satisfies k = F k F < , see, e.g., [3] (Appendix B)), { C n ( F ) } is the sequence of circulant matrices defined by Gray in [1] (Equation (4.32)), see [3] (Lemma 5.2).

3. Sequence of Block Circulant Matrices Considered

We begin this section by presenting a sequence of block circulant matrices { C ^ n ( F ) } , where each block circulant matrix C ^ n ( F ) only depends on the corresponding block Toeplitz matrix T n ( F ) .
Definition 4.
Let F : R C M × N be continuous and 2 π -periodic. For every n N , we define { C ^ n ( F ) } as the n × n block circulant matrix with M × N blocks given by
C ^ n ( F ) : = ( V n I M ) diag 1 k n [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k ( V n I N ) * .
When M = N = 1 and the matrices T n ( F ) are real and symmetric, { C ^ n ( F ) } is the sequence of circulant matrices defined by Pearl in [2].
The following result gives an expression for the blocks of the block circulant matrix C ^ n ( F ) .
Lemma 3.
Consider n N and let F : R C M × N be continuous and 2 π -periodic.
1.
If j , k { 1 , , n } , then
[ C ^ n ( F ) ] j , k = 1 | j k | n F j k + | j k | n F j k n sgn ( j k ) ,
where sgn is the sign function.
2.
If T n ( F ) is real, then C ^ n ( F ) is real.
Proof. 
(1) Fix j , k { 1 , , n } . For convenience, we denote diag 1 k n [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k by D n ( F ) . We have
[ C ^ n ( F ) ] j , k = h = 1 n [ V n I M ] j , h [ D n ( F ) ( V n I N ) * ] h , k = h = 1 n [ V n ] j , h I M l = 1 n [ D n ( F ) ] h , l [ ( V n I N ) * ] l , k = h = 1 n [ V n ] j , h [ D n ( F ) ] h , h [ V n * I N * ] h , k = h = 1 n [ V n ] j , h [ D n ( F ) ] h , h [ V n * ] h , k I N = h = 1 n [ V n ] j , h [ V n ] k , h ¯ [ ( V n I M ) * T n ( F ) ( V n I N ) ] h , h = h = 1 n [ V n ] j , h [ V n ] k , h ¯ l = 1 n [ ( V n I M ) * ] h , l [ T n ( F ) ( V n I N ) ] l , h = h = 1 n [ V n ] j , h [ V n ] k , h ¯ l = 1 n [ V n * ] h , l I M s = 1 n [ T n ( F ) ] l , s [ V n I N ] s , h = h = 1 n [ V n ] j , h [ V n ] k , h ¯ l = 1 n [ V n ] l , h ¯ s = 1 n F l s [ V n ] s , h I N = h , l , s = 1 n [ V n ] j , h [ V n ] k , h ¯ [ V n ] l , h ¯ [ V n ] s , h F l s = 1 n 2 h , l , s = 1 n e 2 π ( j 1 ) ( h 1 ) n i e 2 π ( k 1 ) ( h 1 ) n i e 2 π ( l 1 ) ( h 1 ) n i e 2 π ( s 1 ) ( h 1 ) n i F l s = 1 n 2 h , l , s = 1 n e 2 π ( h 1 ) ( k j + l s ) n i F l s = 1 n 2 l , s = 1 n h = 1 n e 2 π ( k j + l s ) n i h 1 F l s = 1 n 2 l , s = 1 k j + l s n Z n n F l s + l , s = 1 k j + l s n Z n 1 e 2 π ( k j + l s ) n i n 1 e 2 π ( k j + l s ) n i F l s = 1 n l , s = 1 k j + l s n Z n F l s ,
where z ¯ denotes the conjugate of z C .
If j = k , then
[ C ^ n ( F ) ] j , k = 1 n l , s = 1 l s = 0 n F l s = 1 n l = 1 n F 0 = F 0 .
If j > k , then
[ C ^ n ( F ) ] j , k = 1 n l , s = 1 k j + l s { n , 0 } n F l s = 1 n l , s = 1 l s = j k n F l s + 1 n l , s = 1 l s = j k n n F l s = 1 n ( n | j k | ) F j k + 1 n ( n | j k n | ) F j k n = 1 | j k | n F j k + j k n F j k n .
If j < k , then
[ C ^ n ( F ) ] j , k = 1 n l , s = 1 k j + l s { 0 , n } n F l s = 1 n l , s = 1 l s = j k n F l s + 1 n l , s = 1 l s = j k + n n F l s = 1 n ( n | j k | ) F j k + 1 n ( n | j k + n | ) F j k + n = 1 | j k | n F j k + k j n F j k + n .
(2) It is a direct consequence of Assertion (1). ☐
When M = N = 1 and the matrices T n ( F ) are real and symmetric, from Lemma 3, we obtain [2] (Equation (7)).
We now show that the block Toeplitz matrix T n ( F ) can be approximated by the block circulant matrix C ^ n ( F ) for large n.
Lemma 4.
If F : R C M × N is continuous and 2 π -periodic, then
1.
C ^ n ( F ) 2 T n ( F ) 2 σ 1 ( F ) for all n N , where σ 1 ( F ) : = sup ω [ 0 , 2 π ] σ 1 ( F ( ω ) ) < .
2.
{ T n ( F ) } { C ^ n ( F ) } .
Proof. 
(1) From [3] (Theorem 4.3) T n ( F ) 2 σ 1 ( F ) for all n N . Consequently, to prove Assertion (1), we only need to show that C ^ n ( F ) 2 T n ( F ) 2 for all n N . Fix n N . As the spectral norm is unitarily invariant (see, e.g., [3] (Section 2.1)) and V n I M and ( V n I N ) * are unitary matrices, we have
C ^ n ( F ) 2 = diag 1 k n [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k 2 = max 1 k n { [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k 2 } .
Consider k 0 { 1 , , n } and x 0 C N × 1 satisfying
C ^ n ( F ) 2 = [ ( V n I M ) * T n ( F ) ( V n I N ) ] k 0 , k 0 2 = [ ( V n I M ) * T n ( F ) ( V n I N ) ] k 0 , k 0 x 0 2 x 0 2 .
Let x ^ 0 C n N × 1 with [ x ^ 0 ] j , 1 = δ j , k 0 x 0 for all j { 1 , , n } , then
0 C ^ n ( F ) 2 ( V n I M ) * T n ( F ) ( V n I N ) x ^ 0 2 x 0 2 = ( V n I M ) * T n ( F ) ( V n I N ) x ^ 0 2 x ^ 0 2 ( V n I M ) * T n ( F ) ( V n I N ) 2 = T n ( F ) 2 .
(2) From Lemma 1, ( V n I M ) * C n ( F ) ( V n I N ) is an n × n block diagonal matrix with M × N blocks:
( V n I M ) * C n ( F ) ( V n I N ) = diag 1 k n [ ( V n I M ) * C n ( F ) ( V n I N ) ] k , k .
Therefore,
C n ( F ) = ( V n I M ) diag 1 k n [ ( V n I M ) * C n ( F ) ( V n I N ) ] k , k ( V n I N ) * .
Hence, since the Frobenius norm is unitarily invariant (see, e.g., [3] (Section 2.1)), one obtains
C n ( F ) C ^ n ( F ) F   =   ( V n I M ) diag 1 k n [ ( V n I M ) * C n ( F ) ( V n I N ) ] k , k     diag 1 k n [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k ( V n I N ) * F   =   diag 1 k n [ ( V n I M ) * C n ( F ) ( V n I N ) ] k , k diag 1 k n [ ( V n I M ) * T n ( F ) ( V n I N ) ] k , k F   =   diag 1 k n [ ( V n I M ) * ( C n ( F ) T n ( F ) ) ( V n I N ) ] k , k F     ( V n I M ) * ( C n ( F ) T n ( F ) ) ( V n I N ) F = C n ( F ) T n ( F ) F = T n ( F ) C n ( F ) F .
Thus, from Lemma 2, we conclude that
0 T n ( F ) C ^ n ( F ) F n T n ( F ) C n ( F ) F + C n ( F ) C ^ n ( F ) F n 2 T n ( F ) C n ( F ) F n 0 .
 ☐
We now show that C ^ n ( F ) keeps several properties of T n ( F ) for the case in which M = N .
Lemma 5.
Consider n N and let F : R C N × N be continuous and 2 π -periodic.
1.
If T n ( F ) is Hermitian, then C ^ n ( F ) is Hermitian.
2.
If T n ( F ) is real and symmetric, then C ^ n ( F ) is real and symmetric.
3.
If T n ( F ) is positive semidefinite, then C ^ n ( F ) is positive semidefinite.
4.
If T n ( F ) is positive definite, then C ^ n ( F ) is positive definite.
Proof. 
(1) As T n ( F ) is Hermitian, ( V n I N ) * T n ( F ) ( V n I N ) is Hermitian. Therefore, [ ( V n I N ) * T n ( F ) ( V n I N ) ] k , k is Hermitian for all k { 1 , , n } , and, consequently, diag 1 k n [ ( V n I N ) * T n ( F ) ( V n I N ) ] k , k is also Hermitian. Hence, C ^ n ( F ) = ( V n I N ) diag 1 k n [ ( V n I N ) * T n ( F ) ( V n I N ) ] k , k ( V n I N ) * is Hermitian.
(2) If T n ( F ) is real and symmetric, then T n ( F ) is Hermitian. Applying Lemma 3 and Assertion (1), C ^ n ( F ) is real and Hermitian, and hence, C ^ n ( F ) is real and symmetric.
(3) For every k { 1 , , n } , let u k C n N × N with [ u k ] j , 1 = δ j , k I N for all j { 1 , , n } . If x C n N × 1 , then
x * C ^ n ( F ) x = y * diag 1 k n [ ( V n I N ) * T n ( F ) ( V n I N ) ] k , k y = y * diag 1 k n u k * ( V n I N ) * T n ( F ) ( V n I N ) u k y = k = 1 n [ y ] k , 1 * u k * ( V n I N ) * T n ( F ) ( V n I N ) u k [ y ] k , 1 = k = 1 n z k * T n ( F ) z k ,
where y = ( V n I N ) * x and z k = ( V n I N ) u k [ y ] k , 1 for all k { 1 , , n } . Since T n ( F ) is positive semidefinite, z k * T n ( F ) z k 0 for all k { 1 , , n } , and, thus, x * C ^ n ( F ) x 0 .
(4) Consider x C n N × 1 . Since T n ( F ) is positive definite, it is positive semidefinite, and from Assertion (3), we have x * C ^ n ( F ) x 0 . Suppose that x * C ^ n ( F ) x = 0 . As T n ( F ) is positive definite, applying Assertion (2) yields z k = 0 n N × 1 for all k { 1 , , n } . Consequently, u k [ y ] k , 1 = 0 n N × 1 for all k { 1 , , n } , and, therefore, [ y ] k , 1 = 0 N × 1 for all k { 1 , , n } . Hence, y = 0 n N × 1 , and thus, x = 0 n N × 1 . ☐
Finally, we show that Assertion (1) is also true if T n ( F ) is replaced by C ^ n ( F ) .
Lemma 6.
Let F : R C N × N be continuous and 2 π -periodic. If F ( ω ) is Hermitian for all ω R , then
inf ( F ) λ n N ( C ^ n ( F ) ) λ 1 ( C ^ n ( F ) ) sup ( F ) n N .
Proof. 
Applying Lemma 5 C ^ n ( F ) is Hermitian for all n N . Fix n N and x C n N × 1 . From Assertion (2), we have
x * C ^ n ( F ) x = k = 1 n z k * T n ( F ) z k .
Suppose that x is an eigenvector of C ^ n ( F ) with x 2 = 1 . Consequently,
λ j ( C ^ n ( F ) ) = λ j ( C ^ n ( F ) ) x * x = x * λ j ( C ^ n ( F ) ) x = x * C ^ n ( F ) x = k = 1 n z k * T n ( F ) z k
for some j { 1 , , n N } . Let F ( ω ) = U ( ω ) diag ( λ 1 ( F ( ω ) ) , , λ N ( F ( ω ) ) ) ( U ( ω ) ) 1 be a unitary diagonalization (i.e., an eigenvalue decomposition where the eigenvector matrix U ( ω ) is unitary) of F ( ω ) for all ω [ 0 , 2 π ] . Then,
λ j ( C ^ n ( F ) ) = k = 1 n [ z k * T n ( F ) z k ] 1 , 1 = k = 1 n h = 1 n [ z k * ] 1 , h [ T n ( F ) z k ] h , 1 = k = 1 n h = 1 n [ z k * ] 1 , h l = 1 n [ T n ( F ) ] h , l [ z k ] l , 1 = k = 1 n h = 1 n [ z k ] h , 1 * l = 1 n 1 2 π 0 2 π e ( h l ) ω i F ( ω ) d ω [ z k ] l , 1 = k = 1 n 1 2 π 0 2 π h = 1 n [ z k ] h , 1 * l = 1 n e ( h + l ) ω i F ( ω ) [ z k ] l , 1 d ω = k = 1 n 1 2 π 0 2 π h = 1 n e h ω i [ z k ] h , 1 * F ( ω ) l = 1 n e l ω i [ z k ] l , 1 d ω = k = 1 n 1 2 π 0 2 π ( U ( ω ) ) * h = 1 n e h ω i [ z k ] h , 1 * diag ( λ 1 ( F ( ω ) ) , , λ N ( F ( ω ) ) ) ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 d ω = k = 1 n 1 2 π 0 2 π s = 1 N ( U ( ω ) ) * h = 1 n e h ω i [ z k ] h , 1 * 1 , s λ s ( F ( ω ) ) ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 s , 1 d ω = k = 1 n 1 2 π 0 2 π s = 1 N λ s ( F ( ω ) ) ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 s , 1 2 d ω .
Therefore,
inf ( F ) k = 1 n 1 2 π 0 2 π s = 1 N ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 s , 1 2 d ω λ j ( C ^ n ( F ) ) sup ( F ) k = 1 n 1 2 π 0 2 π s = 1 N ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 s , 1 2 d ω .
Since
1 2 π 0 2 π e m ω i d ω = 1 if   m = 0 , 0 if   m Z \ { 0 } ,
we obtain
k = 1 n 1 2 π 0 2 π s = 1 N ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 s , 1 2 d ω = k = 1 n 1 2 π 0 2 π ( U ( ω ) ) * h = 1 n e h ω i [ z k ] h , 1 * ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 d ω = k = 1 n 1 2 π 0 2 π h = 1 n e h ω i [ z k ] h , 1 * U ( ω ) ( U ( ω ) ) * l = 1 n e l ω i [ z k ] l , 1 d ω = k = 1 n h = 1 n [ z k * ] 1 , h l = 1 n 1 2 π 0 2 π e ( h + l ) ω i d ω [ z k ] l , 1 = k = 1 n h = 1 n [ z k * ] 1 , h [ z k ] h , 1 = k = 1 n z k * z k = k = 1 n [ y ] k , 1 * u k * ( V n I N ) * ( V n I N ) u k [ y ] k , 1 = k = 1 n [ y * ] 1 , k u k * u k [ y ] k , 1 = k = 1 n [ y * ] 1 , k [ y ] k , 1 = y * y = x * ( V n I N ) ( V n I N ) * x = x * x = 1 ,
which completes the proof. ☐

4. Coding WSS Vector Processes by Using the Sequence of Block Circulant Matrices Considered

Consider that { x n : n N } is a real zero-mean Gaussian N-dimensional vector process. From [8], we know that the RDF of the real zero-mean Gaussian vector x n : 1 is given by
R n ( D ) = 1 n N k = 1 n N max 0 , 1 2 ln λ k E x n : 1 x n : 1 θ n , n N ,
where θ n is a real number satisfying
D = 1 n N k = 1 n N min θ n , λ k E x n : 1 x n : 1 .
We recall that R n ( D ) represents the lowest possible required rate (measured in nats) for jointly encoding (compressing) n symbols from N sources with mean square error (MSE) distortion D.
We assume that { x n : n N } is WSS with continuous PSD X and inf ( X ) > 0 . Fix D ( 0 , inf ( X ) ] , and from Assertion (1), we obtain that θ n = D , and, consequently,
R n ( D ) = 1 n N k = 1 n N max 0 , 1 2 ln λ k E x n : 1 x n : 1 D = 1 n N k = 1 n N 1 2 ln λ k E x n : 1 x n : 1 D = 1 2 n N k = 1 n N ln λ k T n ( X ) D n N .
For every n N , let us compute the discrete Fourier transform (DFT) of the N sources as
y n : 1 = V n * I N x n : 1 = V n I N * x n : 1 .
The correlation matrix of y n : 1 is given by
E y n : 1 y n : 1 * = E V n I N * x n : 1 x n : 1 * V n I N = V n I N * E x n : 1 x n : 1 * V n I N = V n I N * E x n : 1 x n : 1 V n I N = V n I N * T n ( X ) V n I N .
It is clear from Assertion (3) that the lowest possible required rate for encoding y n : 1 with MSE distortion D is the same as for x n : 1 , i.e., R n ( D ) .
Let us compress the Gaussian vector y n : 1 as if the samples y k 1 and y k 2 were uncorrelated for all k 1 , k 2 { 1 , , n } with k 1 k 2 , or, equivalently, as if its correlation matrix were diag 1 k n [ ( V n I N ) * T n ( X ) ( V n I N ) ] k , k instead of ( V n I N ) * T n ( X ) ( V n I N ) . In this case, from Lemma 6, we obtain that the corresponding rate is given by
R ^ n ( D ) = 1 2 n N k = 1 n N ln λ k ( C ^ n ( X ) ) D n N .
We now study the rate loss R ^ n ( D ) R n ( D ) . We have
0 R ^ n ( D ) R n ( D ) = 1 2 n N k = 1 n N ln λ k ( C ^ n ( X ) ) λ k T n ( X ) = 1 2 n N ln k = 1 n N λ k ( C ^ n ( X ) ) λ k T n ( X ) = 1 2 n N ln det ( C ^ n ( X ) ) det T n ( X ) = 1 2 n N ln det ( C ^ n ( X ) ( T n ( X ) ) 1 ) = 1 2 n N ln k = 1 n N λ k ( C ^ n ( X ) ( T n ( X ) ) 1 ) 1 2 n N ln 1 n N k = 1 n N λ k ( C ^ n ( X ) ( T n ( X ) ) 1 ) n N = 1 2 ln 1 n N k = 1 n N λ k ( C ^ n ( X ) ( T n ( X ) ) 1 ) = 1 2 ln 1 n N tr ( C ^ n ( X ) ( T n ( X ) ) 1 ) 1 2 ln n N n N C ^ n ( X ) ( T n ( X ) ) 1 F = 1 2 ln 1 n N ( C ^ n ( X ) T n ( X ) ) ( T n ( X ) ) 1 + I n N F 1 2 ln 1 n N n N + ( C ^ n ( X ) T n ( X ) ) ( T n ( X ) ) 1 F 1 2 ln 1 n N n N + C ^ n ( X ) T n ( X ) F ( T n ( X ) ) 1 2 = 1 2 ln 1 + C ^ n ( X ) T n ( X ) F n N ( T n ( X ) ) 1 2 .
Applying Assertion (1) yields
0 R ^ n ( D ) R n ( D ) 1 2 ln 1 + 1 N inf ( X ) C ^ n ( X ) T n ( X ) F n ,
and hence, from Lemma 4, we conclude that
lim n R ^ n ( D ) R n ( D ) = 0 .
The obtained result (4) on the RDF of WSS vector processes shows that there is no rate loss for large enough n if we consider the correlation matrix of y n : 1 as if it were a block diagonal matrix. Obviously, encoding all the samples y 1 , , y n separately involves a notably lower computational complexity than encoding them jointly. The complexity of coding a WSS vector process in this way is O ( n log n ) if the fast Fourier transform (FFT) algorithm is used.

5. Filtering WSS Vector Processes by Using the Sequence of Block Circulant Matrices Considered

Consider a zero-mean WSS M-dimensional vector process { x n : n N } with continuous power spectral density (PSD) X. Let { y n : n N } be a zero-mean WSS N-dimensional vector process with continuous PSD Y. Assume that those two processes are jointly WSS with continuous joint PSD Z.
For every n N , if x ^ n : 1 is an estimation of x n : 1 from y n : 1 of the form
x ^ n : 1 = W y n : 1
with W C n M × n N , the MSE per sample is given by
1 n MSE ( W ) : = 1 n E x n : 1 x ^ n : 1 2 2 = 1 n E x n : 1 x ^ n : 1 * x n : 1 x ^ n : 1 = 1 n tr E x n : 1 x ^ n : 1 x n : 1 x ^ n : 1 * = 1 n tr E x n : 1 x n : 1 * E x n : 1 y n : 1 * W * W E y n : 1 x n : 1 * + W E y n : 1 y n : 1 * W * = 1 n tr T n ( X ) T n ( Z ) W * W T n ( Z ) * + W T n ( Y ) W * .
The minimum MSE (MMSE) is given by MMSE = MSE ( W 0 ) , where
W 0 = E x n : 1 y n : 1 * ( E y n : 1 y n : 1 * ) 1 = T n ( Z ) ( T n ( Y ) ) 1
whenever det ( T n ( Y ) ) 0 (or, equivalently, whenever T n ( Y ) is positive definite). The filter W 0 is known as the Wiener filter.
Consider the following filter:
W C : = C ^ n ( Z ) ( C ^ n ( Y ) ) 1 = ( V n I M ) diag 1 k n [ ( V n I M ) * T n ( Z ) ( V n I N ) ] k , k [ ( V n I N ) * T n ( Y ) ( V n I N ) ] k , k 1 ( V n I N ) * .
The filter W C is well defined because, from Lemma 5, C ^ n ( Y ) is positive definite, and, therefore, it is invertible.
We now study the effect on the MSE when the optimal filter W 0 is substituted by W C . We have
0 MSE ( W C ) n MMSE n = MSE ( W C ) n MMSE n = 1 n tr ( T n ( X ) T n ( Z ) W C * W C ( T n ( Z ) ) * + W C ( T n ( Y ) ) W C * ) 1 n tr ( T n ( X ) T n ( Z ) ( T n ( Y ) ) 1 ( T n ( Z ) ) * ) = 1 n tr ( T n ( Z ) W C * W C ( T n ( Z ) ) * + W C ( T n ( Y ) ) W C * + T n ( Z ) ( T n ( Y ) ) 1 ( T n ( Z ) ) * ) = 1 n | tr ( ( T n ( Z ) ( T n ( Y ) ) 1 W C ) ( ( T n ( Z ) ) * T n ( Y ) W C * ) ) | n M n ( T n ( Z ) ( T n ( Y ) ) 1 W C ) ( ( T n ( Z ) ) * T n ( Y ) W C * ) F M n T n ( Z ) ( T n ( Y ) ) 1 W C F ( T n ( Z ) ) * T n ( Y ) W C * 2 M n T n ( Z ) ( T n ( Y ) ) 1 C ^ n ( Z ) ( C ^ n ( Y ) ) 1 F ( T n ( Z ) ) * T n ( Y ) ( C ^ n ( Y ) ) 1 ( C ^ n ( Z ) ) * 2 .
Since
T n ( Z ) ( T n ( Y ) ) 1 C ^ n ( Z ) ( C ^ n ( Y ) ) 1 F   =   T n ( Z ) ( T n ( Y ) ) 1 C ^ n ( Z ) ( T n ( Y ) ) 1 + C ^ n ( Z ) ( T n ( Y ) ) 1 C ^ n ( Z ) ( C ^ n ( Y ) ) 1 F     ( T n ( Z ) C ^ n ( Z ) ) ( T n ( Y ) ) 1 F + C ^ n ( Z ) ( ( T n ( Y ) ) 1 ( C ^ n ( Y ) ) 1 ) F     T n ( Z ) C ^ n ( Z ) F ( T n ( Y ) ) 1 2 + C ^ n ( Z ) 2 ( T n ( Y ) ) 1 ( C ^ n ( Y ) ) 1 F     T n ( Z ) C ^ n ( Z ) F ( T n ( Y ) ) 1 2     + C ^ n ( Z ) 2 ( C ^ n ( Y ) ) 1 C ^ n ( Y ) ( T n ( Y ) ) 1 ( C ^ n ( Y ) ) 1 T n ( Y ) ( T n ( Y ) ) 1 F     T n ( Z ) C ^ n ( Z ) F ( T n ( Y ) ) 1 2 + C ^ n ( Z ) 2 ( C ^ n ( Y ) ) 1 ( C ^ n ( Y ) T n ( Y ) ) ( T n ( Y ) ) 1 F     T n ( Z ) C ^ n ( Z ) F ( T n ( Y ) ) 1 2 + C ^ n ( Z ) 2 ( C ^ n ( Y ) ) 1 2 ( C ^ n ( Y ) T n ( Y ) ) ( T n ( Y ) ) 1 F     T n ( Z ) C ^ n ( Z ) F ( T n ( Y ) ) 1 2 + C ^ n ( Z ) 2 ( C ^ n ( Y ) ) 1 2 C ^ n ( Y ) T n ( Y ) F ( T n ( Y ) ) 1 2
and
( T n ( Z ) ) * T n ( Y ) ( C ^ n ( Y ) ) 1 ( C ^ n ( Z ) ) * 2 ( T n ( Z ) ) * 2 + T n ( Y ) ( C ^ n ( Y ) ) 1 ( C ^ n ( Z ) ) * 2 T n ( Z ) 2 + T n ( Y ) ( C ^ n ( Y ) ) 1 2 ( C ^ n ( Z ) ) * 2 T n ( Z ) 2 + T n ( Y ) 2 ( C ^ n ( Y ) ) 1 2 C ^ n ( Z ) 2 ,
we obtain
0 MSE ( W C ) n MMSE n M ( T n ( Y ) ) 1 2 ( T n ( Z ) 2 + T n ( Y ) 2 ( C ^ n ( Y ) ) 1 2 C ^ n ( Z ) 2 ) C ^ n ( Z ) T n ( Z ) F n + C ^ n ( Z ) 2 ( C ^ n ( Y ) ) 1 2 C ^ n ( Y ) T n ( Y ) F n .
Consequently, applying Assertion (1) and Lemma 4 yields
0 MSE ( W C ) n MMSE n M ( T n ( Y ) ) 1 2 σ 1 ( Z ) ( 1 + sup ( Y ) ( C ^ n ( Y ) ) 1 2 ) C ^ n ( Z ) T n ( Z ) F n + σ 1 ( Z ) ( C ^ n ( Y ) ) 1 2 C ^ n ( Y ) T n ( Y ) F n .
If we assume that inf ( Y ) > 0 , from Assertion (1), we obtain that T n ( Y ) is positive definite for all n N . Moreover, applying Assertion (1) and Lemma 6 yields
0 MSE ( W C ) n MMSE n M σ 1 ( Z ) inf ( Y ) 1 + sup ( Y ) inf ( Y ) C ^ n ( Z ) T n ( Z ) F n + σ 1 ( Z ) inf ( Y ) C ^ n ( Y ) T n ( Y ) F n ,
and, therefore, from Lemma 4, we conclude that
lim n MSE ( W C ) n MMSE n = 0 .
The obtained result (6) shows that there is no difference in the MSE for large enough n if we substitute the optimal filter W 0 by W C . Obviously, the computational complexity of the operation (5) is notably reduced when applying this substitution and the FFT algorithm is used. Specifically, the complexity is reduced from O ( n 2 ) to O ( n log n ) .

6. Conclusions

In this paper, we present a sequence of block circulant matrices and we apply it to reduce the complexity of coding and filtering WSS vector processes. Specifically, in both applications, the complexity is reduced from O ( n 2 ) to O ( n log n ) , which is the complexity of performing an FFT.

Acknowledgments

This work was supported in part by the Spanish Ministry of Economy and Competitiveness through the projects RACHEL (TEC2013-47141-C4-2-R) and CARMEN (TEC2016-75067-C4-3-R).

Author Contributions

Authors are listed in order of their degree of involvement in the work, with the most active contributors listed first. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gray, R.M. Toeplitz and circulant matrices: A review. Found. Trends Commun. Inf. Theory 2006, 2, 155–239. [Google Scholar] [CrossRef]
  2. Pearl, J. On coding and filtering stationary signals by discrete Fourier transforms. IEEE Trans. Inf. Theory 1973, 19, 229–232. [Google Scholar] [CrossRef]
  3. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Block Toeplitz matrices: Asymptotic results and applications. Found. Trends Commun. Inf. Theory 2011, 8, 179–257. [Google Scholar] [CrossRef]
  4. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Asymptotically equivalent sequences of matrices and Hermitian block Toeplitz matrices with continuous symbols: Applications to MIMO systems. IEEE Trans. Inf. Theory 2008, 54, 5671–5680. [Google Scholar] [CrossRef]
  5. Bhatia, R. Matrix Analysis; Springer: Berlin/Heidelberg, Germany, 1997. [Google Scholar]
  6. Gutiérrez-Gutiérrez, J.; Crespo, P.M. Asymptotically equivalent sequences of matrices and multivariate ARMA processes. IEEE Trans. Inf. Theory 2011, 57, 5444–5454. [Google Scholar] [CrossRef]
  7. Gray, R.M. On the asymptotic eigenvalue distribution of Toeplitz matrices. IEEE Trans. Inf. Theory 1972, 18, 725–730. [Google Scholar] [CrossRef]
  8. Kolmogorov, A.N. On the Shannon theory of information transmission in the case of continuous signals. IRE Trans. Inf. Theory 1956, 2, 102–108. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Gutiérrez-Gutiérrez, J.; Zárraga-Rodríguez, M.; Insausti, X.; Hogstad, B.O. On the Complexity Reduction of Coding WSS Vector Processes by Using a Sequence of Block Circulant Matrices. Entropy 2017, 19, 95. https://doi.org/10.3390/e19030095

AMA Style

Gutiérrez-Gutiérrez J, Zárraga-Rodríguez M, Insausti X, Hogstad BO. On the Complexity Reduction of Coding WSS Vector Processes by Using a Sequence of Block Circulant Matrices. Entropy. 2017; 19(3):95. https://doi.org/10.3390/e19030095

Chicago/Turabian Style

Gutiérrez-Gutiérrez, Jesús, Marta Zárraga-Rodríguez, Xabier Insausti, and Bjørn O. Hogstad. 2017. "On the Complexity Reduction of Coding WSS Vector Processes by Using a Sequence of Block Circulant Matrices" Entropy 19, no. 3: 95. https://doi.org/10.3390/e19030095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop