Next Article in Journal
A Finite-Time Thermal Cycle Variational Optimization with a Stefan–Boltzmann Law for Three Different Criteria
Previous Article in Journal
Reliability Analysis of Free Jet Scour Below Dams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Capacity Bounds and Mapping Design for Binary Symmetric Relay Channels

School of Electrical Engineering and ACCESS Linnaeus Center, Royal Institute of Technology (KTH), Stockholm, 100 44, Sweden
*
Authors to whom correspondence should be addressed.
Entropy 2012, 14(12), 2589-2610; https://doi.org/10.3390/e14122589
Submission received: 1 September 2012 / Revised: 4 December 2012 / Accepted: 6 December 2012 / Published: 17 December 2012
(This article belongs to the Special Issue Information Theory Applied to Animal Communication)

Abstract

:
Capacity bounds for a three-node binary symmetric relay channel with orthogonal components at the destination are studied. The cut-set upper bound and the rates achievable using decode-and-forward (DF), partial DF and compress-and-forward (CF) relaying are first evaluated. Then relaying strategies with finite memory-length are considered. An efficient algorithm for optimizing the relay functions is presented. The Boolean Fourier transform is then employed to unveil the structure of the optimized mappings. Interestingly, the optimized relay functions exhibit a simple structure. Numerical results illustrate that the rates achieved using the optimized low-dimensional functions are either comparable to those achieved by CF or superior to those achieved by DF relaying. In particular, the optimized low-dimensional relaying scheme can improve on DF relaying when the quality of the source-relay link is worse than or comparable to that of other links.

1. Introduction

In this paper we consider a relay network consisting of a source, a relay, and a destination, as illustrated in Figure 1. This channel is considered in [1,2]. The communication task is to reproduce the transmitted message M, uniformly chosen from the set M = { 1 , 2 , , 2 n R } , at the destination such that the probability of error is arbitrarily small. The transmission of a message consumes n channel uses. We desire to quantify the supremum of the set of rates, R, for which the average message error probability at the destination can be made to approach zero as the number of channel uses n goes to infinity. This number is the capacity, C, in the communication between the source and destination. The channel capacity for the general relay channel is still an open problem.
Figure 1. Three-node relay network.
Figure 1. Three-node relay network.
Entropy 14 02589 g001
One special case of the general relay channel is the Gaussian relay channel, which is studied, e.g., in [2,3,4,5,6,7,8,9]. In this paper, however, we focus on a relay network in which each link is a binary symmetric channel. Some instances of the binary symmetric relay channel (BSRC) are considered in [10,11,12,13,14]. The work of [10,11] considers BSRCs with correlated noises and [12,13] focuses on the detect-and-forward protocol for multi-hop relaying and the relay channel, respectively. In this paper, we consider a model that is also studied in an independent work [14]. We study decode-and-forward (DF), partial decode-and-forward (PDF), compress-and-forward (CF) and general finite-memory relay mappings. One of our main contributions is to propose a general structure for finite dimensional relaying. We show that one can improve on the result presented in [14] and illustrate that it is possible to approach the capacity upper bound for some cases by using the proposed low-dimensional mappings. Interestingly we recover the lower bound in [14] by a simple one-dimensional mapping at the relay. We also illustrate that one can obtain higher reliable rates by employing optimized relaying functions with larger memories. The rate obtained via optimized finite-length mappings can be superior to those achieved by DF protocol in some cases and are comparable to those achieved by CF protocol in general.
The remaining part of the paper is organized as follows. In Section 2, we present the three-node BSRC with orthogonal components at the destination. Section 3 derives capacity bounds for the BSRC when the relay is assumed to have infinite memory. Section 4 investigates capacity bounds for the BSRC when the relay is assumed to have a finite memory. We present an algorithm to optimize the relay mapping in Section 4. In Section 5 we compare the achievable rates as a function of memory length. Finally, Section 6 concludes the paper.
Notation: We use the following notation for brevity.
  • GF b denotes the binary Galois filed, i.e., { 0 , 1 } .
  • GF b k denotes the k-dimensional binary Galois filed, i.e., { 0 , 1 } k .
  • H b ( ϵ ) : = - ϵ log ( ϵ ) - ( 1 - ϵ ) log ( 1 - ϵ ) denotes binary entropy function where ϵ [ 0 , 1 ] .
  • D H ( x 1 k , y 1 k ) denotes the Hamming distance between the two binary sequences of length k.
  • The operation * is defined as ϵ 1 * ϵ 2 : = ϵ 1 + ϵ 2 - 2 ϵ 1 ϵ 2 .
  • We denote that the binary random variable Z has a Bernoulli distribution by Z Ber ( ϵ ) where Pr ( Z = 1 ) = ϵ and Pr ( Z = 0 ) = 1 - ϵ .
  • We denote the binary Kronecker delta function by δ b ( x ) , where δ b ( 0 ) = 1 and δ b ( 1 ) = 0 for x GF b .

2. Binary Symmetric Relay Channel

In this section, we introduce the channel model that we consider in the paper. Figure 2 shows a BSRC consisting of three nodes: a source, a relay and a destination. In this model, we however assume that received signals at the destination are orthogonal. That is, the signals transmitted from the source and the relay do not interfere with each other. We further assume that all links are corrupted with modulo-sum noises distributed according to the Bernoulli distribution and all quantities are binary; i.e., X , X r , Y 1 , Y 2 , Y r { 0 , 1 } .
The received signal at the relay Y r is given by
Y r = X Z r
where X is the transmitted symbol from the source and Z r Ber ( ϵ r ) is the additive Bernoulli noise. (The Bernoulli noise models for example the conventional communication set-up using BPSK modulation followed by a hard decision (see also [14]).) The received signal from the source at the destination Y 1 is given by
Y 1 = X Z 1
where Z 1 Ber ( ϵ 1 ) is the additive Bernoulli noise. Similarly, the received signal from the relay at the destination Y 2 is given by
Y 2 = X r Z 2
where X r is the transmitted symbol from the relay and Z 2 Ber ( ϵ 2 ) is the additive Bernoulli noise. We assume that the random variables Z r , Z 1 , and Z 2 are mutually independent. Note that the addition in Equations (1)–(3) is done in GF ( 2 ) .
Remark 1. Figure 3 shows a BSRC with non-orthogonal reception at the destination. The received signal at the destination is given by
Y = X X r Z
where X and X r respectively denote the symbols transmitted by the source and the relay. The random variable Z Ber ( ϵ ) is the additive Bernoulli noise and is independent of Z r . The capacity of this channel is
C n o = 1 - H b ( ϵ )
By setting X r = 0 , we have Y = X Z . Then the achievability follows since max p ( x ) I ( X ; Y | X r = 0 ) = C n o . The converse follows from the cut-set bound. Invoking the multiple access bound, we have
C I ( X , X r ; Y ) = H ( Y ) - H ( Y | X , X r ) = H ( Y ) - H ( Z ) 1 - H b ( ϵ )
In the sequel, we present various capacity bounds for the orthogonal BSRC.
Figure 2. Orthogonal Binary Symmetric Relay Channel.
Figure 2. Orthogonal Binary Symmetric Relay Channel.
Entropy 14 02589 g002
Figure 3. Non-Orthogonal Binary Symmetric Relay Channel.
Figure 3. Non-Orthogonal Binary Symmetric Relay Channel.
Entropy 14 02589 g003

3. Capacity Bounds for the Orthogonal BSRC: Infinite Memory Relay Case

In this section, we consider the cut-set upper bound and three lower bounds on the capacity: DF, PDF and CF. These three bounds are evaluated based on the results in [2] for the general relay channel where the relay has infinite memory and unlimited processing capability.
Proposition 1 (cutset bound). For the relay channel in Figure 2, the capacity is upper bounded by
R U B = min 1 + H b ( ϵ 1 * ϵ r ) - H b ( ϵ 1 ) - H b ( ϵ r ) , 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) .
Proof. See Appendix A and also [14].  ☐
Proposition 2 (DF lower bound). For the relay channel in Figure 2, the capacity is lower bounded by
R D F = max { 1 - H b ( ϵ 1 ) , min { 1 - H b ( ϵ r ) , 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) } }
where the bound is achieved using the block Markov encoding scheme in [2] (known as DF relaying) and the only-direct-transmission (i.e., when the relay is off).
Proof. See Appendix B. ☐
Note that in Equation (8), we take the maximum of the rates achieved using the conventional DF [2] and the only-direct-link transmission. In some cases it is possible to improve on DF by using PDF at the relay. That is, the relay only decodes a part of the transmitted message. The achievable rate of PDF is given by
R P D F = max p ( u , x , x r ) min { I ( X , X r ; Y 1 , Y 2 ) , I ( U ; Y r | X r ) + I ( X ; Y 1 , Y 2 | X r , U ) }
where U denotes the part of the transmitted message that the relay decodes. (See Theorem 7 in [2] and also [15].)
Proposition 3 (PDF lower bound). For the relay channel in Figure 2, generalized block Markov encoding attains the same rate as the modified DF given in Equation (8).
Proof. See Appendix C.  ☐
Corollary 1. DF is an optimal relaying strategy if H b ( ϵ 1 ) + H b ( ϵ 2 ) - H b ( ϵ r ) 1 .
Proof. The proof follows from Propositions 2 and 1.  ☐
Proposition 4 (CF lower bound). For the relay channel in Figure 2, the capacity is lower bounded by
R C F = 1 + H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ r * ϵ q ) - H b ( ϵ 1 ) , if 1 < H b ( ϵ 2 ) + H b ( ϵ 1 * ϵ r ) 1 + H b ( ϵ 1 * ϵ r ) - H b ( ϵ r ) - H b ( ϵ 1 ) , if 1 H b ( ϵ 2 ) + H b ( ϵ 1 * ϵ r )
where ϵ q satisfies
H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ q ) + H b ( ϵ 2 ) = 1
Proof. See Appendix D.  ☐
The bound in Equation (10) is achieved using the side-information encoding scheme in [2], known as CF relaying.
Corollary 2. CF is an optimal relaying strategy if H b ( ϵ 1 * ϵ r ) + H b ( ϵ 2 ) 1 .
Proof. The proof follows from Propositions 4 and 1.  ☐
Figure 4. Optimal regions for ϵ 2 = 0.05 .
Figure 4. Optimal regions for ϵ 2 = 0.05 .
Entropy 14 02589 g004
Figure 4 and Figure 5 show optimal regions in the plane [0, 1]×[0, 1] for ϵ2 set to 0.05 and 0.2, respectively. One can see that for some parameters of ϵr and ϵ1, CF or DF is optimal.
From Figure 4 and Figure 5, we see that DF is optimal when ϵr is small and ϵ1 is close to 0.5. When ϵ1 = 0.5, the BSRC simplifies to a two-hop channel for which DF is capacity achieving. However, DF is also optimal when the relay is relatively strong compared with the direct link.
From Figure 4 and Figure 5, we similarly see that CF is optimal at the corner of the plane, i.e., when ϵ1 and ϵr are small. As ϵ2 increases, the area of the region for which CF is optimal shrinks. This is due to the fact that small value of ϵ2 allows the relay to use higher rate to describe the received signal to destination. Small value of ϵ1 means the destination has better side-information in order to decode the reproduction sequence generated at the relay, and small value of ϵr means that the relay receives a less-noisy sequence on average that makes compression of the “true” signal easier.
Figure 5. Optimal regions for ϵ 2 = 0.2 .
Figure 5. Optimal regions for ϵ 2 = 0.2 .
Entropy 14 02589 g005

4. Capacity Bounds for the Orthogonal BSRC: Finite Memory Relay Case

We next consider the case when the relay has a finite memory length. For the Gaussian relay channel, optimized memoryless relaying is investigated in [16,17,18,19]. Here we consider higher-dimension mappings for the BSRC.
If the relay has a storage memory of k bits, it can process the last k - 1 and the presently received symbol to generate k new symbols using k possibly different k-dimensional functions. This results in a low-complexity relaying protocol suitable for delay-sensitive or inexpensive applications. In the following, we denote the relay functions by
f i : GF b k GF b x r , i = f i ( y r , 1 , , y r , k )
for i { 1 , 2 , , k } . Note that we here consider the whole sub-block of k symbols to generate k new symbols to be transmitted to the destination. This is different from the classical definition with strictly causal relaying. We are allowed to do this, without any particular condition in signal reception at the relay, since the relay has an orthogonal link to the destination.

4.1. Achievable Rate

For a given set of relay functions { f i } i = 1 k , the channel is parameterized by the pmf p ( y | x ) , by defining y ( y 1 , 1 k , y 2 , 1 k ) and x x 1 k , where
p y | x : = p y 1 , 1 k , y 2 , 1 k | x 1 k = y r , 1 k GF b k p y 1 , 1 k , y 2 , 1 k | y r , 1 k , x 1 k p y r , 1 k | x 1 k = y r , 1 k GF b k p y 1 , 1 k | x 1 k p y 2 , 1 k | y r , 1 k p y r , 1 k | x 1 k = y r , 1 k GF b k i = 1 k p y 1 , i | x i p y r , i | x i p ( y 2 , i | f i ( y r , 1 k ) ) = y r , 1 k GF b k i = 1 k ( 1 - ϵ 1 ) δ b y 1 , i x i + ϵ 1 δ b y 1 , i x i 1 × ( 1 - ϵ r ) δ b y r , i x i + ϵ r δ b y r , i x i 1 × ( 1 - ϵ 2 ) δ b ( y 2 , i f i ( y r , 1 k ) ) + ϵ 2 δ b ( y 2 , i f i ( y r , 1 k ) 1 )
Now one can apply the standard random coding argument for the equivalent discrete memoryless point-to-point communication link with the input x and the output y whose relation is governed by the pmf p y | x as follows. Generate 2 ( n k ) C k i.i.d. codewords where each has length k n and each k subsequent symbols in every codeword is distributed according to p ( x 1 k ) (i.e., p ( x 1 k n ) = i = 1 n p ( x 1 k ) ). Thus, the achievable rate using the finite memory relay is given by
C k = sup { f i } i = 1 k , p ( x 1 k ) 1 k I ( X 1 k ; Y 1 , 1 k , Y 2 , 1 k )
where the supremum is taken over the set of Boolean functions { f i } i = 1 k and the joint pmf p ( x 1 k ) of k symbols at the source. Since the channel is used k times, the mutual information in Equation (14) is divided by k (see also [1,4]).
Achievable Rate for k = 1 : The simplest case is the memoryless relay in which the relay just transmits the received noisy bit to the destination without any further processing. That is x r , i = y r , i for 1 i n . For this relay function, the optimal input distribution is X Ber ( 1 2 ) .
Proposition 5. For the relay channel in Figure 2, the rate
C 1 = 1 + H b ( ϵ 1 * ϵ r * ϵ 2 ) - H b ( ϵ r * ϵ 2 ) - H b ( ϵ 1 )
is achievable.
Proof. See Appendix E.  ☐
We note that the rate C 1 is also derived in [14] via a suboptimal evaluation of the CF lower bound. However, here we arrive at this rate using a one-dimensional mapping without any need for a compression codebook at the relay.
Computation of C k for k 2 is a cumbersome task. Nevertheless, there is no unique set of relay functions that is optimal for all channel parameters ϵ r , ϵ 1 , ϵ 2 . To see this, consider the case with ϵ 2 = 0 . For this case, the simple strategy with k = 1 used in Proposition 5 is optimal. However, this relay function is not necessarily optimal for cases with ϵ 2 0 since one can potentially provide error protection on the relay-destination link by utilizing functions with higher dimensions.

4.2. Mapping Optimization for an Arbitrary k

In the following, we confine the pmf at the source to be p ( x 1 k ) = i = 1 k p ( x ) , and p ( x ) = 1 2 δ b ( x ) + 1 2 δ b ( x 1 ) , i.e., X Ber ( 1 2 ) .
Lemma 1. The mutual information given in Equation (14) can be written as
1 k I ( X 1 k ; Y 1 , 1 k , Y 2 , 1 k ) = 1 - H b ( ϵ 1 ) - 1 k E log ( p ( y 2 , 1 k | y 1 , 1 k ) ) + 1 k E log ( p ( y 2 , 1 k | x 1 k ) )
where
p ( y 2 , 1 k | x 1 k ) = ( 1 - ϵ 2 ) k ( 1 - ϵ r ) k y r , 1 k GF b k ϵ 2 1 - ϵ 2 D H ( y 2 , 1 k , f ( y r , 1 k ) ) ϵ r 1 - ϵ r D H ( y r , 1 k , x 1 k )
and
p ( y 2 , 1 k | y 1 , 1 k ) =
( 1 - ϵ 1 ) k ( 1 - ϵ 2 ) k ( 1 - ϵ r ) k y r 1 k GF b k ϵ 2 1 - ϵ 2 D H ( y 2 , 1 k , f ( y r , 1 k ) ) x 1 k GF b k ϵ r 1 - ϵ r D H ( y r , 1 k , x 1 k ) ϵ 1 1 - ϵ 1 D H ( y 1 , 1 k , x 1 k ) .
Proof. See Appendix F. ☐
To compute the rate in Equation (14), one needs to select the best functions among 2 k 2 k possible choices, which has a exponential complexity. (For k = 4 , there are approximately 1.8 × 10 19 possible functions.) In order to cope with the complexity, we implement an efficient hill-climbing search algorithm as follows. For a given k, we first initialize the relay functions with a random mapping and compute the rate using Lemma 1. Then we randomly select one function and one corresponding dimension, and flip the mapping and recompute the rate. If the new mapping provides a higher rate, we accept the change. Otherwise we repeat the process until the mapping converges. Since the algorithm by construction may terminate in a local optimum, we repeat the whole algorithm with different initializations and pick the mapping that attains the highest rate.
One example of the optimized mapping for k = 4 when ϵ 1 = ϵ 2 = ϵ r = 0.01 is
F = 0 0 1 1 1 1 0 0 1 1 0 0 0 0 1 1 1 0 0 1 0 1 1 0 1 0 0 1 0 1 1 0 1 0 0 1 1 0 0 1 0 1 1 0 0 1 1 0 1 0 1 0 0 1 0 1 0 1 0 1 1 0 1 0 .
Here F i j denotes the output of the relay along the ith dimension for the jth input configuration (we have 1 i 4 and 1 j 2 4 ). The relay for a given combination of the received bits finds the decimal representation and transmits the bits in the column given by the decimal representation plus one. For example, for the received string 0000, the relay transmits the bits in the first column, i.e., 0111. By studying the matrix F, we can get some insight into the underlying structure of the mapping. Finding an efficient structure at the relay simplifies the design and the implementation of relaying. We employ the Fourier transform to accomplish this task. Our use of the binary Fourier (or Hadamard) transform is related to how it was used in, e.g., [20,21], to analyze the performance of quantizers over noisy channels.

4.3. Fourier Spectrum of the Optimized Mappings

In order to define the Fourier transform, we need an orthonormal basis [22]. Consider the following set of functions
X S ( x ) : { - 1 , + 1 } k { - 1 , + 1 } X S ( x ) = i S x i
where S { 1 , 2 , , k } . Then, any function f : { 0 , 1 } k { 0 , 1 } can uniquely be represented as (We use the one-to-one mapping 0 + 1 and 1 - 1 .)
f = S { 1 , 2 , , k } f ^ ( S ) X S
where f ^ ( S ) is the Fourier coefficient of f and is given by
f ^ ( S ) = f , X S = 1 2 k x { + 1 , - 1 } k f ( x ) X S ( x ) = E f · X S
The expectation in Equation (21) is taken uniformly over x { + 1 , - 1 } k . Note that the Fourier expansion of f can potentially have up to 2 k terms. As an example consider the following randomly chosen function.
x3x2x1f
+ 1+ 1+ 1+ 1
+ 1+ 1− 1+ 1
+ 1− 1+ 1− 1
+ 1− 1− 1− 1
− 1+ 1+ 1− 1
− 1+ 1− 1+ 1
− 1− 1+ 1+ 1
− 1− 1− 1+ 1
This function can be expanded as
f ( x 1 , x 2 , x 3 ) = 1 4 - 1 4 x 1 + 1 4 x 2 - 1 4 x 3 - 1 4 x 1 x 2 + 1 4 x 1 x 3 + 3 4 x 2 x 3 + 1 4 x 1 x 2 x 3
That is, the Fourier spectrum has eight terms and the function is not linear. In general, we would like a sparse Fourier spectrum for efficient implementation. (Sparsity of mappings allows to realize them with much fewer multiplications and additions. This can be also compared to codes with low-density generator/parity matrices that allow simpler encoding and decoding in general.)
Table 1 presents the Fourier expansion of optimized relay functions for k = 2 , 3 , 4 , 6 when ϵ 1 = ϵ 2 = ϵ r = 0.01 . Interestingly, we see that that the Fourier expansion of the optimized functions is indeed sparse. Using the results in Table 1, we can rewrite the functions in the following form
f : { 0 , 1 } k { 0 , 1 } k x r = A k y r + b k
where x r = [ x r , 1 , , x r , k ] T and y r = [ y r , 1 , , y r , k ] T , and A k { 0 , 1 } [ k × k ] and b k { 0 , 1 } [ k × 1 ] . For example for k = 6 , we have
A 6 = 0 1 1 0 1 0 1 0 1 1 0 1 0 1 0 1 1 0 1 1 1 1 0 0 1 0 0 0 1 1 0 1 1 1 0 1 , b 6 = 0 1 1 1 0 1
Note that the mapping given in Equation (23) is not linear in the binary field, when b k 0 . However, the linear mapping x r = A k y r gives the same performance as x r = A k y r + b k . In other words the bias term b k does not improve the rate. This essentially follows from the data processing inequality. Therefore, the underlying relay functions define a linear code of rate one on the noisy received bits at the relay. Additionally, the code used at the relay performs joint source–channel coding, it therefore should be good for both source and channel coding.
Table 1. Fourier expansion of the optimized relay functions for the orthogonal BSRC with ϵ 1 = ϵ 2 = ϵ r = 0.01 .
Table 1. Fourier expansion of the optimized relay functions for the orthogonal BSRC with ϵ 1 = ϵ 2 = ϵ r = 0.01 .
k = 2 f 1 ( y r 1 k ) = y r 1 f 2 ( y r 1 k ) = y r 2 y r 1
k = 3 f 1 ( y r 1 k ) = y r 3 y r 2 y r 1 f 2 ( y r 1 k ) = - y r 3 y r 1 f 3 ( y r 1 k ) = y r 3 y r 2
k = 4 f 1 ( y r 1 k ) = - y r 4 y r 3 y r 1 f 2 ( y r 1 k ) = - y r 4 y r 2 y r 1 f 3 ( y r 1 k ) = - y r 3 y r 2 y r 1 f 4 ( y r 1 k ) = y r 4 y r 3 y r 2
k = 6 f 1 ( y r 1 k ) = y r 5 y r 3 y r 2 f 2 ( y r 1 k ) = - y r 6 y r 4 y r 3 y r 1 f 3 ( y r 1 k ) = - y r 5 y r 4 y r 2 f 4 ( y r 1 k ) = - y r 4 y r 3 y r 2 y r 1 f 5 ( y r 1 k ) = y r 6 y r 5 y r 1 f 6 ( y r 1 k ) = - y r 6 y r 4 y r 3 y r 2

4.4. Effect of Channel Parameters on the Structure of the Optimized Mappings

In this section, we investigate the structure of the optimized mappings for different channel parameters. Our numerical search indicates that the linear mapping x r = A k y r is an efficient strategy among all classes of mappings for low-dimensional relaying. That is, the relay employs the binary matrix A k to generate the relay outputs using k received bits.
Table 2 shows the optimized generator matrices A k for various values of ϵ r when k = 6 , ϵ 1 = ϵ 2 = 0.05 . In particular, for ϵ r = 0.25 the optimized generator matrix is the identity matrix, i.e., A 6 = I 6 . For this case, the relay is better off transmitting the received noisy bits without any further processing. However, as ϵ r decreases, the relay starts combining the received bits at the relay before transmitting. The number of ones in a row of the generator matrix indicates the number of inputs that the relay combines. The density of ones in the generator matrices ρ : = # of ones k 2 are also shown in Table 2. We see that as ϵ r decreases, ρ increases. That is, the relay starts to transmit combinations of more bits in one single channel use. This occurs because of two main reasons: firstly when ϵ r decreases the relay receives less noisy bits on average and secondly the destination has some partial knowledge of individual bits via the received signal from the source.
Table 3 shows the optimized generator matrices for various values of ϵ 1 when k = 6 , ϵ r = 0.01 and ϵ 2 = 0.1 . We similarly see that as ϵ 1 decreases, ρ increases. This is due to the fact that when ϵ 1 decreases, the destination receives better descriptions of the transmitted bits via the source-destination link. The relay then forwards combinations of several incoming bits when the destination has access to more reliable side information.
Table 2. Optimized generator matrices as a function of ϵ r for k = 6 and ϵ 1 = ϵ 2 = 0.05 .
Table 2. Optimized generator matrices as a function of ϵ r for k = 6 and ϵ 1 = ϵ 2 = 0.05 .
ϵ r 0.25 0.1 0.05 0.01 0.001 0.0001
A 6 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 1 1 0 1 0 0 0 1 0 1 1 1 1 1 1 1 0 0 0 1 0 1 1 1 1 0 0 1 0 1 0 0 0 1 0 0 1 1 1 0 1 0 0 1 0 0 0 0 0 1 1 1 1 0 0 0 1 0 1 0 1 1 0 0 0 1 1 1 0 1 0 1 1 0 0 1 0 1 0 1 1 0 1 1 0 1 1 1 1 1 0 0 1 0 1 1 1 0 0 0 1 1 0 1 1 1 1 0 1 1 0 1 0 0 0 1 1 1 0 1 1 1 0 1 0 0 1 0 1 0 1 0 1 0 1 1 1 1 0 0 1 1 1 1 1 1 0 1 1 0 0 1 1 0 0 1 1 1 1
ρ0.16670.36110.44440.55560.61110.6389
Table 3. Optimized generator matrices as a function of ϵ 1 for k = 6 and ϵ r = 0.01 , ϵ 2 = 0.1 .
Table 3. Optimized generator matrices as a function of ϵ 1 for k = 6 and ϵ r = 0.01 , ϵ 2 = 0.1 .
ϵ 1 0.4 0.25 0.1 0.05 0.01 0.001
A 6 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 1 1 0 0 0 1 0 0 1 0 0 0 0 1 0 0 1 1 0 0 0 0 1 0 1 0 0 1 0 0 0 0 1 1 0 1 1 0 0 1 1 0 1 0 1 0 1 0 1 1 0 1 0 1 0 0 1 1 1 1 0 1 0 1 0 1 0 1 1 0 0 1 1 1 0 1 0 1 0 1 1 0 0 1 0 1 0 1 1 0 1 1 1 1 1 1 0 0 1 1 0 1 1 0 1 0 1 1 0 1 0 1 1 1 1 0 1 0 1 0 1 0 1 1 1 1 0 1 1 1 0 1 1 1 0 1 0 0 1 1 1 0 1 1 1 1 1 0 0 0 1 1 1 1 1 1 1 1 0 1 0 1 0 1 1 1 1 1 1 1 0 0 1 1 1 0 1 1 1
ρ0.16670.33330.55560.63890.66670.7500
Table 4. Optimized generator matrices as a function of ϵ 2 for k = 6 and ϵ r = 0.01 , ϵ 1 = 0.1 .
Table 4. Optimized generator matrices as a function of ϵ 2 for k = 6 and ϵ r = 0.01 , ϵ 1 = 0.1 .
ϵ 2 0.4 0.25 0.05 0.01 0.001 0.0001
A 6 0 1 1 1 1 1 1 1 1 1 0 1 1 0 1 1 1 1 1 1 1 0 1 1 1 1 1 1 1 0 1 1 0 1 1 1 1 1 0 1 1 0 0 1 1 0 1 1 1 1 1 1 0 0 1 1 0 1 0 1 1 0 1 0 1 1 0 0 1 1 1 1 0 1 1 1 1 1 1 0 1 1 0 1 0 1 0 1 0 1 1 1 1 0 1 1 0 1 0 0 1 1 1 0 0 0 1 1 0 1 1 0 1 0 0 1 0 1 1 0 0 0 0 1 1 1 1 1 1 1 1 0 1 0 1 0 1 1 1 0 0 1 0 1 0 0 1 0 0 1 0 1 0 1 0 1 1 1 0 0 0 1 0 1 1 0 1 0 1 0 0 0 0 0 0 1 0 1 1 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1
ρ0.83330.66670.63890.58330.41670.1667
Table 4 shows the optimized generator matrices for various values of ϵ 2 when k = 6 , ϵ r = 0.01 and ϵ 1 = 0.1 . We can see that as ϵ 2 decreases, the density of ones ρ decreases as well. For ϵ 2 = 0.0001 , we have A 6 = I 6 . This is due to the fact that for small value of ϵ 2 , the relay can reliably transmits the received bits to the destination. (Note that this strategy is optimal if ϵ 2 = 0 ). However, for high values of ϵ 2 , the relay combines several bits prior to the transmission. This helps the destination combat the noise on the relay-destination link.

5. Numerical Examples

Figure 6 shows the capacity results for the orthogonal BSRC shown in Figure 2 as a function of ϵ r when ϵ 1 = ϵ 2 = 0.01 . In this figure, we have plotted the cut-set upper bound (UB) (Equation (7)), rates achieved using decode-and-forward (DF) (Equation (8)), compress-and-forward (CF) (Equation (10)), and optimized finite memory relay (Equation (14)) for different memory size. The relay functions are optimized for the channel parameters ϵ 1 = ϵ 2 = ϵ r = 0.01 and are given in Table 1.
From Figure 6, we see that the achievable rate of DF decreases as ϵ r increases to 0.01 . For ϵ r 0.01 , the achievable rate of DF coincides with that achievable with direct transmission (i.e., the relay is off). On the other hand, the rates achieved by CF coincides with the upper bound for the chosen channel parameters since the condition in Corollary 2 is satisfied. More interestingly, optimized low-dimensional relaying with k = 6 achieves rates close to those achieved by CF and operates close to the capacity.
Figure 6. Capacity results for the binary symmetric relay channel as a function of ϵ r when ϵ 1 = ϵ 2 = 0.01 .
Figure 6. Capacity results for the binary symmetric relay channel as a function of ϵ r when ϵ 1 = ϵ 2 = 0.01 .
Entropy 14 02589 g006

6. Summary and Concluding Remarks

We introduced a binary symmetric relay channel with orthogonal components at the destination, and investigated three main relaying strategies: decode-and-forward (DF) relaying, compress-and-forward (CF) relaying, and optimized low-dimensional relaying. We used a bit-switching numerical algorithm to find optimized mappings. We initialized our algorithm with arbitrary random nonlinear mappings, and after optimization based on Fourier analysis, we observed that all optimized mappings that we found were linear. We also illustrated that one can obtain rates very close to the upper bound by using the proposed optimized low-dimensional relaying scheme. It is worth noting that DF and CF require codebooks with infinite block length codewords at the relay. This stands in a sharp contrast to the proposed low-dimensional relaying scheme. Additionally, the suggested relaying protocol has low-delay processing and paves the way for implementation of inexpensive relaying protocols. We finally note that the sufficiency of linear mappings for the problem of optimal relaying remains open.

Acknowledgements

The authors wish to acknowledge Abbas El Gamal for helpful input, and the Swedish Research Council for funding the work in part.

Claims

The material in this paper in part was presented in proceedings of the IEEE ITW, Cairo, Egypt, 6–8 January 2010 [23].

Appendix

A. Proof of Proposition 1

In order to proceed, we first recite a lemma given in [13] (We thank one of the anonymous reviewers for bringing this result to our attention.) that we occasionally use in the sequel.
Lemma 2. Consider a binary symmetric channel with input X and outputs Y 1 and Y 2 , where
Y 1 = X Z 1 Y 2 = X Z 2
The random variable Z 1 Ber ( ϵ 1 ) and is independent of Z 2 Ber ( ϵ 2 ) . The capacity of this channel is given by
C = 1 + H b ( ϵ 1 * ϵ 2 ) - H b ( ϵ 1 ) - H b ( ϵ 2 )
Proof. A proof of this lemma can be found in [24], but for completeness we present a slightly different proof in the following. The channel is a standard 1 × 2 SIMO (single-input multiple-output) link and its capacity is given by
C = max p ( x ) I ( X ; Y 1 , Y 2 )
Next consider
C = H ( Y 1 , Y 2 ) - H ( Y 1 , Y 2 | X ) = H ( Y 1 , Y 2 ) - H ( Z 1 , Z 2 | X ) = H ( Y 1 , Y 2 ) - H ( Z 1 ) - H ( Z 2 ) = H ( Y 1 ) + H ( Y 2 | Y 1 ) - H b ( ϵ 1 ) - H b ( ϵ 2 ) = H ( Y 1 ) + H ( Y 2 Y 1 | Y 1 ) - H b ( ϵ 1 ) - H b ( ϵ 2 ) H ( Y 1 ) + H ( Y 2 Y 1 ) - H b ( ϵ 1 ) - H b ( ϵ 2 ) = H ( Y 1 ) + H ( Z 2 Z 1 ) - H b ( ϵ 1 ) - H b ( ϵ 2 ) 1 + H b ( ϵ 1 * ϵ 2 ) - H b ( ϵ 1 ) - H b ( ϵ 2 )
We finally note that the upper bound can be achieved by choosing X Ber ( 0.5 ) .  ☐
Proof of Proposition 1: Using the cut-set bound [2], we have
C max p ( x , x r ) min { I ( X , X r ; Y 1 , Y 2 ) , I ( X ; Y 1 , Y 2 , Y r | X r ) }
Now we bound each term.
I ( X , X r ; Y 1 , Y 2 ) = H ( Y 1 , Y 2 ) - H ( Y 1 , Y 2 | X , X r ) = H ( Y 1 , Y 2 ) - H ( Z 1 , Z 2 | X , X r ) = H ( Y 1 , Y 2 ) - H ( Z 1 ) - H ( Z 2 ) H ( Y 1 ) + H ( Y 2 ) - H ( Z 1 ) - H ( Z 2 ) 2 - H b ( ϵ 1 ) - H b ( ϵ 1 )
Similarly, we have
I ( X ; Y r , Y 1 , Y 2 | X r ) = I ( X ; Y 1 , Y r | X r ) + I ( X ; Y 2 | Y 1 , Y r , X r ) = I ( X ; Y 1 , Y r | X r ) + H ( Y 2 | Y 1 , Y r , X r ) - H ( Y 2 | Y 1 , Y r , X r , X ) = I ( X ; Y 1 , Y r | X r ) + H ( Z 2 | Y 1 , Y r , X r ) - H ( Z 2 | Y 1 , Y r , X r , X ) = I ( X ; Y 1 , Y r | X r ) = H ( Y 1 , Y r | X r ) - H ( Y 1 , Y r | X , X r ) = H ( Y 1 , Y r | X r ) - H ( Z 1 , Z r | X , X r ) = H ( Y 1 , Y r | X r ) - H ( Z 1 , Z r ) H ( Y 1 , Y r ) - H ( Z 1 , Z r ) = I ( X ; Y 1 , Y r ) 1 + H b ( ϵ 1 * ϵ r ) - H b ( ϵ 1 ) - H b ( ϵ r )
where the last inequality follows from Lemma 2. Combining (27) and (28) and noting that (27) and (28) are maximized when the input distribution is chosen as
p ( x , x r ) = p ( x ) p ( x r ) = 0.5 δ b ( x ) + 0.5 δ b ( x 1 ) 0.5 δ b ( x r ) + 0.5 δ b ( x r 1 )
yield the result.

B. Proof of Proposition 2

Using Theorem 1 in [2], the rate
R D F = max p ( x , x r ) min { I ( X , X r ; Y 1 , Y 2 ) , I ( X ; Y r | X r ) }
is achievable. The first term is evaluated in Appendix A and is maximized when X and X r are independent and have uniform distribution. One can show the same distribution maximizes the second term. This yields
R D F = min { 1 - H b ( ϵ r ) , 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) } .

C. Proof of Proposition 3

Using Theorem 7 in [2] , the rate
R P D F = max p ( u , x , x r ) min { I ( X , X r ; Y 1 , Y 2 ) , I ( U ; Y r | X r ) + I ( X ; Y 1 , Y 2 | X r , U ) }
is achievable. (See also [15].) The first term is evaluated in Appendix A and is bounded as
I ( X , X r ; Y 1 , Y 2 ) 2 - H b ( ϵ 1 ) - H b ( ϵ 2 )
Next consider
I ( U ; Y r | X r ) + I ( X ; Y 1 , Y 2 | X r , U ) = I ( U ; Y r | X r ) + H ( Y 1 , Y 2 | X r , U ) - H ( Y 1 , Y 2 | X , X r , U ) = I ( U ; Y r | X r ) + H ( Y 1 , Y 2 | X r , U ) - H ( Z 1 , Z 2 | X , X r , U ) = I ( U ; Y r | X r ) + H ( Y 1 , Y 2 | X r , U ) - H ( Z 1 ) - H ( Z 2 ) = I ( U ; Y r | X r ) + H ( Y 1 | X r , U ) + H ( Y 2 | X r , U , Y 1 ) - H ( Z 1 ) - H ( Z 2 ) = I ( U ; Y r | X r ) + H ( Y 1 | X r , U ) + H ( Z 2 | X r , U , Y 1 ) - H ( Z 1 ) - H ( Z 2 ) = H ( Y r | X r ) - H ( Y r | X r , U ) + H ( Y 1 | X r , U ) - H ( Z 1 ) H ( Y r ) - H ( Y r | X r , U ) + H ( Y 1 | X r , U ) - H ( Z 1 ) 1 - H b ( ϵ 1 ) + H ( Y 1 | X r , U ) - H ( Y r | X r , U )
Now we bound H ( Y 1 | X r , U ) - H ( Y r | X r , U ) . First, define V : = ( X r , U ) where p ( V = v i ) = p i and i p i = 1 . Further assume that p ( x = 0 | V = v i ) = δ i , p ( x = 1 | V = v i ) = 1 - δ i . Next consider
H ( Y 1 | X r , U ) - H ( Y r | X r , U ) = H ( X + Z 1 | V ) - H ( X + Z r | V ) = i H ( X + Z 1 | V = v i ) - H ( X + Z r | V = v i ) p i i max δ i H ( X + Z 1 | V = v i ) - H ( X + Z r | V = v i ) p i = max δ i H ( X + Z 1 | V = v i ) - H ( X + Z r | V = v i ) i p i = max δ i H ( X + Z 1 | V = v i ) - H ( X + Z r | V = v i ) = max δ i H b ( δ i * ϵ 1 ) - H b ( δ i * ϵ r )
In the following, let h ( δ ) : = H b ( δ * ϵ 1 ) - H b ( δ * ϵ r ) and max { ϵ 1 , ϵ r } 0.5 . We then obtain
h δ = ( 1 - 2 ϵ 1 ) log 1 - ( δ * ϵ 1 ) δ * ϵ 1 - ( 1 - 2 ϵ r ) log 1 - ( δ * ϵ r ) δ * ϵ r 2 h δ 2 = ( 1 - 2 ϵ r ) 2 ( δ * ϵ r ) ( 1 - ( δ * ϵ r ) ) - ( 1 - 2 ϵ 1 ) 2 ( δ * ϵ 1 ) ( 1 - ( δ * ϵ 1 ) )
Now let g ( ϵ ) : = ( 1 - 2 ϵ ) 2 ( δ * ϵ ) ( 1 - ( δ * ϵ ) ) . One can show that g ϵ 0 if ϵ 0.5 and hence g ( ϵ ) is a non-increasing function. Thus we conclude that
ϵ r ϵ 1 g ( ϵ r ) g ( ϵ 1 ) 2 h δ 2 0 h ( δ ) is convex ϵ r ϵ 1 g ( ϵ r ) g ( ϵ 1 ) 2 h δ 2 0 h ( δ ) is concave
Therefore
h ( δ ) max { h ( 0 ) , h ( 1 ) } , if ϵ r ϵ 1 h ( δ ) h ( δ ) , if ϵ r ϵ 1
where δ is the solution of h δ = 0 . Finally we obtain the following bound
H ( Y 1 | X r , U ) - H ( Y r | X r , U ) H b ( ϵ 1 ) - H b ( ϵ r ) , if ϵ r ϵ 1 0 , otherwise
Combining Equations (30), (31) and (36) proves that partial DF does not improve on DF for the BSRC.

D. Proof of Proposition 4

We use the equivalent formulation of the original CF (Theorem 6 in [2]) given in [4]. CF achieves the rate
R C F = max p ( q ) p ( x | q ) p ( x r | q ) p ( y ^ r | x r , y r , q ) min I ( X , X r ; Y 1 , Y 2 | Q ) - I ( Y r ; Y ^ r | X , X r , Y 1 , Y 2 , Q ) , I ( X ; Y 1 , Y 2 , Y ^ r | X r , Q )
First consider
I ( X , X r ; Y 1 , Y 2 | Q ) = I ( X ; Y 1 , Y 2 | Q ) + I ( X r ; Y 1 , Y 2 | X , Q ) = I ( X ; Y 1 | Q ) + I ( X ; Y 2 | Y 1 , Q ) + I ( X r ; Y 2 | X , Q ) + I ( X r ; Y 1 | Y 2 , X , Q ) = I ( X ; Y 1 | Q ) + I ( X r ; Y 2 | X , Q ) = I ( X ; Y 1 | Q ) + I ( X r ; Y 2 | Q ) 2 - H b ( ϵ 1 ) - H b ( ϵ 2 )
where the upper bound can be achieved by choosing X Ber ( 0.5 ) and X r Ber ( 0.5 ) .
In order to proceed we choose the following binary test channel:
Y ^ r = Y r Z q
where Z q Ber ( ϵ q ) and is independent of other random variables.
This yields
I ( Y r ; Y ^ r | X , X r , Y 1 , Y 2 , Q ) = H ( Y ^ r | X , X r , Y 1 , Y 2 , Q ) - H ( Y ^ r | Y r , X , X r , Y 1 , Y 2 , Q ) = H ( Z r Z q ) - H ( Z q ) = H b ( ϵ r * ϵ q ) - H b ( ϵ q )
I ( X ; Y 1 , Y 2 , Y ^ r | X r , Q ) = I ( X ; Y 1 , Y ^ r | X r , Q ) + I ( X ; Y 2 | Y 1 , Y ^ r , X r , Q ) = I ( X ; Y 1 , Y ^ r | Q ) 1 + H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ r * ϵ q ) - H b ( ϵ 1 )
where the last inequality follows from Lemma 2 and it is achieved by choosing X Ber ( 0.5 ) .
Putting all together, the following rate is achievable
R C F = max ϵ q [ 0 , 1 ] min { 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) - H b ( ϵ r * ϵ q ) + H b ( ϵ q ) , 1 + H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ r * ϵ q ) - H b ( ϵ 1 ) }
Now define
R 1 ( ϵ q ) : = 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) - H b ( ϵ r * ϵ q ) + H b ( ϵ q ) R 2 ( ϵ q ) : = 1 + H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ r * ϵ q ) - H b ( ϵ 1 )
Using the fact that f ( ϵ ) : = H b ( ϵ * δ ) - H b ( ϵ ) is convex δ [ 0 , 1 ] , we conclude that R 1 ( ϵ q ) is concave and R 2 ( ϵ q ) is convex in ϵ q . We next note that
max ϵ q R 1 = R 1 ( 0.5 ) = 2 - H b ( ϵ 1 ) - H b ( ϵ 2 ) min ϵ q R 2 = R 2 ( 0.5 ) = 1 - H b ( ϵ 1 )
Since R 2 ( 0.5 ) R 1 ( 0.5 ) we only need to consider two following cases:
  • Case 1: R 1 ( 0 ) < R 2 ( 0 )
    If R 1 ( 0 ) < R 2 ( 0 ) we have 1 - H b ( ϵ 2 ) < H b ( ϵ 1 * ϵ r ) and there exists ϵ q such that R 1 ( ϵ q ) = R 2 ( ϵ q ) . Thus
    R C F = 1 + H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ r * ϵ q ) - H b ( ϵ 1 )
    is achievable where ϵ q satisfies
    H b ( ϵ 1 * ϵ r * ϵ q ) - H b ( ϵ q ) + H b ( ϵ 2 ) = 1
  • Case 2: R 1 ( 0 ) R 2 ( 0 )
    If R 1 ( 0 ) R 2 ( 0 ) we have 1 - H b ( ϵ 2 ) H b ( ϵ 1 * ϵ r ) . Thus
    R C F = max ϵ q min { R 1 ( ϵ q ) , R 2 ( ϵ q ) } = max ϵ q R 2 ( ϵ q ) = R 2 ( 0 ) = 1 + H b ( ϵ 1 * ϵ r ) - H b ( ϵ r ) - H b ( ϵ 1 )
    is achievable.

E. Proof of Proposition 5

For the memoryless relay, we have x r = y r and hence
Y 1 = X Z 1 Y 2 = X r Z 2 = X Z r Z 2 = X Z e q
where Z e q : = Z r Z 2 Ber ( ϵ r * ϵ 2 ) . Then the achievable rate is given by C 1 = max p ( x ) I ( X ; Y 1 , Y 2 ) . Now using Lemma 2, we obtain
C 1 = 1 + H b ( ϵ 1 * ϵ r * ϵ 2 ) - H b ( ϵ r * ϵ 2 ) - H b ( ϵ 1 )

F. Proof of Lemma 1

Consider
I ( X 1 k ; Y 1 , 1 k , Y 2 , 1 k ) = I ( X 1 k ; Y 1 , 1 k ) + I ( X 1 k ; Y 2 , 1 k | Y 1 , 1 k ) = ( a ) k I ( X 1 ; Y 1 ) + I ( X 1 k ; Y 2 , 1 k | Y 1 , 1 k ) = k I ( X 1 ; Y 1 ) + H ( Y 2 , 1 k | Y 1 , 1 k ) - H ( Y 2 , 1 k | Y 1 , 1 k , X 1 k ) = ( b ) k I ( X 1 ; Y 1 ) + H ( Y 2 , 1 k | Y 1 , 1 k ) - H ( Y 2 , 1 k | X 1 k ) = k I ( X 1 ; Y 1 ) + H ( Y 2 , 1 k | Y 1 , 1 k ) - H ( Y 2 , 1 k | X 1 k ) = k ( 1 - H b ( ϵ 1 ) ) - E log 2 ( p ( y 2 , 1 k | y 1 , 1 k ) ) + E log 2 ( p ( y 2 , 1 k | x 1 k ) )
where ( a ) holds since { X i } i = 1 k are i.i.d. and the channel is memoryless and ( b ) holds since Y 1 , 1 k = X 1 k Z 1 , 1 k and Z 1 , 1 k is independent of other random sequences. The conditional probabilities can be computed as follows
p ( y 2 , 1 k | x 1 k ) = y r 1 k p ( y 2 , 1 k | x 1 k , y r , 1 k ) p ( y r , 1 k | x 1 k ) = y r 1 k p ( y 2 , 1 k | y r , 1 k ) p ( y r , 1 k | x 1 k ) = y r 1 k p ( y 2 , 1 k | y r , 1 k ) i = 1 k p ( y r , i | x i ) = y r 1 k i = 1 k p ( y 2 , i | f i ( y r , 1 k ) ) i = 1 k p ( y r , i | x i ) = y r 1 k ϵ 2 D H ( y 2 , 1 k , f ( y r , 1 k ) ) ( 1 - ϵ 2 ) k - D H ( y 2 , 1 k , f ( y r , 1 k ) ) × ϵ r D H ( y r , 1 k , x 1 k ) ( 1 - ϵ r ) k - D H ( y r , 1 k , x 1 k ) = ( 1 - ϵ 2 ) k ( 1 - ϵ r ) k × y r 1 k ϵ 2 1 - ϵ 2 D H ( y 2 , 1 k , f ( y r , 1 k ) ) ϵ r 1 - ϵ r D H ( y r , 1 k , x 1 k )
We similarly obtain
p ( y 2 , 1 k | y 1 , 1 k ) = y r 1 k p ( y 2 , 1 k | y 1 , 1 k , y r , 1 k ) p ( y r , 1 k | y 1 , 1 k ) = y r 1 k p ( y 2 , 1 k | y r , 1 k ) p ( y r , 1 k | y 1 , 1 k ) = y r 1 k p ( y 2 , 1 k | y r , 1 k ) x 1 k p ( y r , 1 k | y 1 , 1 k , x 1 k ) p ( x 1 k | y 1 k ) = y r 1 k p ( y 2 , 1 k | y r , 1 k ) x 1 k p ( y r , 1 k | x 1 k ) p ( y 1 , 1 k | x 1 k ) p ( x 1 k ) p ( y 1 , 1 k ) = y r 1 k i = 1 k p ( y 2 , i | f i ( y r , 1 k ) ) x 1 k i = 1 k p ( y r , i k | x i ) i = 1 k p ( y 1 , i | x i ) = ( 1 - ϵ 1 ) k ( 1 - ϵ 2 ) k ( 1 - ϵ r ) k × y r 1 k ϵ 2 1 - ϵ 2 D H ( y 2 , 1 k , f ( y r , 1 k ) ) × x 1 k ϵ r 1 - ϵ r D H ( y r , 1 k , x 1 k ) ϵ 1 1 - ϵ 1 D H ( y 1 , 1 k , x 1 k )

References

  1. van der Meulen, E.C. Three-terminal communication channels. Adv. Appl. Probab. 1971, 3, 120–154. [Google Scholar] [CrossRef]
  2. Cover, T.M.; El Gamal, A. Capacity theorems for the relay channel. IEEE Trans. Inform. Theory 1979, 25, 572–584. [Google Scholar] [CrossRef]
  3. Høst-Madsen, A.; Zhang, J. Capacity bounds and power allocation for wireless relay channels. IEEE Trans. Inform. Theory 2006, 51, 2020–2040. [Google Scholar] [CrossRef]
  4. El Gamal, A.; Mohseni, M.; Zahedi, S. Bounds on capacity and minimum energy-per-bit for AWGN relay channels. IEEE Trans. Inform. Theory 2006, 52, 1545–1561. [Google Scholar] [CrossRef]
  5. Kramer, G.; Gastpar, M.; Gupta, P. Cooperative strategies and capacity theorems for relay networks. IEEE Trans. Inform. Theory 2005, 51, 3037–3063. [Google Scholar] [CrossRef]
  6. Dabora, R.; Servetto, S.D. On the role of estimate-and-forward with time sharing in cooperative communication. IEEE Trans. Inform. Theory 2008, 541, 4409–4431. [Google Scholar] [CrossRef]
  7. Laneman, J.N.; Wornell, G.W.; Tse, D.N.C. Cooperative diversity in wireless networks: Efficient protocols and outage behavior. IEEE Trans. Inform. Theory 2004, 50, 3062–3080. [Google Scholar] [CrossRef]
  8. Sendonaris, A.; Erkip, E.; Aazhang, B. User cooperation diversity-Part I: System description. IEEE Trans. Commun. 2003, 51, 1927–1938. [Google Scholar] [CrossRef]
  9. Sendonaris, A.; Erkip, E.; Aazhang, B. User cooperation diversity-Part II: Implementation aspects and performance analysis. IEEE Trans. Commun. 2003, 51, 1939–1948. [Google Scholar] [CrossRef]
  10. Aleksic, M.; Razaghi, P.; Yu, W. Capacity of a class of modulo-sum relay channels. IEEE Trans. Inform. Theory 2009, 55, 921–930. [Google Scholar] [CrossRef]
  11. Kim, Y.H. Coding Techniques for Primitive Relay Channels. In Proceedings of the Forty-Fifth Annual Allerton Conference, Allerton House, UIUC, IL, USA, 26–28 September 2007.
  12. Lau, A.P.T.; Cui, S. Joint power minimization in wireless relay channels. IEEE Trans. Wirel. Commun. 2007, 6, 2820–2824. [Google Scholar]
  13. Karystinos, G.N.; Liavas, A.P. Outage Capacity of a Cooperative Scheme with Binary Input and a Simple Relay. In Proceedings of the IEEE ICASSP 2008-International Conference Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March– 4 April 2008.
  14. Sagar, Y.; Kwon, H.M.; Ding, Y. Capacity of Modulo-Sum Simple Relay Network. In Proceedings International Zurich Seminar on Communications, Zurich, Switzerland, 3–5 March 2010.
  15. El Gamal, A.; Aref, M. The capacity of the semi-deterministic relay channel. IEEE Trans. Inform. Theory 1982, 28, 536. [Google Scholar] [CrossRef]
  16. Khormuji, M.N.; Larsson, E.G. Rate-optimized constellation rearrangement for the relay channel. IEEE Commun. Lett. 2008, 12, 618–620. [Google Scholar] [CrossRef]
  17. Khormuji, M.N.; Skoglund, M. Piecewise linear relaying: Low complexity parametric relaying. In Proceedings of the IEEE SPAWC, Perugia, Italy, 21–24 June 2009.
  18. Khormuji, M.N.; Skoglund, M. On instantaneous relaying. IEEE Trans. Inform. Theory 2010, 56, 3378–3394. [Google Scholar] [CrossRef]
  19. Zaidi, A.; Khormuji, M.N.; Yao, S.; Skoglund, M. Rate-maximizing Mappings for Memoryless Relaying. In Proceedings of the IEEE ISIT, Seoul, Korea, 28 June–3 July, 2009.
  20. Skoglund, M. On channel-constrained vector quantization and index assignment for discrete memoryless channels. IEEE Trans Inform. Theory 1999, 45, 2615–2622. [Google Scholar] [CrossRef]
  21. Mehes, A.; Zeger, K. Performance of quantizers on noisy channels using structured families of codes. IEEE Trans. Inform. Theory 2000, 46, 2468–2476. [Google Scholar]
  22. Rudin, W. Fourier Analysis on Groups; John Wiley And Sons Ltd: Hoboken, NJ, USA, 1990. [Google Scholar]
  23. Khormuji, M.N.; Skoglund, M. On the Capacity of the Binary Symmetric Relay Channel with a Finite Memory Relay. In Proceedings of the IEEE ITW, Cairo, Egypt, 6–8 January 2010.
  24. Liavas, A. Outage Capacity of a Cooperative Scheme with Binary Input and a Simple Relay. Year-1 Work-Package-1 report of European Commission FET Project FP6-033533-COOPCOM, November 2007. [Google Scholar]

Share and Cite

MDPI and ACS Style

Khormuji, M.N.; Skoglund, M. Capacity Bounds and Mapping Design for Binary Symmetric Relay Channels. Entropy 2012, 14, 2589-2610. https://doi.org/10.3390/e14122589

AMA Style

Khormuji MN, Skoglund M. Capacity Bounds and Mapping Design for Binary Symmetric Relay Channels. Entropy. 2012; 14(12):2589-2610. https://doi.org/10.3390/e14122589

Chicago/Turabian Style

Khormuji, Majid Nasiri, and Mikael Skoglund. 2012. "Capacity Bounds and Mapping Design for Binary Symmetric Relay Channels" Entropy 14, no. 12: 2589-2610. https://doi.org/10.3390/e14122589

Article Metrics

Back to TopTop