Skip to main content
Entropy logoLink to Entropy
. 2019 May 22;21(5):516. doi: 10.3390/e21050516

The Arbitrarily Varying Relay Channel

Uzi Pereg 1,*, Yossef Steinberg 1
PMCID: PMC7515005  PMID: 33267230

Abstract

We study the arbitrarily varying relay channel, which models communication with relaying in the presence of an active adversary. We establish the cutset bound and partial decode-forward bound on the random code capacity. We further determine the random code capacity for special cases. Then, we consider conditions under which the deterministic code capacity is determined as well. In addition, we consider the arbitrarily varying Gaussian relay channel with sender frequency division under input and state constraints. We determine the random code capacity, and establish lower and upper bounds on the deterministic code capacity. Furthermore, we show that as opposed to previous relay models, the primitive relay channel has a different behavior compared to the non-primitive relay channel in the arbitrarily varying scenario.

Keywords: arbitrarily varying channel, relay channel, decode-forward, Markov block code, minimax theorem, deterministic code, random code, symmetrizability

1. Introduction

The relay channel was first introduced by van der Meulen [1] to describe point-to-point communication with the help of a relay, which receives a noisy version of the transmitter signal and transmits a signal of its own to the destination receiver. The relay channel is generally perceived as a fundamental building block for multihop networks (see e.g., [2,3], Chapter 16), where some nodes receive and transmit in order to assist the information flow between other nodes. The capacity of the relay channel is not known in general, however, Cover and El Gamal established the cutset upper bound, the decode-forward lower bound, and the partial decode-forward lower bound [4]. It was also shown in [4] that for the reversely degraded relay channel, direct transmission is capacity achieving. For the degraded relay channel, the decode-forward lower bound and the cutset upper bound coincide, thus characterizing the capacity for this model [4].

In general, the partial decode-forward lower bound is tighter than both direct transmission and decode-forward lower bounds. El Gamal and Zahedi [5] determined the capacity of the relay channel with orthogonal sender components, by showing that the partial decode-forward lower bound and cutset upper bound coincide. A variation of the relay channel, referred to as the primitive relay channel, was introduced by Kim [2], and attracted a lot of attention (see e.g., [6,7,8,9,10,11,12] and references therein). Recently, there has also been a growing interest in the Gaussian relay channel, as e.g., in [5,7,9,13,14,15,16] and references therein. In particular, El Gamal and Zahedi [5] introduced the Gaussian relay channel with sender frequency division (SFD), as a special case of a relay channel with orthogonal sender components. There are many other relaying scenarios, including secrecy [17,18], networking [15,19,20,21,22], parallel relaying [23,24,25], diamond channels [26,27,28], side information [29,30,31,32,33], etc.

In practice, the channel statistics are not necessarily known in exact, and they may even change over time. The arbitrarily varying channel (AVC) is an appropriate model to describe such a situation [34]. In real systems, such variations are caused by fading in wireless communication [35,36,37,38,39,40,41,42], memory faults in storage [43,44,45,46,47], malicious attacks on identification and authorization systems [48,49], etc. It is especially relevant to communication in the presence of an adversary, or a jammer, attempting to disrupt communication. Jamming attacks are not limited to point-to-point communication, and cause a major security concern for cognitive radio networks [50] and wireless sensor networks [42,51,52,53,54], for instance.

Considering the AVC without a relay, Blackwell et al. determined the random code capacity [34], i.e., the capacity achieved by stochastic-encoder stochastic-decoder coding schemes with common randomness. It was also demonstrated in [34] that the random code capacity is not necessarily achievable using deterministic codes. A well-known result by Ahlswede [55] is the dichotomy property of the AVC. Specifically, the deterministic code capacity either equals the random code capacity or else, it is zero. Subsequently, Ericson [56] and Csiszár and Narayan [57] established a simple single-letter condition, namely non-symmetrizability, which is both necessary and sufficient for the capacity to be positive. Ahlswede’s Robustification Technique (RT) is a useful technique for the AVC analysis, developed and applied to classical AVC settings [58,59]. Essentially, the RT uses a reliable code for the compound channel to construct a random code for the AVC applying random permutations to the codeword symbols. A continuing line of works on arbitrarily varying networks includes among others the arbitrarily varying broadcast channel [60,61,62,63,64,65], multiple-access channel [60,66,67,68,69,70,71,72,73,74,75], and wiretap channel [76,77,78,79,80,81,82,83,84]. The reference lists here are far from being exhaustive.

In this work, we introduce a new model, namely, the arbitrarily varying relay channel (AVRC). The AVRC combines the previous models, i.e., the relay channel and the AVC, and we believe that it is a natural problem to consider, in light of the jamming attacks on current and future networks, as mentioned above. In the analysis, we incorporate the block Markov coding schemes of [4] in Ahlswede’s Robustification and Elimination Techniques [55,59]. A straightforward application of Ahlswede’s RT fails to comply with the strictly causal relay transmission. In a recent work [85,86], by the authors of this paper, a modified RT technique was presented and applied to the point-to-point AVC with causal side information under input and state constraints, without a relay. This was the first time where the application of the RT exploited the structure of the original compound channel code to construct a random code for the AVC, as opposed to earlier work where the original code is treated as a “black box”. Here, we present another modification of the RT, which also exploits the structure of the original compound channel code, but in a different manner. The analysis also requires to redefine the compound channel, and we refer to the newly defined channel as the block-compound relay channel.

We establish the cutset upper bound and the full/partial decode-forward lower bound on the random code capacity of the AVRC. The random code capacity is determined in special cases of the degraded AVRC, the reversely degraded AVRC, and the AVRC with orthogonal sender components. Then, we give extended non-symmetrizability conditions under which the deterministic code capacity coincides with the random code capacity. We show by example that the deterministic code capacity can be strictly lower than the random code capacity of the AVRC. Then, we consider the Gaussian AVRC with SFD, under input and state constraints. The random code capacity is determined using the previous results, whereas the deterministic code capacity is lower and upper bounded using an independent approach. Specifically, we extend the techniques from [87], where Csiszár and Narayan determine the capacity of the Gaussian AVC under input and state constraint. It is shown that for low values on the input constraint, the deterministic code capacity can be strictly lower than the random code capacity, but yet non-zero.

Furthermore, we give similar bounds for the primitive AVRC, where there is a noiseless link between the relay and the receiver of limited capacity [2]. We find the capacity of the primitive counterpart of the Gaussian AVRC with SFD, in which case the deterministic and random code capacities coincide, regardless of the value of the input constraint. We deduce that Kim’s assertion—that “the primitive relay channel captures most essential features and challenges of relaying, and thus serves as a good testbed for new relay coding techniques” [2]—is not true in the arbitrarily varying scenario.

This work is organized as follows. In Section 2, the basic definitions and notation are provided. In Section 3, we give the main results on the general AVRC. The Gaussian AVRC with SFD is introduced in Section 4, and the main results are given in Section 5. The definition and results on the primitive AVRC are in Section 6.

2. Definitions

2.1. Notation

We use the following notation conventions throughout. Calligraphic letters X,S,Y, are used for finite sets. Lowercase letters x,s,y, stand for constants and values of random variables, and uppercase letters X,S,Y, stand for random variables. The distribution of a random variable X is specified by a probability mass function (pmf) PX(x)=p(x) over a finite set X. The set of all pmfs over X is denoted by P(X). We use xj=(x1,x2,,xj) to denote a sequence of letters from X. A random sequence Xn and its distribution PXn(xn)=p(xn) are defined accordingly. For a pair of integers i and j, 1ij, we define the discrete interval [i:j]={i,i+1,,j}. The notation x=(x1,x2,,xn) is used when it is understood from the context that the length of the sequence is n, and the 2-norm of x is denoted by x.

2.2. Channel Description

A state-dependent discrete memoryless relay channel (X,X1,S,WY,Y1|X,X1,S,Y,Y1) consists of five sets, X, X1, S, Y and Y1, and a collection of conditional pmfs WY,Y1|X,X1,S. The sets stand for the input alphabet, the relay transmission alphabet, the state alphabet, the output alphabet, and the relay input alphabet, respectively. The alphabets are assumed to be finite, unless explicitly said otherwise. The channel is memoryless without feedback, and therefore

WYn,Y1n|Xn,X1n,Sn(yn,y1n|xn,x1n,sn)=i=1nWY,Y1|X,X1,S(yi,y1,i|xi,x1,i,si). (1)

Communication over a relay channel is depicted in Figure 1. Following [29], a relay channel WY,Y1|X,X1,S is called degraded if the channel can be expressed as

WY,Y1|X,X1,S(y,y1|x,x1,s)=WY1|X,X1,S(y1|x,x1,s)WY|Y1,X1,S(y|y1,x1,s), (2)

and it is called reversely degraded if

WY,Y1|X,X1,S(y,y1|x,x1,s)=WY|X,X1,S(y|x,x1,s)WY1|Y,X1,S(y1|y,x1,s). (3)

We say that the relay channel is strongly degraded or reversely degraded, if the respective definition holds such that the sender-relay marginal is independent of the state. That is, WY,Y1|X,X1,S is strongly degraded if WY,Y1|X,X1,S=WY1|X,X1WY|Y1,X1,S, and similarly, WY,Y1|X,X1,S is strongly reversely degraded if WY,Y1|X,X1,S=WY|X,X1,SWY1|Y,X1. For example, if Y1=X+Z and Y=Y1+X1+S, where Z is an independent additive noise, then WY,Y1|X,X1,S is strongly degraded. Whereas, if Y=X+X1+S and Y1=Y+Z, then WY,Y1|X,X1,S is strongly reversely degraded.

Figure 1.

Figure 1

Communication over the arbitrarily varying relay channel L={WY,Y1|X,X1,S}. Given a message M, the encoder transmits Xn=f(M). At time i[1:n], the relay transmits X1,i based on all the symbols of the past Y1i1 and then receives a new symbol Y1,i. The decoder receives the output sequence Yn, and finds an estimate of the message M^=g(Yn).

The arbitrarily varying relay channel (AVRC) is a discrete memoryless relay channel (X,X1,S,WY,Y1|X,X1,S,Y,Y1) with a state sequence of unknown distribution, not necessarily independent nor stationary. That is, Snq(sn) with an unknown joint pmf q(sn) over Sn. In particular, q(sn) can give mass 1 to some state sequence sn. We use the shorthand notation L={WY,Y1|X,X1,S} for the AVRC, where the alphabets are understood from the context.

To analyze the AVRC, we consider the compound relay channel. Different models of compound relay channels have been considered in the literature [30,88]. Here, we define the compound relay channel as a discrete memoryless relay channel (X,X1,S, WY,Y1|X,X1,S,Y,Y1) with a discrete memoryless state, where the state distribution q(s) is not known in exact, but rather belongs to a family of distributions Q, with QP(S). That is, Sni=1nq(si), with an unknown pmf qQ over S. We use the shorthand notation LQ for the compound relay channel, where the transition probability WY,Y1|X,X1,S and the alphabets are understood from the context.

In the analysis, we also use the following model. Suppose that the user transmits B>0 blocks of length n, and the jammer is entitled to use a different state distribution qb(s)Q for every block b[1:B], while the encoder, relay and receiver are aware of this jamming scheme. In other words, every block is governed by a different memoryless state. We refer to this channel as the block-compound relay channel, denoted by LQ×B. Although this is a toy model, it is a useful tool for the analysis of the AVRC.

2.3. Coding

We introduce some preliminary definitions, starting with the definitions of a deterministic code and a random code for the AVRC L. Note that in general, the term ‘code’, unless mentioned otherwise, refers to a deterministic code.

Definition 1 

(A code, an achievable rate and capacity). A (2nR,n) code for the AVRC L consists of the following; a message set [1:2nR], where it is assumed throughout that 2nR is an integer, an encoder f:[1:2nR]Xn, a sequence of n relaying functions f1,i:Y1i1X1,i, i[1:n], and a decoding function g:Yn[1:2nR].

Given a message m[1:2nR], the encoder transmits xn=f(m). At time i[1:n], the relay transmits x1,i=f1,i(y1i1) and then receives y1,i. The relay codeword is given by x1n=f1n(y1n)f1,i(y1i1)i=1n. The decoder receives the output sequence yn, and finds an estimate of the message m^=g(yn) (see Figure 1). We denote the code by C=f(·),f1n(·),g(·). Define the conditional probability of error of the code C given a state sequence snSn by

Pe|sn(n)(C)=12nRm=12nR(yn,y1n):g(yn)mi=1nWY,Y1|X,X1,S(yi,y1,i|fi(m),f1,i(y1i1),si). (4)

Now, define the average probability of error of C for some distribution q(sn)P(Sn),

Pe(n)(q,C)=snSnq(sn)·Pe|sn(n)(C). (5)

Observe that Pe(n)(q,C) is linear in q, and thus continuous. We say that C is a (2nR,n,ε) code for the AVRC L if it further satisfies

Pe(n)(q,C)ε,forallq(sn)P(Sn). (6)

A rate R is called achievable if for every ε>0 and sufficiently large n, there exists a (2nR,n,ε) code. The operational capacity is defined as the supremum of the achievable rates and it is denoted by C(L). We use the term ‘capacity’ referring to this operational meaning, and in some places we call it the deterministic code capacity in order to emphasize that achievability is measured with respect to deterministic codes.

We proceed now to define the parallel quantities when using stochastic-encoders stochastic-decoder triplets with common randomness. The codes formed by these triplets are referred to as random codes.

Definition 2 

(Random code). A (2nR,n) random code for the AVRC L consists of a collection of (2nR,n) codes {Cγ=(fγ,f1,γn,gγ)}γΓ, along with a probability distribution μ(γ) over the code collection Γ. We denote such a code by CΓ=(μ,Γ,{Cγ}γΓ). Analogously to the deterministic case, a (2nR,n,ε) random code has the additional requirement

Pe(n)(q,CΓ)=γΓμ(γ)Pe(n)(q,Cγ)ε,forallq(sn)P(Sn). (7)

The capacity achieved by random codes is denoted by C(L), and it is referred to as the random code capacity.

3. Main Results—General AVRC

We present our results on the compound relay channel and the AVRC.

3.1. The Compound Relay Channel

We establish the cutset upper bound and the partial decode-forward lower bound for the compound relay channel. Consider a given compound relay channel LQ. Let

RCS(LQ)infqQmaxp(x,x1)minIq(X,X1;Y),Iq(X;Y,Y1|X1), (8)

and

RPDF(LQ)maxp(u,x,x1)min{infqQIq(U,X1;Y)+infqQIq(X;Y|X1,U),infqQIq(U;Y1|X1)+infqQIq(X;Y|X1,U)}, (9)

where the subscripts ‘CS’ and ‘DF’ stand for ‘cutset’ and ‘partial decode-forward’, respectively.

Lemma 1.

The capacity of the compound relay channel LQ is bounded by

C(LQ)RPDF(LQ), (10)
C(LQ)RCS(LQ). (11)

Specifically, if R<RPDF(LQ), then there exists a (2nR,n,ean) block Markov code over LQ for sufficiently large n and some a>0.

The proof of Lemma 1 is given in Appendix A. The achievability proof is based on block Markov coding interlaced with the partial decode-forward scheme. That is, the encoder sends a sequence of messages over multiple blocks. The message in each block consists of two components, a decode-forward component, and a direct transmission component, where only the former is decoded by the relay. The name ‘decode-forward component’ stands for the fact that the relay decodes this message component and sends its estimation forwards, to the destination receiver. Once the decoder has received all blocks, the decode-forward components are decoded backwards, i.e., starting with the message in the last block going backwards. Using the estimation of the decode-forward components, the direct transmission components are decoded forwards, i.e., starting with the message in the first block going forwards. The ambiguity of the state distribution needs to be dealt with throughout all of those estimations. In both decoding stages, the receiver performs joint typicality decoding using a set of types that “quantizes” the set Q of state distributions.

Remark 1.

If the set of state distributions Q is convex, then the upper bound expression in the RHS of Equation (8) has a minmax form. On the other hand, in the lower bound expression in the RHS of Equation (9), the maximum comes first, and then we have multiple min terms, which makes this expression a lot more complicated than the classical partial decode-forward bound [4] (see also [3], Theorem 16.3), where Markov properties lead to a simpler expression. We note that this phenomenon (or one might say, disturbance) where the lower bound has multiple min terms is not exclusive to the AVRC. A noteworthy example is the arbitrarily varying wiretap channel [76,89], where the lower bound has the form of max[minIq(U;Y)maxIq(U;Z)]. While the capacity of the classical wiretap channel is known, the arbitrarily varying counterpart has remained an open problem for several years.

Observe that taking U= in (9) gives the direct transmission lower bound,

C(LQ)RPDF(LQ)maxp(x,x1)infqQIq(X;Y|X1). (12)

Taking U=X in (9) results in a full decode-forward lower bound,

C(LQ)RPDF(LQ)maxp(x,x1)infqQminIq(X,X1;Y),Iq(X;Y1|X1). (13)

This yields the following corollary. The corollary uses the terms of a strongly degraded relay channel, for which WY,Y1|X,X1,S=WY1|X,X1WY|Y1,X1,S, and a strongly reversely degraded relay channel, for which WY,Y1|X,X1,S=WY|X,X1,SWY1|Y,X1, as defined in Section 2.2.

Corollary 1.

Let LQ be a compound relay channel, where Q is a compact convex set.

  • 1. 
    If WY,Y1|X,X1,S is strongly reversely degraded, then
    C(LQ)=RPDF(LQ)=RCS(LQ)=minqQmaxp(x,x1)Iq(X;Y|X1). (14)
  • 2. 
    If WY,Y1|X,X1,S is strongly degraded, then
    C(LQ)=RPDF(LQ)=RCS(LQ)=maxp(x,x1)minminqQIq(X,X1;Y),I(X;Y1|X1). (15)

The proof of Corollary 1 is given in Appendix B. Part 1 follows from the direct transmission and cutset bounds, (12) and (8), respectively, while part 2 is based on the full decode-forward and cutset bounds, (13) and (8), respectively, along with the convexity considerations in the remark below.

Remark 2.

On a technical level, there are two purposes for considering the strongly degraded relay channel, for which the marginal channel to the relay is independent of the state, i.e., WY1|X,X1,S=WY1|X,X1 (see Section 2.2). First, this ensures that X(X1,Y1)Y form a Markov chain, without conditioning on S. Secondly, as pointed out in Remark 1, there is a difference between the order of the min and max in the lower and upper bounds (cf. (8) and (9)). Thereby, proving the capacity results of Corollary 1 above, we apply the minimax theorem. In general, a pointwise minimum of two convex functions may not necessarily yield a convex function. Nevertheless, having assumed that the relay channel is strongly degraded, the functional G(p,q)=min{Iq(X,X1;Y),I(X;Y1|X1)} is quasi-convex in the state distribution, i.e.,

G(p,(1α)q1+αq2))maxG(p,q1),G(p,q2), (16)

for every pP(X×X1), q1,q2Q, and 0α1. The quasi-convex shape is illustrated in Figure 2, which depicts G(p,q) for an example given in the sequel. By [90] (Theorem 3.4), the minimax theorem applies to quasi-convex functions as well, which alleviates the proof of Corollary 1.

Figure 2.

Figure 2

The functional G(p,q)=min{Iq(X,X1;Y),I(X;Y1|X1)}, for SBernoulli(q), 0q1, as a function of q. The figure corresponds to Example 1, where G(p,q)=min{112h(q),1h(θ)}, for p(x,x1)=p(x)p(x1), with XBernoulli(12) and X1Bernoulli(12), and θ=0.08. Clearly, G(p,q) is not convex in q, but rather quasi-convex in q.

The following corollary is a direct consequence of Lemma 1 and it is significant for the random code analysis of the AVRC.

Corollary 2.

The capacity of the block-compound relay channel LQ×B is bounded by

C(LQ×B)RPDF(LQ), (17)
C(LQ×B)RCS(LQ). (18)

Specifically, if R<RPDF(LQ), then there exists a (2nR,n,ean) block Markov code over LQ×B for sufficiently large n and some a>0.

The proof of Corollary 2 is given in Appendix C.

3.2. The AVRC

We give lower and upper bounds, on the random code capacity and the deterministic code capacity, for the AVRC L.

3.2.1. Random Code Lower and Upper Bounds

The random code bounds below are obtained through a modified version of Ahlswede’s RT, using our results on the block-compound relay channel in Corollary 2. Define

RPDF(L)RPDF(LQ)|Q=P(S),RCS(L)RCS(LQ)|Q=P(S). (19)
Theorem 1.

The random code capacity of an AVRC L is bounded by

RPDF(L)C(L)RCS(L). (20)

The proof of Theorem 1 is given in Appendix D. To prove Theorem 1 we modify Ahlswede’s RT. A straightforward application of Ahlswede’s RT fails to comply with the strictly causal relay transmission. Essentially, the RT uses a reliable code for the compound channel code to construct a random code for the AVC, applying random permutations to the transmitted codeword. However, the relay cannot apply permutations to its transmission, since at time i[1:n], the relay cannot compute f1,j(y1j1), for j>i, as the relay encoder only knows the past received symbols y1,1,,y1,i1, and does not have access to the symbols y1,i,,y1,j1 which will be received in the future. To resolve this difficulty, we use a block Markov code for the block compound channel. In a block Markov coding scheme, the relay sends x1,bn in block b, using the sequence of symbols y1,b1n received in the previous block. Since the entire sequence y1,b1n is known to the relay encoder, permutations can be applied to the transmission in each block separately. Hence, our proof exploits the structure of the original block-compound channel code to construct a random code for the AVRC, as opposed to classical works where the RT is used such that the original code is treated as a “black box” [59].

Remark 3.

Block Markov coding with partial decode-forward is not a simple scheme by itself, and thus, using the RT requires careful attention. In particular, by close inspection of the proof of Theorem 1, one may recognize that the necessity of using the block-compound relay channel, rather than the standard compound channel, stems from the fact that for the AVRC, the state sequences may have completely different types in each block. For each block, we use the RT twice. First, the RT is applied to the probability of the backward decoding error, for the message component which is decoded by the relay. Then, it is applied to the probability of forward decoding error, for the message component which is transmitted directly.

Together with Corollary 1, the theorem above yields another corollary.

Corollary 3.

Let L be an AVRC.

  • 1. 
    If WY,Y1|X,X1,S is strongly reversely degraded,
    C(L)=RPDF(L)=RCS(L)=minq(s)maxp(x,x1)Iq(X;Y|X1). (21)
  • 2. 
    If WY,Y1|X,X1,S is strongly degraded,
    C(L)=RPDF(L)=RCS(L)=maxp(x,x1)minminq(s)Iq(X,X1;Y),I(X;Y1|X1). (22)

Before we proceed to the deterministic code capacity, we note that Ahlswede’s Elimination Technique [55] can be applied to the AVRC as well. Hence, the size of the code collection of any reliable random code can be reduced to polynomial size.

3.2.2. Deterministic Code Lower and Upper Bounds

In the next statements, we characterize the deterministic code capacity of the AVRC L. We consider conditions under which the deterministic code capacity is positive, and it coincides with the random code capacity, and conditions under which it is lower. For every x1X1, let W1(x1) and W(x1) denote the marginal AVCs from the sender to the relay and from the sender to the destination receiver, respectively,

W1(x1)={WY1|X,X1,S(·|·,x1,·)},W(x1)={WY|X,X1,S(·|·,x1,·)}. (23)

See Figure 3.

Figure 3.

Figure 3

The marginals of the arbitrarily varying relay channel. For every relay transmission x1X1, the marginal sender-relay AVC is denoted by W1(x1)={WY1|X,X1,S(·|·,x1,·)}, and the marginal sender-receiver AVC is denoted by W(x1)={WY|X,X1,S(·|·,x1,·). A sufficient condition, under which the deterministic code capacity is the same as the random code capacity of the AVRC, is given in Lemma 2. This condition is also a sufficient condition for positive capacity, but as explained in Remark 4, it is not a necessary condition.

Lemma 2 gives a condition under which the deterministic code capacity is the same as the random code capacity. The condition is given in terms of the marginal AVCs W1(x1) and W(x1).

Lemma 2.

If the marginal sender-relay and sender-reciever AVCs have positive capacities, i.e., C(W1(x1,1))>0 and C(W(x1,2))>0, for some x1,1,x1,2X1, then the capacity of the AVRC L is positive, and it coincides with the random code capacity, i.e., C(L)=C(L)>0.

The proof of Lemma 2 is given in Appendix E, extending Ahlswede’s Elimination Technique [55].

Next, we give a computable sufficient condition, under which the deterministic code capacity coincides with the random code capacity. For the point to point AVC, this occurs if and only if the channel is non-symmetrizable [56,57] (Definition 2). Our condition here is given in terms of an extended definition of symmetrizability, akin to [67] (Definition 3.2).

Definition 3.

A state-dependent relay channel WY,Y1|X,X1,S is said to be symmetrizable-X|X1 if for some conditional distribution J(s|x),

sSWY,Y1|X,X1,S(y,y1|x,x1,s)J(s|x˜)=sSWY,Y1|X,X1,S(y,y1|x˜,x1,s)J(s|x),x,x˜X,x1X1,yY,y1Y1. (24)

Equivalently, for every given x1X1, the channel WY¯|X,X1,S(·|·,x1,·) is symmetrizable, where Y¯=(Y,Y1).

A similar definition applies to the marginals WY|X,X1,S and WY1|X,X1,S. Note that symmetrizability of each of these marginals can be checked, without reference to whether the channel is degraded or strongly degraded.

Corollary 4.

Let L be an AVRC.

  • 1. 
    If WY|X,X1,S and WY1|X,X1,S are non-symmetrizable-X|X1, then C(L)=C(L)>0. In this case,
    RPDF(L)C(L)RCS(L). (25)
  • 2. 
    If WY,Y1|X,X1,S is strongly reversely degraded, where WY1|X,X1,S is non-symmetrizable-X|X1, then
    C(L)=C(L)=RPDF(L)=RCS(L)=minq(s)maxp(x,x1)Iq(X;Y|X1). (26)
  • 3. 
    If WY,Y1|X,X1,S is strongly degraded, where WY|X,X1,S is non-symmetrizable-X|X1 and WY1|X,X1(y1|x,x1)WY1|X,X1(y1|x˜,x1) for some x,x˜X, x1X1 and y1Y1, then
    C(L)=C(L)=RPDF(L)=RCS(L)=maxp(x,x1)minminq(s)Iq(X,X1;Y),I(X;Y1|X1). (27)

The proof of Corollary 4 is given in Appendix F.

Remark 4.

By Corollary 4, we have that non-symmetrizability of the marginal AVCs, W1(x1,1) and W(x1,2), for some x1,1,x1,2X1, is a sufficient condition for positive capacity (see Figure 3). This raises the question whether it is a necessary condition as well. In other words: If W1(x1) and W(x1) are symmetrizable for all x1X1, does that necessarily imply that the capacity is zero? The answer is no. We show this using a very simple example. Suppose that Y1=S and Y=(X1,X+S), where all variables are binary. It is readily seen that for both Y1 and Y, the input and the state are symmetric, for every given X1=x1. Hence, W1(x1) and W(x1) are symmetrizable for all x1X1. Nevertheless, we note that since the relay can send X1=Y1=S, this is equivalent to an AVC with state information at the decoder. As the decoder can use X1 to eliminate the state, the capacity of this AVRC is C(L)=1. In Lemma 3 below, we give a stronger condition which is a necessary condition for positive capacity.

Remark 5.

Note that there are 4 symmetrizability cases in terms of the sender-relay channel WY1|X,X1,S and the sender-receiver channel WY|X,X1,S. For the case where WY1|X,X1,S and WY|X,X1,S are both non-symmetrizable-X|X1, the lemma above asserts that the capacity coincides with the random code capacity. In other cases, one may expect the capacity to be lower than the random code capacity. For instance, if WY|X,X1,S is non-symmetrizable-X|X1, while WY1|X,X1,S is symmetrizable-X|X1, then the capacity is positive by direct transmission. Furthermore, in this case, if the channel is reversely degraded, then the capacity coincides with the random code capacity. However, it remains in question whether this is true in general, when the channel is not reversely degraded.

Next, we consider conditions under which the capacity is zero. Observe that if WY,Y1|X,X1,S is symmetrizable-X|X1 then so are WY|X,X1,S and WY1|X,X1,S. Intuitively, if the AVRC is symmetrizable-X|X1, then it is a poor channel. For example, say Y1=X+X1+S and Y=X·X1·S, with S=X. Then, the jammer can confuse the decoder by taking the state sequence Sn to be some codeword. The following lemma validates this intuition.

Lemma 3.

If the AVRC L is symmetrizable-X|X1, then it has zero capacity, i.e., C(L)=0. Equivalently, non-symmetrizability-X|X1 of the AVRC L is a necessary condition for positive capacity.

Lemma 3 is proved in Appendix G, using an extended version of Ericson’s technique [56]. For a strongly degraded AVRC, we have a simpler symmetrizability condition under which the capacity is zero.

Definition 4.

Let WY,Y1|X,X1,S=WY1|X,X1WY|Y1,X1,S be a strongly degraded relay channel. We say that WY,Y1|X,X1,S is symmetrizable-X1×Y1 if for some conditional distribution J(s|x1,y1),

sSWY|Y1,X1,S(y|y1,x1,s)J(s|x˜1,y˜1)=sSWY|Y1,X1,S(y|y˜1,x˜1,s)J(s|x1,y1),x˜1,x1X1,yY,y1,y˜1Y1. (28)

Equivalently, the channel WY|Y¯1,S is symmetrizable, where Y¯1=(Y1,X1).

Lemma 4.

If the AVRC L is strongly degraded and symmetrizable-X1×Y1, then it has zero capacity, i.e., C(L)=0.

Lemma 4 is proved in Appendix H. An example is given below.

Example 1.

Consider a state-dependent relay channel WY,Y1|X,X1,S, specified by

Y1=X+Zmod2,Y=X1+S,

where X=X1=Z=S=Y1={0,1} and Y={0,1,2}, and the additive noise is distributed according to ZBernoulli(θ), 0θ1. It is readily seen that WY,Y1|X,X1,S is strongly degraded and symmetrizable-X1×Y1, by (2) and (28). In particular, (28) is satisfied with J(s|x1,y1)=1 for s=x1, and J(s|x1,y1)=0 otherwise. Hence, by Lemma 4, the capacity is C(L)=0. On the other hand, we show that the random code capacity is given by C(L)=min12,1h(θ), using Corollary 3. The derivation of the random code capacity is given in Appendix I.

3.3. AVRC with Orthogonal Sender Components

Consider the special case of a relay channel WY,Y1|X,X1,S with orthogonal sender components [5]; [3] (Section 16.6.2), where X=(X,X) and

WY,Y1|X,X,X1,S(y,y1|x,x,x1,s)=WY|X,X1,S(y|x,x1,s)·WY1|X,X1,S(y1|x,x1,s). (29)

Here, we address the case where the channel output depends on the state only through the relay, i.e., WY|X,X1,S(y|x,x1,s)=WY|X,X1(y|x,x1).

Lemma 5.

Let L = {WY|X,X1WY1|X,X1,S} be an AVRC with orthogonal sender components. The random code capacity of L is given by

C(L)=RPDF(L)=RCS(L)=maxp(x1)p(x|x1)p(x|x1)minI(X,X1;Y),minq(s)Iq(X;Y1|X1)+I(X;Y|X1). (30)

If WY1|X,X1,S is non-symmetrizable-X|X1, and WY|X,X1(y|x,x1)WY|X,X1(y|x˜,x1) for some x1X1, x,x˜X, yY, then the deterministic code capacity is given by C(L)=RPDF(L)=RCS(L).

The proof of Lemma 5 is given in Appendix J. To prove Lemma 5, we apply the methods of [5] to our results. Specifically, we use the partial decode-forward lower bound in Theorem 1, taking U=X (see (9) and (19)).

4. Gaussian AVRC with Sender Frequency Division

We give extended results for the Gaussian AVRC with sender frequency division (SFD), which is a special case of the AVRC with orthogonal sender components [5]. We determine the random code capacity of the Gaussian AVRC with SFD, and give lower and upper bounds on the deterministic code capacity. The derivation of the deterministic code bounds is mostly independent of our previous results, and it is based on the technique by [87]. The Gaussian relay channel WY,Y1|X,X1,S with SFD is a special case of a relay channel with orthogonal sender components [5], specified by

Y1=X+Z,Y=X+X1+S, (31)

where the Gaussian additive noise ZN(0,σ2) is independent of the channel state. As opposed to Lemma 5, the main channel here depends on the state, while the channel to the relay does not. In the case of a Gaussian channel, power limitations need to be accounted for, and thus, we consider the Gaussian relay channel under input and state constraints. Specifically, the user and the relay’s transmission are subject to input constraints Ω>0 and Ω1>0, respectively, and the jammer is under a state constraint Λ, i.e.,

1ni=1n(Xi2+Xi2)Ω,1ni=1nX1,i2Ω1w.p.1,1ni=1nSi2Λw.p.1. (32)

We note that Ahlswede’s Elimination Technique cannot be used under a state constraint (see [57]). Indeed, if the jammer concentrates a lot of power on the shared randomness transmission, then this transmission needs to be robust against a state constraint that is higher than Λ. Thereby, the results given in Section 3.2.2 do not apply to the Gaussian AVRC under input and state constraints.

For the compound relay channel, the state constraint is in the average sense. That is, we say that the Gaussian compound relay channel LQ with SFD is under input constraints Ω and Ω1 and state constraint Λ if

1ni=1n(Xi2+Xi2)Ω,1ni=1nX1,i2Ω1,w.p.1,Q={q(s):ES2Λ}. (33)

Coding definitions and notation are as follows. The definition of a code is similar to that of Section 2.3. The encoding function is denoted by f=(f,f), with f:[1:2nR]Rn and f:[1:2nR]Rn, and the relay encoding function is denoted by f1:RnRn, where f1,i:Ri1R, for i[1:n]. The boldface notation indicates that the encoding functions produce sequences. Here, the encoder and the relay satisfy the input constraints f(m)2+f(m)2nΩ and f1(y1)2nΩ1 for all m[1:2nR] and y1Rn. At time i[1:n], given a message m[1:2nR], the encoder transmits (xi,xi)=(fi(m),fi(m)), and the relay transmits x1,i=f1,i(y1,1,,y1,i1). The decoder receives the output sequence y, and finds an estimate m^=g(y). A (2nR,n,ε) code C for the Gaussian AVRC satisfies Pe|s(n)(C)ε, for all sRn with s2nΛ, where

Pe|s(n)(C)=12nRm=12nRD(m,s)c1(2πσ2)n/2ez2/2σ2dz, (34)

with

D(m,s)=zRn:gf(m)+f1f(m)+z+s=m. (35)

Achivable rates, deterministic code capacity and random code capacity are defined as before. Next, we give our results on the Gaussian compound relay channel and the Gaussian AVRC with SFD.

5. Main Results—Gaussian AVRC with SFD

We give our results on the Gaussian compound and AVRC with SFD. The results on this compound relay channel and on the random code capacity of this AVRC are obtained through a straightforward extension of our previous results and derivations. However, the derivation of the deterministic code bounds is mostly independent of our previous results, and it is based on modifying the technique by Csiszär and Narayan in their paper on the Gaussian AVC [87].

5.1. Gaussian Compound Relay Channel

We determine the capacity of the Gaussian compound relay channel with SFD under input and state constraints. Let

FG(α,ρ)min{12log1+Ω1+αΩ+2ραΩΩ1Λ,12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛ}. (36)

Lemma 6.

The capacity of the Gaussian compound relay channel with SFD, under input constraints Ω and Ω1 and state constraint Λ, is given by

C(LQ)=max0α,ρ1FG(α,ρ), (37)

and it is identical to the random code capacity, i.e., C(LQ)=C(LQ).

The proof of Lemma 6 is given in Appendix K, based on our results in the previous sections. The parameter 0α1 represents the fraction of input power invested in the transmission of the message component which is decoded by the relay, in the partial decode-forward coding scheme. Specifically, in the achievability proof in [5], αΩ and (1α)Ω are the variances of X and X, respectively. The parameter ρ stands for the correlation coefficient between the decode-forward transmission X and the relay transmission X1.

5.2. Gaussian AVRC

We determine the random code capacity of the Gaussian AVRC with SFD under constraints.

Theorem 2.

The random code capacity of the Gaussian AVRC with SFD, under input constraints Ω and Ω1 and state constraint Λ, is given by

C(L)=C(LQ)=max0α,ρ1FG(α,ρ). (38)

The proof of Theorem 2 is given in Appendix L. The proof follows the same considerations as in our previous results.

Next, we give lower and upper bounds on the deterministic code capacity of the Gaussian AVRC with SFD under constraints, obtained by generalizing the non-standard techniques by Csiszár and Narayan in their 1991 paper on the Gaussian AVC [87]. Define

RG,low(L)maxFG(α,ρ)subjectto0α,ρ1,(1ρ2)αΩ>Λ,Ω1Ω(Ω1+ραΩ)2>Λ+(1ρ2)αΩ. (39)

and

RG,up(L)maxFG(α,ρ)subjectto0α,ρ1,Ω1+αΩ+2ραΩ·Ω1Λ. (40)

It can be seen that RG,lowRG,up, since

Ω1+αΩ+2ραΩ·Ω1=(1ρ2)αΩ+(Ω1+ραΩ)2(1ρ2)αΩ. (41)

The analysis is based on the following lemma by [87].

Lemma 7

(see [87] (Lemma 1)). For every ε>0, 8ε<η<1, K>2ε, and M=2nR, with 2εRK, and nn0(ε,η,K), there exist M unit vectors a(m)Rn, m[1:M], such that for every unit vector cRn and 0θ,ζ1,

|m˜[1:M]:a(m˜),cθ|2n[R+12log(1θ2)]++ε, (42)

and if θη and θ2+ζ2>1+η22R, then

1M|m[1:M]:|a(m˜),a(m)|θ,|a(m˜),c|ζ,forsomem˜m|2nε, (43)

where [t]+=max{0,t} and ·,· denotes inner product.

Intuitively, the lemma states that under certain conditions, a codebook can be constructed with an exponentially small fraction of “bad” messages, for which the codewords are non-orthogonal to each other and the state sequence.

Theorem 3.

The deterministic code capacity of the Gaussian AVRC with SFD, under input constraints Ω and Ω1 and state constraint Λ, is bounded by

RG,low(L)C(L)RG,up(L). (44)

The proof of Theorem 3 is given in Appendix M.

Remark 6.

Csiszár and Narayan [87] have shown that for the classical Gaussian AVC, reliable decoding is guaranteed when the input constraint Ω is larger than the state constraint Λ. Here, we use a partial decode-forward coding scheme, where the message has two components, one which is decoded by the relay, and the other is transmitted directly. The respective optimization constraints Ω1Ω(Ω1+ραΩ)2>Λ+(1ρ2)αΩ and (1ρ2)αΩ>Λ in the RHS of (39), guarantee reliability for each decoding step.

Remark 7.

Csiszár and Narayan [87] have further shown that for the classical Gaussian AVC, if ΩΛ, the capacity is zero. The converse proof in [87] follows by considering a jammer who chooses the state sequence to be a codeword. Due to the symmetry between X and S, the decoder cannot distinguish between the transmitted codeword and the impostor sent by the jammer. Here, we consider a jammer who simulates X+X1. Specifically, The jammer draws a codeword X=f(m˜) uniformly at random, and then, generates a sequence Y˜1 distributed according to the conditional distribution PY1|M=m˜. If the sequence S˜=f(m˜)+f1(Y˜1) satisfies the state constraint Λ, then the jammer chooses S˜ as the state sequence. Defining αΩ, Ω1, and ρ as the empirical decode-forward transmission power, relay transmission power, and their correlation coefficient, respectively, we have that the state constraint S˜2nΛ holds with high probability, if Ω1+αΩ+2ραΩ·Ω1<Λ. The details are in Appendix M.

Figure 4 depicts the bounds on the capacity of the Gaussian AVRC with SFD under input and state constraints, as a function of the input constraint Ω=Ω1, under state constraint Λ=1 and σ2=0.5. The top dashed line depicts the random code capacity of the Gaussian AVRC. The solid lines depict the deterministic code lower and upper bounds RG,low(L) and RG,up(L). For low values, Ω<Λ4=0.25, we have that RG,up(L)=0, hence the deterministic code capacity is zero, and it is strictly lower than the random code capacity. The dotted lower line depicts the direct transmission lower bound, which is FG(1,0) for Ω>Λ, and zero otherwise [57]. For intermediate values of Ω, direct transmission is better than the lower bound in Theorem 3. Whereas, for high values of Ω, the optimization constraints in (39) and (40) are inactive, hence, our bounds are tight, and the capacity coincides with the random code capacity, i.e., C(L)=C(L)=RG,low(L)=RG,up(L).

Figure 4.

Figure 4

Bounds on the capacity of the Gaussian AVRC with sender frequency division. The dashed upper line depicts the random code capacity of the Gaussian AVRC as a function of the input constraint Ω=Ω1, under state constraint Λ=1 and σ2=0.5. The solid lines depict the deterministic code lower and upper bounds RG,low(L) and RG,up(L). The dotted lower line depicts the direct transmission lower bound.

6. The Primitive AVRC

In this section, we give our results on the primitive AVRC [2], and then consider the Gaussian case. Part of the motivation given in [2] to consider the primitive relay channel was that the overall behavior and properties are the same as the non primitive (“regular”) relay channel. We show that this is not true in the arbitrarily varying scenario. In particular, the behavior of the primitive Gaussian AVRC with SFD is different compared to the non-primitive counterpart considered above.

6.1. Definitions and Notation

Consider a setup where the sender transmits information over state-dependent memoryless relay channel WY,Y1|X,S, while there is a noiseless link of capacity C1>0 between the relay and the receiver. Communication over a primitive relay channel is depicted in Figure 5. Given a message M[1:2nR], the encoder transmits Xn=f(M) over the channel WY,Y1|X,S, which is referred to as the primitive relay channel. The relay receives Y1n and sends an index L=f1(Y1n) to the receiver, where f1:Y1n[1:2nC1]. The decoder receives both the channel output sequence Yn and the relay output L, and finds an estimate of the message M^=g(Yn,L). In accordance with the previous definitions, the primitive AVRC Lprim={WY,Y1|X,S} has a state sequence of unknown distribution, not necessarily independent nor stationary. The deterministic code capacity and the random code capacity are defined as before, and denoted by C(Lprim) and C(Lprim), respectively.

Figure 5.

Figure 5

Communication over the primitive AVRC L. Given a message M, the encoder transmits Xn=f(M). The relay receives Y1n and sends L=f1(Y1n), where f1:Y1n[1:2nC1]. The decoder receives both the channel output sequence Yn and the relay output L, and finds an estimate of the message M^=g(Yn,L).

6.2. Main Results—Primitive AVRC

We give our results on the primitive AVRC below. However, since the proofs are based on the same arguments as given for the non primitive AVRC, we omit the proofs of the results in this section. The details are given in [91].

Using similar arguments to those given for the non primitive relay channel, we obtain the following bounds on the random code capacity,

RCSminq(s)maxp(x)minIq(X;Y)+C1,Iq(X;Y,Y1), (45)

and

RPDFmaxp(u,x)minminq(s)Iq(U;Y)+minq(s)Iq(X;Y|U)+C1,minq(s)Iq(U;Y1)+minq(s)Iq(X;Y|U). (46)

Theorem 4.

The random code capacity of a primitive AVRC Lprim is bounded by

RPDFC(Lprim)RCS. (47)

Those bounds have the same form as the cutset upper bound and the partical decode-forward lower bound in Section 3 (cf. (8), (9) and (45), (46)). As in Section 3, we can use the bounds above to determine the capacity in the strongly degraded and reversely degraded cases, based on the direct transmission lower bound (for U=), and the full decode-forward lower bound (for U=X).

Corollary 5.

Let Lprim be a primitive AVRC.

  • 1. 
    If WY,Y1|X,S is strongly reversely degraded, i.e., WY,Y1|X,S=WY|X,SWY1|Y, then
    C(Lprim)=minq(s)maxp(x)Iq(X;Y). (48)
  • 2. 
    If WY,Y1|X,X1,S is strongly degraded, i.e., WY,Y1|X,X1,S=WY1|XWY|Y1,S, then
    C(Lprim)=maxp(x)minminq(s)Iq(X;Y)+C1,I(X;Y1). (49)

As for the deterministic code capacity, we give the following theorem.

Theorem 5.

Let Lprim be a primitive AVRC.

  • 1. 
    If WY1|X,S is non-symmetrizable, then C(Lprim)=C(Lprim). In this case,
    RPDFC(Lprim)RCS. (50)
  • 2. 
    If WY,Y1|X,S is strongly reversely degraded, where WY1|X,S is non-symmetrizable, then
    C(Lprim)=minq(s)maxp(x)Iq(X;Y). (51)
  • 3. 
    If WY,Y1|X,S is strongly degraded, such that WY1|X(y1|x)WY1|X(y1|x˜) for some x,x˜X, y1Y1, then
    C(Lprim)=maxp(x)minminq(s)Iq(X;Y)+C1,I(X;Y1). (52)
  • 4. 

    If WY˜|X,S is symmetrizable, where Y˜=(Y,Y1), then C(Lprim)=0.

The proof of Theorem 5 is available in [91]. To illustrate our results, we give the following example of a primitive AVRC.

Example 2.

Consider a state-dependent primitive relay channel WY,Y1|X,S, specified by

Y1=X(1S),Y=X+S,

where X=S=Y1={0,1}, Y={0,1,2}, and C1=1, i.e., the link between the relay and the receiver is a noiseless bit pipe. It can be seen that both the sender-relay and the sender-receiver marginals are symmetrizable. Indeed, WY|X,S satisfies

sSWY|X,S(y1|x,s)J(s|x˜)=sSWY|X,S(y1|x˜,s)J(s|x),x,x˜X,yY, (53)

with J(s|x)=1 for s=x, and J(s|x)=0 otherwise, while WY1|X,S satisfies (53) with J(s|x)=1 for s=1x, and J(s|x)=0 otherwise. Nevertheless, the capacity of the primitive AVRC Lprim={WY,Y1|X,S} is C(Lprim)=1, which can be achieved using a code of length n=1, with f(m)=m, f1(y1)=y1,

g(y,)=g(y,y1)=0y=01y=2y1y=1 (54)

for m,y1{0,1} and y{0,1,2}. This example shows that even if the sender-relay and sender-receiver marginals are symmetrizable, the capacity may still be positive. We further note that the condition in part 4 of Theorem 5 implies that WY|X,S and WY1|X,S are both symmetrizable, but not vice versa, as shown by this example. That is, as the capacity is positive, we have that WY˜|X,S is non-symmetrizable, where Y˜=(Y,Y1), despite the fact that the marginals WY|X,S and WY1|X,S are both symmetrizable.

6.3. Primitive Gaussian AVRC

Consider the primitive Gaussian relay channel with SFD,

Y1=X+Z,Y=X+S, (55)

Suppose that input and state constraints are imposed as before, i.e., 1ni=1n(Xi2+Xi2)Ω and 1ni=1nSi2Λ with probability 1. The capacity of the primitive Gaussian AVRC with SFD, under input constraint Ω and state constraint Λ is given by

C(Lprim)=C(Lprim)=max0α112log1+αΩΛ+minC1,12log1+(1α)ΩΛ. (56)

This result is due to the following. Observe that one could treat this primitive AVRC as two independent channels, one from X to Y and the other from X to Y1, dividing the input power to αΩ and (1α)Ω, respectively. Based on this observation, the random code direct part follows from [92]. Next, the deterministic code direct part follows from part 1 of Theorem 5, and the converse part follows straightforwardly from the cutset upper bound in Theorem 4.

7. Discussion

We have presented the model of the arbitrarily varying relay channel (AVRC), as a state dependent relay channel, where jamming attacks result in either a random or a deterministic state sequence, Snq(sn), where the joint distribution q(sn) is unknown and it is not necessarily of a product form. We have established the cutset upper bound and the partial decode-forward lower bound on the random code capacity of the AVRC. We have determined the random code capacity in special cases of the degraded AVRC, the reversely degraded AVRC, and the AVRC with orthogonal sender components. To do so, we used the direct transmission lower bound and the full decode-forward lower bound, along with quasi-convexity properties which are required in order to use the minimax theorem.

We have provided generalized symmetrizability conditions under which the deterministic code capacity coincides with the random code capacity. Specifically, we have shown that if the sender-relay and sender-receiver marginals are non-symmetrizable for a given relay transmission, then the capacity is positive. We further noted that this is a sufficient condition for positive capacity, which raises the question whether it is also a necessary condition. In other words, if those marginals are symmetrizable for every given relay transmission, does that necessarily imply that the capacity is zero? The answer is no, and we have refuted this assertion using a simple example, where the relay acts as a source of state information to the receiver. Then, we provided a stronger symmetrizability condition, which is necessary for the capacity to be positive. We have shown by example that the deterministic code capacity can be strictly lower than the random code capacity of the AVRC.

The Gaussian AVRC with sender frequency division (SFD) under input and state constraints is also addressed in this paper. The random code capacity is determined using the above results, whereas the deterministic code capacity is lower and upper bounded using an independent approach. Specifically, we extended the technique by Csiszár and Narayan in their 1991 paper on the Gaussian AVC [87]. We have shown that the deterministic code capacity can be strictly lower than the random code capacity, for low values on the input constraint.

Furthermore, we have considered the primitive AVRC, where there is a noiseless link between the relay and the receiver of limited capacity [2]. We tested Kim’s assertion that “the primitive relay channel captures most essential features and challenges of relaying, and thus serves as a good testbed for new relay coding techniques” [2]. We have shown that this assertion is not true in the arbitrarily varying scenario. Specifically, for the primitive Gaussian AVRC with SFD, the deterministic code capacity and the random code capacity are always the same, regardless of the value of the input constraint (see (56)), in contrast to our findings for the non primitive case, as demonstrated in Figure 4.

Abbreviations

The following abbreviations are used in this manuscript:

AVC Arbitrarily varying channel
AVRC Arbitrarily varying relay channel
DMC Discrete memoryless channel
pmf probability mass function
RT Robustification technique
SFD Sender frequency division
Eq. Equation
RHS Right hand side
LHS Left hand side

Appendix A. Proof of Lemma 1

Appendix A.1. Partial Decode-Forward Lower Bound

We construct a block Markov code using the partial decode-forward scheme. That is, the encoder sends a sequence of messages over multiple blocks. The message in each block consists of two components, a decode-forward component, and a direct transmission component, where only the former is decoded by the relay. Once the decoder has received all blocks, the decode-forward components are decoded backwards, i.e., starting with the message in the last block going backwards. Using the estimation of the decode-forward components, the direct transmission components are decoded forwards, i.e., starting with the message in the first block going forwards. The ambiguity of the state distribution needs to be treated throughout all of those estimations. Hence, we use joint typicality with respect to a state type, which is “close” to some qQ. Let δ>0 be arbitrarily small. Define a set of state types Q^n by

Q^n={P^sn:snAδ1(q)forsomeqQ}, (A1)

where

δ1δ2·|S|. (A2)

Namely, Q^n is the set of types that are δ1-close to some state distribution q(s) in Q. A code C for the compound relay channel is constructed as follows.

The encoders use B blocks, each consists of n channel uses to convey (B1) independent messages to the receiver. Furthermore, each message Mb, for b[1:B1], is divided into two independent messages. That is, Mb=(Mb,Mb), where Mb and Mb are uniformly distributed, i.e.,

MbUnif[1:2nR],MbUnif[1:2nR],withR+R=R, (A3)

for b[1:B1]. For convenience of notation, set M0=MB1 and M0=MB1. The average rate B1B·R is arbitrarily close to R.

Codebook Generation: Fix the distribution PU,X,X1(u,x,x1), and let

PX,Y,Y1|U,X1q(x,y,y1|u,x1)=PX|U,X1(x|u,x1)sSq(s)WY,Y1|X,X1,S(y,y1|x,x1,s). (A4)

We construct B independent codebooks. For b[2:B1], generate 2nR independent sequences x1,bn(mb1), mb1[1:2nR], at random, each according to i=1nPX1(x1,i). Then, generate 2nR sequences,

ubn(mb|mb1)i=1nPU|X1(ui|x1,b,i(mb1)),mb[1:2nR], (A5)

conditionally independent given x1,bn(mb1). Then, for every mb[1:2nR], generate 2nR sequences,

xbn(mb,mb|mb1)i=1nPX|U,X1(xi|ub,i(mb|mb1),x1,b,i(mb1)),mb[1:2nR], (A6)

conditionally independent given (ubn(mb|mb1),x1,bn(mb1)). We have thus generated B2 independent codebooks,

Fb=x1,bn(mb1),ubn(mb|mb1),xbn(mb,mb|mb1):mb1,mb[1:2nR],mb[1:2nR], (A7)

for b[2:B1]. The codebooks F1 and FB are generated in the same manner, with fixed m0=mB1 and m0=mB1. Encoding and decoding is illustrated in Figure A1.

Encoding: To send the message sequence (m1,m1,,mB1,mB1), transmit xbn(mb,mb|mb1) at block b, for b[1:B].

Relay Encoding: In block 1, the relay transmits x1,1n(1). Set m˜01. At the end of block b[1:B1], the relay receives y1,bn, and finds some m˜b[1:2nR] such that

(ubn(m˜b|m˜b1),x1,bn(m˜b1),y1,bn)Aδ(PU,X1PY1|U,X1q),forsomeqQ^n. (A8)

If there is none or there is more than one such, set m˜b=1. In block b+1, the relay transmits x1,b+1n(m˜b).

Backward Decoding: Once all blocks (ybn)b=1B are received, decoding is performed backwards. Set m^B=m^B1. For b=B1,B2,,1, find a unique m^b[1:2nR] such that

(ub+1n(m^b+1|m^b),x1,b+1n(m^b),yb+1n)Aδ(PU,X1PY|U,X1q),forsomeqQ^n. (A9)

If there is none, or more than one such m^b[1:2nR], declare an error.

Then, the decoder uses m^1,,m^B1 as follows. For b=B1,B2,,1, find a unique m^b[1:2nR] such that

(ubn(m^b|m^b1),xbn(m^b,m^b|m^b1),x1,b(m^b1),ybn)Aδ(PU,X,X1PY|X,X1q),forsomeqQ^n. (A10)

If there is none, or more than one such m^b[1:2nR], declare an error. We note that using the set of types Q^n instead of the original set of state distributions Q alleviates the analysis, since Q is not necessarily finite nor countable.

Figure A1.

Figure A1

The partial decode-forward coding scheme. The block index b[1:B] is indicated at the top. In the following rows, we have the corresponding elements: (1) sequences transmitted by the encoder; (2) estimated messages at the relay; (3) sequences transmitted by the relay; (4) estimated messages at the destination decoder. The arrows in the second row indicate that the relay encodes forwards with respect to the block index, while the arrows in the fourth row indicate that the receiver decodes backwards.

Analysis of Probability of Error: Assume without loss of generality that the user sent (Mb,Mb)=(1,1), and let q*(s)Q denote the actual state distribution chosen by the jammer. The error event is bounded by the union of the events

E1(b)={M˜b1},E2(b)={M^b1},E3(b)={M^b1},forb[1:B1]. (A11)

Then, the probability of error is bounded by

Pe(n)(q,C)b=1B1PrE1(b)+b=1B1PrE2(b)E1c(b)+b=1B1PrE3(b)E1c(b)E2c(b)E2c(b1), (A12)

with E2(0)=, where the conditioning on (Mb,Mb)=(1,1) is omitted for convenience of notation.

We begin with the probability of erroneous relaying, PrE1(b). Define

E1,1(b)={(Ubn(1|M˜b1),X1,bn(M˜b1),Y1,bn)Aδ(PU,X1PY1|U,X1q)forallqQ^n}E1,2(b)={(Ubn(mb|M˜b1),X1,bn(M˜b1),Y1,bn)Aδ(PU,X1PY1|U,X1q),forsomemb1,qQ^n}. (A13)

For b[1:B1], the relay error event is bounded as

E1(b)E1(b1)E1,1(b)E1,2(b)=E1(b1)E1(b1)cE1,1(b)E1(b1)cE1,2(b), (A14)

with E1(0)=. Thus, by the union of events bound,

PrE1(b)PrE1(b1)+PrE1,1(b)E1(b1)c+PrE1,2(b)E1(b1)c. (A15)

Consider the second term on the RHS of (A15). We now claim that given that E1(b1)c occurred, i.e., M˜b1=1, the event E1,1(b) implies that (Ubn(1|1),X1,bn(1),Y1,bn)Aδ/2(PU,X1PY1|U,X1q) for all qQ. This claim is due to the following. Assume to the contrary that E1,1(b) holds, but (Ubn(1|1),X1,bn(1),Y1,bn)Aδ/2(PU,X1PY1|U,X1q) for some qQ. Then, for a sufficiently large n, there exists a type q(s) such that

|q(s)q(s)|δ1, (A16)

for all sS, and by the definition in (A1), qQ^n. Then, (A16) implies that

|PY1|U,X1q(y1|u,x1)PY1|U,X1q(y1|u,x1)||S|·δ1=δ2, (A17)

for all uU, x1X1 and y1Y1 (see (A4) and (A2)). Hence, (Ubn(1|1),X1,bn(1),Y1,bn)Aδ(PU,X1PY1|U,X1q), which contradicts the first assumption. It follows that

PrE1,1(b)E1(b1)cPr(Ubn(1|1),X1,bn(1),Y1,bn)Aδ/2(PU,X1PY1|U,X1q)forallqQE1(b1)cPr(Ubn(1|1),X1,bn(1),Y1,bn)Aδ/2(PU,X1PY1|U,X1q*)E1(b1)c. (A18)

Since the codebooks F1,,FB are independent, the sequence (Ubn(1|1),X1,bn(1)) from the codebook Fb is independent of the relay estimate M˜b1, which is a function of Y1,b1n and the codebook Fb1. Thus, the RHS of (A18) tends to zero exponentially as n by the law of large numbers and Chernoff’s bound.

We move to the third term in the RHS of (A15). By the union of events bound, the fact that the number of type classes in Sn is bounded by (n+1)|S|, and the independence of the codebooks, we have that

PrE1,2(b)E1(b1)c(n+1)|S|·supqQ^nPr(Ubn(mb|1),X1,bn(1),Y1,bn)Aδ(PU,X1PY1|U,X1q)forsomemb1(n+1)|S|·2nR·supqQ^nun,x1nPUn,X1n(un,x1n)·y1n:(un,x1n,y1n)Aδ(PU,X1PY1|U,X1q)PY1n|X1nq*(y1n|x1n), (A19)

where the last line follows since Ubn(mb|1) is conditionally independent of Y1,bn given X1,bn(1), for every mb1. Let y1n satisfy (un,x1n,y1n)Aδ(PU,X1PY1|U,X1q). Then, (x1n,y1n)Aδ2(PX1,Y1q) with δ2|U|·δ. By Lemmas 2.6 and 2.7 in [93],

PX1n,Y1nq*(x1n,y1n)=2nH(P^x1n,y1n)+D(P^x1n,y1n||PX1,Y1q*)2nH(P^x1n,y1n)2nHq(X1,Y1)ε1(δ),

hence,

PY1n|X1nq*(y1n|x1n)2nHq(Y1|X1)ε2(δ), (A20)

where ε1(δ),ε2(δ)0 as δ0. Therefore, by Equation (A19)−(A20), along with [93] (Lemma 2.13),

PrE1,2(b)E1(b1)c(n+1)|S|·supqQ2n[Iq(U;Y1|X1)Rε3(δ)], (A21)

with ε3(δ)0 as δ0. Using induction, we have by (A15) that PrE1(b) tends to zero exponentially as n, for b[1:B1], provided that R<infqQIq(U;Y1|X1)ε3(δ).

As for the erroneous decoding of Mb at the receiver, observe that given E1(b)c, the relay sends X1,bn(1) in block b+1, hence

(Ub+1n(1|1),Xb+1n(1,1|1),X1,b+1n(1))PU,X,X1(u,x,x1). (A22)

At the destination receiver, decoding is performed backwards, hence the error events have a different form compared to those of the relay (cf. (A13) and the events below). Define the events,

E2,1(b)={(Ub+1n(M^b+1|1),X1,b+1n(1),Yb+1n)Aδ(PU,X1PY|U,X1q)forallqQ^n}E2,2(b)={(Ub+1n(M^b+1|mb),X1,b+1n(mb),Yb+1n)Aδ(PU,X1PY1|U,X1q),forsomemb1,qQ^n} (A23)

For b[1:B1], the error event E2(b) is bounded by

E2(b)E2(b+1)E2,1(b)E2,2(b)=E2(b+1)E2(b+1)cE2,1(b)E2(b+1)cE2,2(b), (A24)

with E2(B)=. Thus,

PrE2(b)E1(b)cPrE2(b+1)E1(b)c+PrE2,1(b)E1(b)c,E2(b+1)c+PrE2,2(b)E1(b)c,E2(b+1)c. (A25)

By similar arguments to those used above, we have that

PrE2,1(b)E1(b)c,E2(b+1)cPr(Ub+1n(1|1),X1,b+1n(1),Yb+1n)Aδ/2(PU,X1PY|U,X1q*)E1(b)c, (A26)

which tends to zero exponentially as n, due to (A22), and by the law of large numbers and Chernoff’s bound. Then, by similar arguments to those used for the bound on PrE1,2(b)E1(b1)c, the third term on the RHS of (A25) tends to zero as n, provided that R<infqQIq(U,X1;Y)ε4(δ), where ε4(δ)0 as δ0. Using induction, we have by (A25) that the second term on the RHS of (A12) tends to zero exponentially as n, for b[1:B1].

Moving to the error event for Mb, define

E3,1(b)={(Ubn(M^b|M^b1),Xbn(M^b,1|M^b1),X1,b(M^b1),Ybn)Aδ(PU,X,X1PY|X,X1q),forallqQ^n}E3,2(b)={(Ubn(M^b|M^b1),Xbn(M^b,mb|M^b1),X1,b(M^b1),Ybn)Aδ(PU,X,X1PY|X,X1q),forsomemb1,qQ^n}. (A27)

Given E2(b)cE2(b1)c, we have that M^b=1 and M^b1=1. Then, by similar arguments to those used above,

PrE3(b)E1(b)cE2(b)cE2(b1)cPrE3,1(b)E1(b)cE2(b)cE2(b1)c+PrE3,2(b)E1(b)cE2(b)cE2(b1)cea0n+(n+1)|S|·supqQmb1Pr(Ubn(1|1),Xbn(1,mb|1),X1,b(1),Ybn)Aδ(PU,X,X1PY|X,X1q)E1(b)cea0n+(n+1)|S|·supqQ2n[Iq(X;Y|U,X1)Rε5(δ)] (A28)

where a0>0 and ε5(δ)0 as δ0. The second inequality holds by (A22) along with the law of large numbers and Chernoff’s bound, and the last inequality holds as Xbn(1,mb|1) is conditionally independent of Ybn given (Ubn(1|1),X1,bn(1)) for every mb1. Thus, the third term on the RHS of (A12) tends to zero exponentially as n, provided that R<infqQIq(X;Y|U,X1)ε5(δ). Eliminating R and R, we conclude that the probability of error, averaged over the class of the codebooks, exponentially decays to zero as n, provided that R<RPDF(LQ). Therefore, there must exist a (2nR,n,ε) deterministic code, for a sufficiently large n. □

Appendix A.2. Cutset Upper Bound

This is a straightforward consequence of the cutset bound in [4]. Assume to the contrary that there exists an achievable rate R>RCS(LQ). Then, for some q*(s) in the closure of Q,

R>maxp(x,x1)minIq*(X,X1;Y),Iq*(X;Y,Y1|X1). (A29)

By the achievability assumption, we have that for every ε>0 and sufficiently large n, there exists a (2nR,n) random code CΓ such that Pe(n)(q,C)ε for every i.i.d. state distribution qQ, and in particular for q*. This holds even if q* is in the closure of Q but not in Q itself, since Pe(n)(q,C) is continuous in q. Consider using this code over a standard relay channel WY,Y1|X,X1 without a state, where WY,Y1|X,X1(y,y1|x,x1)=sSq*(s)WY,Y1|X,X1,S(y,y1|x,x1,s). It follows that the rate R as in (A29) can be achieved over the relay channel WY,Y1|X,X1, in contradiction to [4]. We deduce that the assumption is false, and R>RCS(LQ) cannot be achieved. □

Appendix B. Proof of Corollary 1

This is a straightforward consequence of Lemma 1, which states that the capacity of the compound relay channel is bounded by RPDF(LQ)C(LQ)RCS(LQ). Thus, if WY,Y1|X,X1,S is reversely degraded such that WY,Y1|X,X1,S=WY|X,X1WY1|Y,X1,S, then Iq(X;Y,Y1|X1)=Iq(X;Y|X1), and the bounds coincide by the minimax theorem [90], cf. (8) and (12). Similarly, if WY,Y1|X,X1,S is strongly degraded, i.e., WY,Y1|X,X1,S=WY1|X,X1WY|Y1,X1,S, then Iq(X;Y,Y1|X1)=I(X;Y1|X1), and by (8) and (13),

RCS(LQ)=minq(s)Qmaxp(x,x1)minIq(X,X1;Y),I(X;Y1|X1), (A30)
RPDF(LQ)=maxp(x,x1)minq(s)QminIq(X,X1;Y),I(X;Y1|X1). (A31)

Observe that minIq(X,X1;Y),I(X;Y1|X1) is concave in p(x,x1) and quasi-convex in q(s) (see e.g., [94] (Section 3.4])), hence the bounds (A30) and (A31) coincide by the minimax theorem [90]. □

Appendix C. Proof of Corollary 2

Consider the block-compound relay channel LQ×B, where the state distribution qbQ varies from block to block. Since the encoder, relay and receiver are aware of this jamming scheme, they can use a block coding scheme that is synchronized with the jammer block strategy. Thus, the capacity is the same as that of the ordinary compound channel, i.e., C(LQ×B)=C(LQ) and C(LQ×B)=C(LQ). Hence, (17) and (18) follow from Lemma 1. As for the second part of Corollary 2, observe that the block Markov coding scheme used in the proof of the partial decode-forward lower bound can be applied as is to the block-compound relay channel, since the relay and the destination receiver do not estimate the state distribution while decoding the messages (see Appendix A). Furthermore, the analysis also holds, where the actual state distribution q*, in (A18)–(A20) and (A26), is now replaced by the state distribution qb* which corresponds to block b[1:B]. □

Appendix D. Proof of Theorem 1

First, we explain the general idea. We modify Ahlswede’s Robustification Technique (RT) [59] to the relay channel. Namely, we use codes for the compound relay channel to construct a random code for the AVRC using randomized permutations. However, in our case, the strictly causal nature of the relay imposes a difficulty, and the application of the RT is not straightforward.

In [59], there is noncausal state information and a random code is defined via permutations of the codeword symbols and the received sequence. Here, however, the relay cannot apply permutations to its transmission x1n, because it depends on the received sequence y1n in a strictly causal manner. We resolve this difficulty using block Markov codes for the block-compound relay channel to construct a random code for the AVRC, applying B in-block permutations to the relay transmission, which depends only on the sequence received in the previous block. The details are given below.

Appendix D.1. Partial Decode-Forward Lower Bound

We show that every rate R<RPDF(L) (see (19)) can be achieved by random codes over the AVRC L, i.e., C(L)RPDF(L). We start with Ahlswede’s RT [59], stated below. Let h:Sn[0,1] be a given function. If, for some fixed αn(0,1), and for all q(sn)=i=1nq(si), with qP(S),

snSnq(sn)h(sn)αn, (A32)

then,

1n!πΠnh(πsn)βn,forallsnSn, (A33)

where Πn is the set of all n-tuple permutations π:SnSn, and βn=(n+1)|S|·αn.

According to Corollary 2, for every R<RPDF(L), there exists a (2nR(B1), nB, e2θn) block Markov code for the block-compound relay channel LP(S)×B for some θ>0 and sufficiently large n, where B>0 is arbitrarily large. Recall that the code constructed in the proof in Appendix A has the following form. The encoders use B>0 blocks to convey B1 messages mb, b[1:B1]. Each message consists of two parts, i.e., mb=(mb,mb), where mb[1:2nR] and mb[1:2nR]. In block b[1:B], the encoder sends xbn=fb(mb,mb|mb1), with fixed m0 and mB, and the relay transmits x1,bn=f1,b(y1,b1n), using the sequence received in the previous block. After receiving the entire output sequence (ybn)b=1B, the decoder finds an estimate for the messages. Set m^B=1. The first part of each message is decoded backwards as m^b=gb(yb+1n,m^b+1), for b=B1,B2,,1. Then, the second part of each message is decoded as m^b=gb(ybn,m^1,,m^B1), for b[1:B1]. The overall blocklength is then n·B and the average rate is B1B(R+R).

Given such a block Markov code CBM for the block-compound relay channel LP(S)×B, we have that

PrCBMEb|(Eb+1)ce2θn,PrCBMEb|E1c,,Eb1ce2θn (A34)

for b=B1,,1, where E0=EB=, and Eb={M^bMb}, Eb={M^bMb}, b[1:B1]. That is, for every sequence of state distributions q1,,qb+1, where qt(stn)=i=1nqt(st,i) for t[1:b+1],

s1nSnq1(s1n)s2nSnq2(s2n)sb+1nSnqb+1(sb+1n)·hb(s1n,s2n,,sb+1n)e2θn, (A35)

and

s1nSnq1(s1n)s2nSnq2(s2n)sbnSnqb(sbn)·hb(s1n,s2n,,sbn)e2θn, (A36)

where

hb(s1n,s2n,,sb+1n)=12n(b+1)(R+R)(m1,m1),,(mb+1,mb+1)y1,bnY1nPrY1,bn=y1,bn(M1,M1)=(m1,m1),,(Mb,Mb)=(mb,mb),S1n=s1n,,Sbn=sbn×yb+1n:gb(yb+1n,mb+1)mbWYn|Xn,X1n,Sn(yb+1n|fb+1(mb+1,mb+1|mb),f1,b+1(y1,bn),sb+1n) (A37)

and

hb(s1n,s2n,,sbn)=12nRmb=12nR12nR(B1)m1,,mB1y1,b1nY1nPr(Y1,b1n=y1,b1n|(M1,M1)=(m1,m1),,(Mb1,Mb1)=(mb1,mb1),S1n=s1n,,Sb1n=sb1n)×ybn,y1,bn:gb(ybn,m1,,mB1)mbWYn|Xn,X1n,Sn(ybn|fb(mb,mb|mb1),f1,b(y1,b1n),sbn). (A38)

The conditioning in the equations above can be explained as follows. In (A37), due to the code construction, the sequence Y1,bn received at the relay in block b[1:B] depends only on the messages (Mt,Mt) with tb. The decoded message M^b, at the destination receiver, depends on messages Mt with t>b, since the receiver decodes this part of the message backwards. In (A38), since the second part of the message Mb is decoded after backward decoding is complete, the estimation of Mb at the decoder depends on the entire sequence M^1,,M^B1. By (A35)–(A36), for every t[1:b], hb and hb as functions of st+1n and stn, respectively, satisfy (A32) with αn=e2θn, given that the state sequences in the other blocks are fixed. Hence, applying Ahlswede’s RT recursively, we obtain

1(n!)b+1π1,π2,,πb+1Πnhb(π1s1n,π2s2n,,πb+1sb+1n)(n+1)B|S|e2θneθn,1(n!)bπ1,π2,,πbΠnhb(π1s1n,π2s2n,,πbsbn)(n+1)B|S|e2θneθn, (A39)

for all (s1n,s2n,,sb+1n)S(b+1)n and sufficiently large n, such that (n+1)B|S|eθn.

On the other hand, for every π1,π2,,πb+1Πn, we have that

hb(π1s1n,π2s2n,,πb+1sb+1n)=Ehb(π1s1n,π2s2n,,πb+1sb+1n|Mt,Mt,t=1,,b+1), (A40)

with

hb(π1s1n,π2s2n,,πb+1sb+1n|mt,mt,t=1,,b+1)=y1,1,,y1,bt=0b1WY1n|Xn,X1n,Sn(y1,t+1n|ft+1(mt+1,mt+1|mt),f1,t+1(y1,tn),πt+1st+1n)×yb+1n:gb(yb+1n,mb+1)mbWYn|Xn,X1n,Sn(yb+1n|fb+1(mb+1,mb+1|mb),f1,b+1(y1,bn),πb+1sb+1n)=(a)y1,1,,y1,bt=0b1WY1n|Xn,X1n,Sn(πt+1y1,t+1n|ft+1(mt+1,mt+1|mt),f1,b+1(πty1,tn),πt+1st+1n)
×yb+1n:gb(πb+1yb+1n,mb+1)mbWYn|Xn,X1n,Sn(πb+1yb+1n|fb+1(mb+1,mb+1|mb),f1,b+1(πby1,bn),πb+1sb+1n)=(b)y1,1,,y1,bt=0b1WY1n|Xn,X1n,Sn(y1,t+1n|πt+11ft+1(mt+1,mt+1|mt),πt+11f1,b+1(πty1,tn),st+1n)×yb+1n:gb(πb+1yb+1n,mb+1)mbWYn|Xn,X1n,Sn(yb+1n|πb+11fb+1(mb+1,mb+1|mb),πb+11f1,b+1(πby1,bn),sb+1n), (A41)

where (a) is obtained by changing the order of summation over y1,1n,,y1,bn and yb+1n; and (b) holds because the relay channel is memoryless. Similarly,

hb(π1s1n,π2s2n,,πbsbn)=Ehb(π1s1n,π2s2n,,πbsbn|M1,,MB1,Mt,t=1,,b), (A42)

with

hb(π1s1n,π2s2n,,πbsbn|m1,,mB1,mt,t=1,,b)=y1,1,,y1,b1t=1b1WY1n|Xn,X1n,Sn(y1,tn|ft(mt,mt|mt1),f1,t(y1,tn),πtstn)×ybn:gb(ybn,m1,,mB1)mbWYn|Xn,X1n,Sn(ybn|fb(mb,mb|mb1),f1,b(y1,b1n),πbsbn)=(a)y1,1,,y1,b1t=1b1WY1n|Xn,X1n,Sn(πty1,tn|ft(mt,mt|mt1),f1,t(πt1y1,t1n),πtstn)×ybn:gb(πbybn,m1,,mB1)mbWYn|Xn,X1n,Sn(πbybn|fb(mb,mb|mb1),f1,b(πb1y1,b1n),πbsbn)=(b)y1,1,,y1,b1t=1b1WY1n|Xn,X1n,Sn(y1,tn|πt1ft(mt,mt|mt1),πt1f1,t(πt1y1,t1n),stn)×ybn:gb(πbybn,m1,,mB1)mbWYn|Xn,X1n,Sn(ybn|πb1fb(mb,mb|mb1),πb1f1,b(πb1y1,b1n),sbn). (A43)

Then, consider the (2nR(B1),nB) random Markov block code CBMΠ, specified by

fb,π(mb,mb|mb1)=πb1fb(mb,mb|mb1),f1,b,π(y1,b1n)=πb1f1,b(πb1y1,b1n), (A44a)

and

gb,π(yb+1n,m^b+1)=gb(πb+1yb+1n,m^b+1),gb,π(ybn,m^1,,m^B1)=gb(πybn,m^1,,m^B1), (A44b)

for π1,,πBΠn, with a uniform distribution μ(π1,,πB)=1|Πn|B=1(n!)B. That is, a set of B independent permutations is chosen at random and applied to all blocks simultaneously, while the order of the blocks remains intact. As we restricted ourselves to a block Markov code, the relaying function in a given block depends only on symbols received in the previous block, hence, the relay can implement those in-block permutations, and the coding scheme does not violate the causality requirement.

From (A41) and (A43), we see that using the random code CBMΠ, the error probabilities for the messages Mb and Mb are given by

PrCBMΠEb|(Eb+1)c,S1n=s1n,,Sb+1n=sb+1n=π1,,πBΠnμ(π1,,πB)hb(π1s1n,π2s2n,,πb+1sb+1n),PrCBMΠEb|E1c,,EB1c,S1n=s1n,,Sbn=sbn=π1,,πBΠnμ(π1,,πB)hb(π1s1n,π2s2n,,πbsbn), (A45)

for all s1n,,sb+1nSn, b[1:B1], and therefore, together with (A39), we have that the probability of error of the random code CBMΠ is bounded by Pe(n)(q,CBMΠ)eθn, for every q(snB)P(SnB). That is, CBMΠ is a (2nR(B1),nB,eθn) random code for the AVRC L, where the overall blocklength is nB, and the average rate B1B·R tends to R as B. This completes the proof of the partial decode-forward lower bound.

Appendix D.2. Cutset Upper Bound

The proof immediately follows from Lemma 1, since the random code capacity of the AVRC is bounded by the random code capacity of the compound relay channel, i.e., C(L)C(LP(S)). □

Appendix E. Proof of Lemma 2

We use the approach of [55], with the required adjustments. We use the random code constructed in the proof of Theorem 1. Let R<C(L), and consider the case where the marginal sender-relay and sender-receiver AVCs have positive capacity, i.e.,

C(W1(x1,1))>0,andC(W(x1,2))>0, (A46)

for some x1,1,x1,2X1 (see (23)). By Theorem 1, for every ε>0 and sufficiently large n, there exists a (2nR,n,ε) random code CΓ=μ(γ)=1k,Γ=[1:k],{Cγ}γΓ, where Cγ=(fγn,f1,γ,gγ), for γΓ. Following Ahlswede’s Elimination Technique [55], it can be assumed that the size of the code collection is bounded by k=|Γ|n2. By (A46), we have that for every ε>0 and sufficiently large ν, the code index γ[1:k] can be sent through the relay channel WY1|X,X1,S using a (2νR˜,ν,ε) deterministic code Ci=(f˜ν,g˜), where R˜>0, while the relay repeatedly transmits the symbol x1,1. Since k is at most polynomial, the encoder can reliably convey γ to the relay with a negligible blocklength, i.e., ν=o(n). Similarly, there exists (2νR˜,ν,ε) code Ci=(f˜ν,g˜) for the transmission of γ[1:k] through the channel WY|X,X1,S to the receiver, where ν=o(n) and R˜>0, while the relay repeatedly transmits the symbol x1,2.

Now, consider a code formed by the concatenation of Ci and Ci as consecutive prefixes to a corresponding code in the code collection {Cγ}γΓ. That is, the encoder first sends the index γ to the relay and the receiver, and then it sends the message m[1:2nR] to the receiver. Specifically, the encoder first transmits the (ν+ν)-sequence (f˜ν(γ),f˜ν(γ)) to convey the index γ, while the relay transmits the (ν+ν)-sequence (x˜1ν,x˜1ν), where x˜1ν=(x1,1,x1,1,,x1,1) and x˜1ν=(x1,2,x1,2,,x1,2). At the end of this transmission, the relay uses the first ν symbols it received to estimate the code index as γ^=g˜(y˜1ν).

Then, the message m is transmitted by the codeword xn=fγ(m), while the relay transmits x1n=f1,γ^n(y1n). Subsequently, decoding is performed in two stages as well; the decoder estimates the index at first, with γ^= g˜(y˜ν), and the message is then estimated by m^= gγ^(yn). By the union of events bound, the probability of error is then bounded by εc=ε+ε+ε, for every joint distribution in P(Sν+ν+n). That is, the concatenated code is a (2(ν+ν+n)R˜n,ν+ν+n,εc) code over the AVRC L, where the blocklength is n+o(n), and the rate R˜n=nν+ν+n·R approaches R as n. □

Appendix F. Proof of Corollary 4

Consider part 1. By Definition 3, if WY1|X,X1,S and WY|X,X1,S are not symmetrizable-X|X1 then there exist x1,1,x1,2X1 such that the DMCs WY1|X,X1,S(·|·, x1,1, ·) and WY|X,X1,S(·|·,x1,2,·) are non-symmetrizable in the sense of [57] (Definition 2). This, in turn, implies that C(W1(x1,1))>0 and C(W(x1,2))>0, due to [57] (Theorem 1). Hence, by Lemma 2, C(L)=C(L), and by Theorem 1, RPDF(L)C(L)RCS(L).

Part 3 immediately follows from part 1 and Corollary 3. As for part 2, consider a strongly reversely degraded relay channel. We claim that if WY|X,X1,S is symmetrizable-X|X1, then WY1|X,X1,S is also symmetrizable-X|X1. Indeed, suppose that WY|X,X1,S is symmetrized by some J(s|x,x1) (see Definition 24). Then, for every x,x˜X, x1X1, and y1Y1,

sSJ(s|x˜,x1)WY1|X,X1,S(y1|x,x1,s)=sSJ(s|x˜,x1)yYWY,Y1|X,X1,S(y,y1|x,x1,s)=(a)yYWY1|Y,X1(y1|y,x1)sSJ(s|x˜,x1)WY|X,X1,X(y|x,x1,s)=(b)yYWY1|Y,X1(y1|y,x1)sSJ(s|x,x1)WY|X,X1,X(y|x˜,x1,s)=(c)sSJ(s|x,x1)yYWY,Y1|X,X1,S(y,y1|x˜,x1,s)=sSJ(s|x,x1)WY1|X,X1,S(y1|x˜,x1,s), (A47)

where (a) and (c) hold since WY,Y1|X,X1,S is strongly reversely degraded, and (b) holds since WY|X,X1,S is symmetrized by J(s|x,x1). This means that WY1|X,X1,S is also symmetrizable-X|X1. It can be deduced that given the conditions of part 2, both WY|X,X1,S and WY1|X,X1,S are non-symmetrizable-X|X1. Hence, the proof follows from part 1 and Corollary 3. □

Appendix G. Proof of Lemma 3

The proof is based on generalizing the technique by [56]. Let L be a symmetrizable-X|X1. Assume to the contrary that a positive rate R>0 can be achieved. That is, for every ε>0 and sufficiently large n, there exists a (2nR,n,ε) code C=(f,f1,g). Hence, the size of the message set is at least 2, i.e.,

M2nR2. (A48)

We now show that there exists a distribution q(sn) such that the probability of error Pe(n)(q,C) is bounded from below by a positive constant, in contradiction to the assumption above.

By Definition 3, there exists a conditional distribution J(s|x) that satisfies (24). Then, consider the state sequence distribution q(sn)=1Mm=1MJn(sn|xn(m)), where Jn(sn|xn)=i=1nJ(si|xi) and xn(m)=f(m). For this distribution, the probability of error is given by

Pe(n)(q,C)=snSn1Mm˜=1MJn(sn|xn(m˜))·1Mm=1M(yn,y1n):g(yn)mWn(yn,y1n|xn(m),f1n(y1n),sn)=12M2m=1Mm˜=1M(yn,y1n):g(yn)msnSnWn(yn,y1n|xn(m),f1n(y1n),sn)Jn(sn|xn(m˜))+12M2m=1Mm˜=1M(yn,y1n):g(yn)m˜snSnWn(yn,y1n|xn(m˜),f1n(y1n),sn)Jn(sn|xn(m)) (A49)

with WnWYn,Y1n|Xn,X1n,Sn for short notation, where in the last sum we interchanged the summation indices m and m˜. Then, consider the last sum, and observe that by (24), we have that

snSnWn(yn,y1n|xn(m˜),f1n(y1n),sn)Jn(sn|xn(m))=i=1nsiSW(yi,y1,i|xi(m˜),f1,i(y1i1),si)J(si|xi(m))=i=1nsiSW(yi,y1,i|xi(m),f1,i(y1i1),si)J(si|xi(m˜))=snSnWn(yn,y1n|xn(m),f1n(y1n),sn)Jn(sn|xn(m˜)). (A50)

Substituting (A50) in (A49), we have

Pe(n)(q,C)=12M2m=1Mm˜=1MsnSn[(yn,y1n):g(yn)mWn(yn,y1n|xn(m),f1n(y1n),sn)Jn(sn|xn(m˜))+(yn,y1n):g(yn)m˜Wn(yn,y1n|xn(m),f1n(y1n),sn)Jn(sn|xn(m˜))]
12M2m=1Mm˜msnSnyn,y1nWn(yn,y1n|xn(m),f1n(y1n),sn)Jn(sn|xn(m˜))=M(M1)2M214, (A51)

where the last inequality follows from (A48), hence a positive rate cannot be achieved. □

Appendix H. Proof of Lemma 4

Let L={WY1|X,X1WY|Y1,X1,S} be a symmetrizable-X1×Y1 degraded AVRC. The proof follows similar lines as in Appendix G. First, assume to the contrary that there exists a (2nR,n,ε) code C=(f,f1,g), with M2nR2. By Definition 4, there exists J(s|x1,y1) that satisfies (28). Hence, defining

q(sn)=1Mm=1My1nY1WY1n|Xn,X1n(y1n|f(m),f1n(y1n))Jn(sn|f1n(y1n),y1n), (A52)

where Jn(sn|x1n,y1n)=i=1nJ(si|x1,i,y1,i), we have that

snSnWn(yn|y˜1n,f1n(y˜1n),sn)Jn(sn|f1n(y1n),y1n)=snSnWn(yn|y1n,f1n(y1n),sn)Jn(sn|f1n(y˜1n),y˜1n). (A53)

By similar manipulations as in Appendix G, we obtain

Pe(n)(q,C)=12M2m=1Mm˜=1My1n,y˜1nWY1n|Xn,X1n(y˜1n|f(m˜),f1n(y˜1n))WY1n|Xn,X1n(y1n|f(m),f1n(y1n))×snSn[yn:g(yn)mWn(yn|y1n,f1n(y1n),sn)Jn(sn|f1n(y˜1n),y˜1n)+yn:g(yn)m˜Wn(yn|y1n,f1n(y1n),sn)Jn(sn|f1n(y˜1n),y˜1n)]M(M1)2M214, (A54)

hence a positive rate cannot be achieved. □

Appendix I. Analysis of Example 1

We show that the random code capacity of the AVRC in Example 1 is given by C(L)=min12,1h(θ). As the AVRC is degraded, the random code capacity is given by

C(L)=RPDF(L)=RCS(L)=maxp(x,x1)minmin0q1Iq(X,X1;Y),I(X;Y1|X1), (A55)

due to part 2 of Corollary 3, where qq(1)=1q(0). Now, consider the direct part. Set p(x,x1)=p(x)p(x1), where XBernoulli(MM1/2) and X1Bernoulli(MM1/2). Then,

I(X;Y1|X1)=1h(θ),Hq(Y)=12qlog12q(1q)log12(1q)12log12=1+12h(q),Hq(Y|X,X1)=h(q). (A56)

Hence,

C(L)minmin0q1112h(q),1h(θ)=min12,1h(θ). (A57)

As for the converse part, we have the following bounds,

11C(L)maxp(x,x1)I(X;Y1|X1)=1h(θ), (A58)

and

11C(L)maxp(x,x1)min0q1Iq(X,X1;Y)maxp(x,x1)[Hq(Y)Hq(Y|X,X1)]|q=12=max0p11+12h(p)1=12, (A59)

where pPrX1=1. □

Appendix J. Proof of Lemma 5

The proof follows the lines of [5]. Consider an AVRC L = {WY|X,X1 WY1|X,X1,S} with orthogonal sender components. We apply Theorem 1, which states that RPDF(L)C(L)RCS(L).

Appendix J.1. Achievability Proof

To show achievability, we set U=X and p(x,x,x1)=p(x1)p(x|x1)p(x|x1) in the partial decode-forward lower bound RPDF(L)RPDF(LQ)|Q=P(S). Hence, by (9),

RPDF(L2)maxp(x1)p(x|x1)p(x|x1)minI(X,X,X1;Y),minq(s)Iq(X;Y1|X1)+I(X;Y|X1,X). (A60)

Now, by (29), we have that (X,Y1)(X,X1)Y form a Markov chain. As (X1,X,X)p(x1)p(x|x1)p(x|x1), it further follows that (X,Y1)X1Y form a Markov chain, hence I(X,X,X1;Y)=I(X,X1;Y) and I(X;Y|X1,X)=I(X;Y|X1). Thus, (A60) reduces to the expression in the RHS of (30). If WY1|X,X1,S is non-symmetrizable-X|X1, then (A60) is achievable by deterministic codes as well, due to Corollary 4.

Appendix J.2. Converse Proof

By (8) and (19), the cutset upper bound takes the following form,

RCS(L)=minq(s)maxp(x,x,x1)minI(X,X,X1;Y),Iq(X,X;Y,Y1|X1)=maxp(x,x,x1)minI(X,X,X1;Y),minq(s)Iq(X,X;Y,Y1|X1), (A61)

where the last line is due to the minimax theorem [90]. For the AVRC with orthogonal sender components, as specified by (29), we have the following Markov relations,>

Y1(X,X1)(X,Y), (A62)
(X,Y1)(X,X1)Y. (A63)

Hence, by (A63), I(X,X,X1;Y)=I(X,X1;Y). As for the second mutual information in the RHS of (A61), by the mutual information chain rule,

Iq(X,X;Y,Y1|X1)=Iq(X;Y1|X1)+Iq(X;Y1|X,X1)+Iq(X,X;Y|X1,Y1)=(a)Iq(X;Y1|X1)+Iq(X,X;Y|X1,Y1)=(b)Iq(X;Y1|X1)+Hq(Y|X1,Y1)H(Y|X,X1)(c)Iq(X;Y1|X1)+I(X;Y|X1) (A64)

where (a) is due to (A62), (b) is due to (A63), and (c) holds since conditioning reduces entropy. Therefore,

RCS(L)maxp(x,x,x1)minI(X,X1;Y),minq(s)Iq(X;Y1|X1)+I(X;Y|X1). (A65)

Without loss of generality, the maximization in (A65) can be restricted to distributions of the form p(x,x,x1)=p(x1)· p(x|x1)· p(x|x1). □

Appendix K. Proof of Lemma 6

Consider the Gaussian compound relay channel with SFD under input constraints Ω and Ω1 and state constraint Λ, i.e., Q={q(s):ES2Λ}.

Appendix K.1. Achievability Proof

Consider the direct part. Although we previously assumed that the input, state and output alphabets are finite, our results for the compound relay channel can be extended to the continuous case as well, using standard discretization techniques [3] (Section 3.4.1); [55,95]. In particular, Lemma 1 can be extended to the compound relay channel LQ under input constraints Ω and Ω1 and state constraint Λ, by choosing a distribution p(x,x,x1) such that E(X2+X2)Ω and EX12Ω1. Then, the capacity of LQ is bounded by

C(LQ)RPDF(LQ)maxp(x)p(x,x1):E(X2+X2)Ω,EX12Ω1min{minq(s):ES2ΛIq(X1;Y)+minq(s):ES2ΛIq(X;Y|X1),I(X;Y1)+minq(s):ES2ΛIq(X;Y|X1)}, (A66)

which follows from the partial decode-forward lower bound by taking U=X. Lemma 1 further states that there exists a block Markov code that achieves this rate such that the probability of error decays exponentially as the blocklength increases.

Let 0α,ρ1, and let (X,X,X1) be jointly Gaussian with

XN(0,αΩ),XN(0,(1α)Ω),X1N(0,Ω1), (A67)

where the correlation coefficient of X and X1 is ρ, while X is independent of (X,X1). Hence,

I(X;Y1)=12log1+(1α)Ωσ2. (A68)

Since Gaussian noise is the worst additive noise under variance constraint [96] (Lemma II.2), and as Var(X|X1=x1)=(1ρ2)αΩ for all x1R, we have that

minq(s):ES2ΛIq(X;Y|X1)=12log1+(1ρ2)αΩΛ. (A69)

It is left for us to evaluate the first term in the RHS of (A66). Then, by standard whitening transformation, there exist two independent Gaussian random variables T1 and T2 such that

X+X1=T1+T2, (A70)
T1N(0,(1ρ2)αΩ),T2N(0,Ω1+ρ2αΩ+2ραΩ·Ω1). (A71)

Hence, Y=T1+T2+S, and as Var(X|X1=x1)=Var(T1) for all x1R, we have that

Iq(X1;Y)=Hq(Y)Hq(X+S|X1)=Hq(Y)Hq(T1+S)=Iq(T2;Y) (A72)

Let S¯T1+S. Then, since Gaussian noise is the worst additive noise under variance constraint [96] (Lemma II.2),

minq(s):ES2ΛIq(X1;Y)=minq(s):ES2ΛIq(T2;T2+S¯)=12log1+Var(T2)Var(T1)+Λ=12logΩ1+ρ2αΩ+2ραΩ·Ω1+Λ(1ρ2)αΩ+Λ. (A73)

Substituting (A68), (A69) and (A73) in the RHS of (A66), we have that

C(LQ)max0α,ρ1min{12logΩ1+ρ2αΩ+2ραΩ·Ω1+Λ(1ρ2)αΩ+Λ+12log1+(1ρ2)αΩΛ,12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛ}. (A74)

Observe that the first sum in the RHS of (A74) can be expressed as

12logΩ1+ρ2αΩ+2ραΩ·Ω1+Λ(1ρ2)αΩ+Λ+12log(1ρ2)αΩ+ΛΛ=12logΩ1+ρ2αΩ+2ραΩ·Ω1+ΛΛ=12log1+Ω1+ρ2αΩ+2ραΩ·Ω1Λ. (A75)

Hence, the direct part follows from (A74). □

Appendix K.2. Converse Proof

By Lemma 1, C(LQ)RCS(LQ). Now, observe that

RCS(LQ)=minq(s):ES2Λmaxp(x)p(x,x1):E(X2+X2)Ω,EX12ΩminIq(X,X1;Y),I(X;Y1)+Iq(X;Y|X1)maxp(x)p(x,x1):E(X2+X2)Ω,EX12ΩminIq(X,X1;Y),I(X;Y1)+Iq(X;Y|X1)|SN(0,Λ)=max0α,ρ1min{12log1+Ω1+ρ2αΩ+2ραΩ·Ω1Λ,12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛ}, (A76)

where the last equality is due to [5]. □

Appendix L. Proof of Theorem 2

Appendix L.1. Achievability Proof

To show that C(L)C(LQ), we follow the steps in the proof of Theorem 1, where we replace Ahlswede’s original RT with the modified version in [85, 86] (Lemma 9), plugging ln(sn)=1ni=1nsi2. Then, by Lemma 6, it follows that

C(L)max0α,ρ1min{12log1+(1+α+2ρα)ΩΛ,12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛ}. (A77)

The details are omitted. □

Appendix L.2. Converse Proof

Assume to the contrary that there exists an achievable rate R such that

R>max0α,ρ1min{12log1+(1+α+2ρα)ΩΛδ,12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛδ} (A78)

using random codes over the Gaussian AVRC L, under input constraints Ω and Ω1 and state constraint Λ, where δ>0 is arbitrarily small. That is, for every ε>0 and sufficiently large n, there exists a (2nR,n) random code CΓ=(μ,Γ,{Cγ}γΓ) for the Gaussian AVRC L, under input constraints Ω and Ω1 and state constraint Λ, such that

Pe|s(CΓ)ε, (A79)

for all m[1:2nR] and sRn with s2nΛ.

Consider using the random code CΓ over the Gaussian compound relay channel LQ under state constraint (Λδ), i.e., with

Q={q(s):ES2Λδ}, (A80)

under input constraints Ω and Ω1. Let q¯(s)Q be a given state distribution. Then, define a sequence of i.i.d. random variables S¯1,,S¯nq¯(s). Letting q¯(sn)i=1nq¯(si), the probability of error is bounded by

Pe(n)(q¯,CΓ)sn:ln(sn)Λq¯n(sn)Pe|sn(n)(CΓ)+Pr1ni=1nS¯i2>Λ. (A81)

Then, the first sum is bounded by (A79), and the second term vanishes as well by the law of large numbers, since q¯(s) is in (A80). Hence, the rate R in (A78) is achievable for the Gaussian compound relay channel LQ, in contradiction to Lemma 6. We deduce that the assumption is false, and (A78) cannot be achieved. □

Appendix M. Proof of Theorem 3

Consider the Gaussian AVRC L with SFD under input constraints Ω and Ω1 and state constraint Λ. In the proof, we modify the techniques by Csiszár and Narayan [87]. In the direct part, we use their correlation binning technique within the decode-forward coding scheme, and in the converse part, we consider a jamming scheme which simulates the transmission sum by the encoder and the relay.

Appendix M.1. Lower Bound

We construct a block Markov code using backward minimum-distance decoding in two steps. The encoders use B blocks, each consists of n channel uses, to convey (B1) independent messages to the receiver, where each message Mb, for b[1:B1], is divided into two independent messages. That is, Mb=(Mb,Mb), where Mb and Mb are uniformly distributed, i.e.,

MbUnif[1:2nR],MbUnif[1:2nR],withR+R=R, (A82)

for b[1:B1]. For convenience of notation, set M0=MB1 and M0=MB1. The average rate B1B·R is arbitrarily close to R.

Codebook Construction: Fix 0α,ρ1 with

(1ρ2)αΩ>Λ, (A83)
Ω1Ω(Ω1+ραΩ)2>Λ+(1ρ2)αΩ. (A84)

We construct B codebooks Fb of the following form,

Fb=x1(mb1),x(mb,mb|mb1),x(mb):mb1,mb[1:2nR],mb[1:2nR], (A85)

for b[2:B1]. The codebooks F1 and FB have the same form, with fixed m0=mB1 and m0=mB1.

The sequences x(mb), mb[1:2nR] are chosen as follows. Observe that the channel from the sender to the relay, Y1=X+Z, does not depend on the state. Thus, by Shannon’s well-known result on the point-to-point Gaussian channel [97], the message mb can be conveyed to the relay reliably, under input constraint (1α)Ω, provided that R<12log1+(1α)Ωσ2δ1, where δ1 is arbitrarily small (see also [98] (Chapter 9)). That is, for every ε>0 and sufficiently large n, there exists a (2nR,n,ε) code C=(x(mb),g1(y1,b)), such that x(mb)2n(1α)Ω for all mb[1:2nR].

Next, we choose the sequences x1(mb1) and x(mb,mb| mb1), for mb1, mb[1:2nR], mb[1:2nR]. Applying Lemma 7 by [87] repeatedly yields the following.

Lemma A1.

For every ε>0, 8ε<η<1, K>2ε, 2εRK, 2εRK, and nn0(ε,η,K),

  • 1. 
    there exist 2nR unit vectors,
    a(mb1)Rn,mb1[1:2nR], (A86)
    such that for every unit vector cRn and 0θ,ζ1,
    m˜b1[1:2nR]:a(m˜b1),cθ2n[R+12log(1θ2)]++ε, (A87)
    and if θη and θ2+ζ2>1+η22R, then
    12nR|{mb1[1:2nR]:|a(m˜b1),a(mb1)|θ,|a(m˜b1),c|ζ,forsomem˜b1mb1}|2nε. (A88)
  • 2. 
    Furthermore, for every mb[1:2nR], there exist 2nR unit vectors,
    v(mb,mb)Rn,mb[1:2nR], (A89)
    such that for every unit vector cRn and 0θ,ζ1,
    m˜b[1:2nR]:v(mb,m˜b),cθ2n[R+12log(1θ2)]++ε, (A90)
    and if θη and θ2+ζ2>1+η22R, then
    12nR|{mb[1:2nR]:|v(mb,m˜b),v(mb,mb)|θ,|v(mb,m˜b),c|ζ,forsomem˜bmb}|2nε. (A91)

Then, define

x1(mb1)=nγ(Ωδ)·a(mb1),x(mb,mb|mb1)=ραγ1·x1(mb1)+β·v(mb,mb), (A92)

where

βn(1ρ2)α(Ωδ),γΩ1/Ω. (A93)

Note that x1(mb1)2=nγ(Ωδ)<nΩ1, for all mb1[1:2nR]. On the other hand, x(mb,mb|mb1)2 could be greater than nαΩ due to the possible correlation between x1(mb1) and v(mb,mb).

Encoding: Let (m1,m1,,mB1,mB1) be a sequence of messages to be sent. In block b[1:B], if x(mb,mb|mb1)2 nαΩ, transmit (x(mb,mb|mb1),x(mb)). Otherwise, transmit (0,x(mb)).

Relay Encoding: In block 1, the relay transmits x1(1). At the end of block b[1:B1], the relay receives y1,b, and finds an estimate m¯b=g1(y1,b). In block b+1, the relay transmits x1(m¯b).

Backward Decoding: Once all blocks (yb)b=1B are received, decoding is performed backwards. Set m^0=m^01. For b=B1,B2,,1, find a unique m^b[1:2nR] such that

yb+1(1+ραγ1)x1(m^b)yb+1(1+ραγ1)x1(mb),forallmb[1:2nR]. (A94)

If there is more than one such m^b[1:2nR], declare an error.

Then, the decoder uses m^1,,m^B1 as follows. For b=B1,B2,,1, find a unique m^b[1:2nR] such that

ybx1(m^b1)x(m^b,m^b|m^b1)ybx1(m^b1)x(m^b,mb|m^b1),forallmb[1:2nR]. (A95)

If there is more than one such m^b[1:2nR], declare an error.

Analysis of Probability of Error: Fix sSn, and let

c0ss. (A96)

The error event is bounded by the union of the following events. For b[1:B1], define

E1(b)={M¯bMb},E2(b)={M^bMb},E3(b)={M^bMb}. (A97)

Then, the conditional probability of error given the state sequence s is bounded by

Pe|s(C)b=1B1PrE1(b)+b=1B1PrE2(b)E1c(b)+b=1B1PrE3(b)E1c(b1)E2c(b)E2c(b1), (A98)

with E1(0)=E2(0)=, where the conditioning on S=s is omitted for convenience of notation. Recall that we have defined C as a (2nR,n,ε) code for the point-to-point Gaussian channel Y1=X+Z. Hence, the first sum in the RHS of (A98) is bounded by B·ε, which is arbitrarily small.

To be more concise, we only give the details for erroneous decoding of Mb at the receiver. Consider the following events,

E2(b)={Yb+1(1+ραγ1)x1(m˜b)Yb+1(1+ραγ1)x1(Mb),forsomem˜bMb},E2,1(b)={|a(Mb),c0|η},E2,2(b)={|a(Mb),v(Mb+1)|η}E2,3(b)={|v(Mb+1),c0|η}E˜2(b)=E2(b)E1c(b)E2,1c(b)E2,2c(b)E2,3c(b), (A99)

where Mb+1=(Mb+1,Mb+1). Then,

E2(b)E1c(b)E2,1(b)E2,2(b)E2,3(b)(E2(b)E1c(b))=E2,1(b)E2,2(b)E2,3(b)E˜2(b). (A100)

Hence, by the union of events bound, we have that

PrE2(b)E1c(b)PrE2,1(b)+PrE2,2(b)+PrE2,3(b)+PrE˜2(b). (A101)

By Lemma A1, given R>12log(1η2), the first term is bounded by

PrE2,1(b)=Pra(Mb),c0η+Pra(Mb),c0η2·12nR·2nR+12log(1η2)+ε2·2n12η2+ε, (A102)

since log(1+t)t for tR. As η28ε, the last expression tends to zero as n. Similarly, PrE2,2(b) and PrE2,3(b) tend to zero as well. Moving to the fourth term in the RHS of (A101), observe that for a sufficiently small ε and η, the event E2,2c(b) implies that x(Mb+1|Mb)2nαΩ, while the event E1c(b) means that M¯b=Mb. Hence, the encoder transmits (x(Mb+1|Mb),x(Mb+1)), the relay transmits x1(Mb), and we have that

Yb+1(1+ραγ1)x1(m˜b)2Yb+1(1+ραγ1)x1(Mb)2=(1+ραγ1)x1(Mb)+βv(Mb+1)+s(1+ραγ1)x1(m˜b)2βv(Mb+1)+s2=2(1+ραγ1)212x1(Mb)2+12x1(m˜b)2x1(m˜b),x1(Mb)+2(1+ραγ1)x1(Mb),βv(Mb+1)+sx1(m˜b),βv(Mb+1)+s (A103)

Then, since x1(mb)2=nγ(Ωδ) for all mb[1:2nR], we have that

E2(b)E1c(b)E2,2c(b){(1+ραγ1)x1(m˜b),x1(Mb)+x1(m˜b),βv(Mb+1)+sn(1+ραγ1)γ(Ωδ)+x1(Mb),βv(Mb+1)+s,forsomem˜bMb}. (A104)

Observe that for sufficiently small ε and η, the event E2,1c(b)E2,2c(b)E2,3c(b) implies that

x1(Mb),βv(Mb+1)+sδ, (A105)

and

βv(Mb+1)+s2n[(1ρ2)αΩ+Λ]. (A106)

Hence, by (A104) and (A105),

E˜2(b)=E2(b)E1c(b)E2,1c(b)E2,2c(b)E2,3c(b){(1+ραγ1)x1(m˜b),x1(Mb)+x1(m˜b),βv(Mb+1)+sn(1+ραγ1)γ(Ω2δ),forsomem˜bMb}. (A107)

Dividing both sides of the inequality by n(1+ραγ1), we obtain

E˜2(b)1nx1(m˜b),x1(Mb)+x1(m˜b),βv(Mb+1)+sn(1+ραγ1)γ(Ω2δ),forsomem˜bMb. (A108)

Next, we partition the set of values of 1nx1(m˜b),x1(Mb) to K bins. Let τ1<τ2<<τK be such partition, where

τ1=γ(Ω2δ)(Ωδ)[(1ρ2)αΩ+Λ]1+ραγ1,τK=γ(Ω3δ),τk+1τkγ·δ,fork=[1:K1], (A109)

where K is a finite constant which is independent of n, as in Lemma A1. By (A106) and (A108), given the event E˜2(b), we have that

1nx1(m˜b),x1(Mb)τ1>0, (A110)

where the last inequality is due to (A84), for sufficiently small δ>0. To see this, observe that the inequality in (A84) is strict, and it implies that

γ·(γΩ+ραΩ)>(1ρ2)αΩ+Λ. (A111)

Hence, for sufficiently small δ>0, τ1>0 as

τ1=Ω2δ1+ραγ1·γ(γ(Ω2δ)+ρα(Ω2δ))ΩδΩ2δ[(1ρ2)αΩ+Λ]. (A112)

Furthermore, if τk1nx1(m˜b),x1(Mb)<τk+1, then

x1(m˜b),βv(Mb+1)+sn(1+ραγ1)γ(Ω2δ)τk+1γ(Ω3δ)τk. (A113)

Thus,

PrE˜2(b)k=1K1Pr(1n|x1(m˜b),x1(Mb)|τk,|x1(m˜b),βv(Mb+1)+s|n(1+ραγ1)γ(Ω3δ)τk,forsomem˜bMb)+Pr1n|x1(m˜b),x1(Mb)|τK,forsomem˜bMb. (A114)

By (A106), this can be further bounded by

PrE˜2(b)k=1KPr|a(m˜b),a(Mb)|θk,|a(m˜b),c(Mb+1)|μk,forsomem˜bMb, (A115)

where

c(mb+1)βv(mb+1)+sβv(mb+1)+s, (A116)

and

θkτkγ(Ωδ),ζk(1+ραγ1)γ(Ω3δ)τkγ(Ωδ)((1ρ2)αΩ+Λ),fork[1:K1];θKτKΩδ,ζK=0. (A117)

By Lemma A1, the RHS of (A115) tends to zero as n provided that

θkηandθk2+ζk2>1+ηe2R,fork=[1:K]. (A118)

For sufficiently small ε and η, we have that ηθ1=τ1γ(Ωδ), hence the first condition is met. Then, observe that the second condition is equivalent to G(τk)>1+ηe2R, for k[1:K1], where

G(τ)=(Aτ)2+D2(Lτ)2, (A119)

with

A=1γ(Ωδ),D=1+ραγ1γ(Ωδ)((1ρ2)αΩ+Λ),L=γ(Ω3δ). (A120)

By differentiation, we have that the minimum value of this function is given by minτ1ττKG(τ)=A2D2L2A2+D2=D2A2+D2δ1, where δ10 as δ0. Thus, the RHS of (A115) tends to zero as n, provided that

R<12log1+ηD2A2+D2+δ1=12logη+δ1+(1ρ2)αΩ+Λ(γ+α+2ραγ)Ω+Λδγ(1+ραγ1)2. (A121)

This is satisfied for R=Rα(L)δ, with

Rα(L)=12log(γ+α+2ραγ)Ω+Λ(1ρ2)αΩ+Λ=12log(1ρ2)αΩ+Λ(γ+α+2ραγ)Ω+Λ. (A122)

and arbitrary δ>0, if η and δ are sufficiently small.

As for the error event for Mb, a similar derivation shows that the probability term in the last sum in (A98) exponentially tends to zero as n, provided that

R<12log1+ηβ2(12δ)2β2+nΛ<12logη+Λ(1ρ2)α(Ωδ)+Λ. (A123)

This is satisfied for R=Rα(L)δ, with

Rα(L)=12log(1ρ2)αΩ+ΛΛ=12logΛ(1ρ2)αΩ+Λ (A124)

for an arbitrary δ>0, if η and δ are sufficiently small.

We have thus shown achievability of every rate

R<minRα(L)+Rα(L),12log1+(1α)Ωσ2+Rα(L), (A125)

where

Rα(L)+Rα(L)=12log(γ+α+2ραγ)Ω+Λ(1ρ2)αΩ+Λ+12log(1ρ2)αΩ+ΛΛ=12log1+Ω1+αΩ+2ραΩ·Ω1Λ (A126)

(see (A93)). This completes the proof of the lower bound.

Appendix M.2. Upper Bound

Let R>0 be an achievable rate. Then, there exists a sequence of (2nR,n,εn*) codes Cn=(f,f1,g) for the Gaussian AVRC L with SFD such that εn*0 as n, where the encoder consists of a pair f=(f,f), with f:[1:2nR]Rn and f:[1:2nR]Rn. Assume without loss of generality that the codewords have zero mean, i.e.,

12nRm=12nR1ni=1nfi(m)=0,dy1·12nRm[1:2nR]PY1|M(y1|m)·1ni=1nf1,i(y1,1,y1,2,,y1,i1)=0, (A127)

where Py1|M(y1|m)=1(2πσ2)MMn/2ey1f(m)2/2σ2. If this is not the case, redefine the code such that the mean is subtracted from each codeword. Then, define

α1nΩ·12nRm[1:2nR]f(m)2,α11nΩ1·12nRm[1:2nR]dy1·PY1|M(y1|m)·f1(y1)2.ρ1nαΩ·α1Ω1dy1·12nRm[1:2nR]PY1|M(y1|m)·f(m),f1(y1), (A128)

Since the code satisfies the input constraints Ω and Ω1, we have that α, α1 and ρ are in the interval [0,1].

First, we show that if

Λ>Ω1+αΩ+2ραΩ·Ω1+δ, (A129)

then the capacity is zero, where δ>0 is arbitrarily small. Consider the following jamming strategy. The jammer draws a message M˜[1:2nR] uniformly at random, and then, generates a sequence Y˜1Rn distributed according to PY1|M(y˜1|m˜). Let S˜=f(M˜)+f1(Y˜1). If 1nS˜2Λ, the jammer chooses S˜ to be the state sequence. Otherwise, let the state sequence consist of all zeros. Observe that

ES˜2=Ef(M˜)+f1(Y˜1)2=Ef(M˜)2+Ef1(Y˜1)2+2Ef(M˜),f1(Y˜1)=n(αΩ+α1Ω1+2ραΩ·α1Ω1)n(αΩ+Ω1+2ραΩ·Ω1)<n(Λδ). (A130)

where the second equality is due to (A128), and the last inequality is due to (A129). Thus, by Chebyshev’s inequality, there exists κ>0 such that

Pr1nS˜2Λκ. (A131)

The state sequence S is then distributed according to

PS|{1nS˜2Λ}(s)=12nRm˜[1:2nR]y˜1:f(m˜)+f1(y˜1)=sdy1PY1|M(y1|m),PrS=01nS˜2>Λ=1. (A132)

Assume to the contrary that a positive rate can be achieved when the channel is governed by such state sequence, hence the size of the message set is at least 2, i.e., M2nR2. The probability of error is then bounded by

Pe(n)(q,C)=ds·q(s)Pe|s(n)(C)Pr1nS˜2Λ·s:1ns2Λds·PS|{1nS˜2Λ}(s)·Pe|s(n)(C)κ·s:1ns2Λds·PS|{1nS˜2Λ}(s)·Pe|s(n)(C) (A133)

where the inequality holds by (A131). Next, we have that

Pe|s(n)(C)=1Mm=1Mdy1·PY1|M(y1|m)·1y1:g(f(m)+f1(y1)+s)m, (A134)

where we define the indicator function G(y1)=1{y1A} such that G(y1)=1 if y1A, and G(y1)=0 otherwise. Substituting (A132) and (A134) into (A133) yields

Pe(n)(q,C)κ·s:1ns2Λds·1Mm˜=1My˜1:f(m˜)+f1(y˜1)=sdy˜1·PY1|M(y˜1|m˜)×1Mm=1Mdy1·PY1|M(y1|m)·1y1:g(f(m)+f1(y1)+s)m. (A135)

Eliminating s=f(m˜)+f1(y˜1), and adding the constraint f(m)+f1(y1)2Λ, we obtain the following,

Pe(n)(q,C)κM2m=1Mm˜=1M(y1,y˜1):1nf(m)+f1(y1)2Λ,1nf(m˜)+f1(y˜1)2Λdy1dy˜1·PY1|M(y1|m)PY1|M(y˜1|m˜)×1y1:g(f(m)+f1(y1)+f(m˜)+f1(y˜1))m. (A136)

Now, by interchanging the summation variables (m,y1) and (m˜,y˜1), we have that

Pe(n)(q,C)κ2M2m=1Mm˜=1M(y1,y˜1):1nf(m)+f1(y1)2Λ,1nf(m˜)+f1(y˜1)2Λdy1dy˜1·PY1|M(y1|m)PY1|M(y˜1|m˜)
×1y1:g(f(m)+f1(y1)+f(m˜)+f1(y˜1))m+κ2M2m=1Mm˜=1M(y1,y˜1):1nf(m)+f1(y1)2Λ,1nf(m˜)+f1(y˜1)2Λdy1dy˜1·PY1|M(y1|m)PY1|M(y˜1|m˜)×1y1:g(f(m)+f1(y1)+f(m˜)+f1(y˜1))m˜. (A137)

Thus,

Pe(n)(q,C)κ2M2m=1Mm˜m(y1,y˜1):1nf(m)+f1(y1)2Λ,1nf(m˜)+f1(y˜1)2Λdy1dy˜1·PY1|M(y1|m)PY1|M(y˜1|m˜)×[1y1:g(f(m)+f1(y1)+f(m˜)+f1(y˜1))m+1y1:g(f(m)+f1(y1)+f(m˜)+f1(y˜1))m˜]. (A138)

As the sum in the square brackets is at least 1 for all m˜m, it follows that

Pe(n)(q,C)κ2M2m=1Mm˜m(y1,y˜1):1nf(m)+f1(y1)2Λ,1nf(m˜)+f1(y˜1)2Λdy1dy˜1·PY1|M(y1n|m)PY1|M(y˜1|m˜)κ4·Pr1nf(M)+f1(Y1)2Λ,1nf(M˜)+f1(Y˜1)2Λ,M˜M. (A139)

Then, recall that by (A130), the expectation of 1nf(M)+f1(Y1n)2 is strictly lower than Λ, and for a sufficiently large n, the conditional expectation of 1n||f(M˜)+f1(Y˜1)||2 given M˜M is also strictly lower than Λ. Thus, by Chebyshev’s inequality, the probability of error is bounded from below by a positive constant. Following this contradiction, we deduce that if the code is reliable, then Λ(1+α+2ρα)Ω.

It is left for us to show that for α and ρ as defined in (A128), we have that R<FG(α,ρ) (see (36)). For a (2nR,n,εn*) code,

Pe|s(n)(C)εn*, (A140)

for all sRn with s2nΛ. Then, consider using the code C over the Gaussian relay channel WY,Y1|X,X1q¯, specified by

Y1=X+Z,Y=X+X1+S¯, (A141)

where the sequence S¯ is i.i.d. q¯=N(0,Λδ). First, we show that the code C is reliable for this channel, and then we show that R<FG(α,ρ). Using the code C over the channel WY,Y1|X,X1q¯, the probability of error is bounded by

Pe(n)(q¯,C)=Pr1nS¯>Λ+s:1nS¯Λds·Pe|s(n)(C)εn*+εn**, (A142)

where we have bounded the first term by εn** using the law of large numbers and the second term using (A140), where εn**0 as n. Since WY,Y1|X,X1q¯ is a channel without a state, we can now show that R<FG(α,ρ) by following the lines of [4] and [5]. By Fano’s inequality and [4] (Lemma 4), we have that

R1ni=1nIq¯(Xi,Xi,X1,i;Yi)+εn,R1ni=1nIq¯(Xi,Xi;Yi,Y1,i|X1,i)+εn, (A143)

where q¯=N(0,Λδ), X=f(M), X=f(M), X1=f1(y1), and εn0 as n. For the Gaussian relay channel with SFD, we have the following Markov relations,

Y1,iXi(Xi,X1,i,Y1,i), (A144)
(Xi,Y1,i)(Xi,X1,i)Yi. (A145)

Hence, by (A145), Iq¯(Xi,Xi,X1,i;Yi)=Iq¯(Xi,X1,i;Yi). Moving to the second bound in the RHS of (A143), we follow the lines of [5]. Then, by the mutual information chain rule, we have

Iq¯(Xi,Xi;Yi,Y1,i|X1,i)=I(Xi;Y1,i|X1,i)+I(Xi;Y1,i|Xi,X1,i)+Iq¯(Xi,Xi;Yi|X1,i,Y1,i)=(a)I(Xi;Y1,i|X1,i)+Iq¯(Xi,Xi;Yi|X1,i,Y1,i)=(b)[H(Y1,i|X1,i)H(Y1,i|Xi)]+[Hq¯(Yi|X1,i,Y1,i)Hq¯(Yi|Xi,X1,i)](c)Iq1(Xi;Y1,i)+I(Xi;Yi|X1,i) (A146)

where (a) is due to (A144), (b) is due to (A145), and (c) holds since conditioning reduces entropy. Introducing a time-sharing random variable KUnif[1:n], which is independent of X, X, X1, Y, Y1, we have that

RεnIq¯(XK,X1,K;YK|K)RεnI(XK;Y1,K|K)+Iq¯(XK;YK|X1,K,K). (A147)

Now, by the maximum differential entropy lemma (see e.g., [98] (Theorem 8.6.5)),

Iq¯(XK,X1,K;YK|K)12logE[(XK+X1,K)2]+(Λδ)Λδ=12log1+αΩ+α1Ω1+2ραΩ·α1Ω1Λδ (A148)

and

I(XK;Y1,K|K)+Iq¯(XK;YK|X1,K,K)12logEXK2+σ2σ2+12log1(E(XK·X1,K))2EXK2·EX1,K2EXK2+(Λδ)Λδ=12log1+(1α)Ωσ2+12log1+(1ρ2)αΩΛδ, (A149)

where α, α1 and ρ are given by (A128). Since δ>0 is arbitrary, and α11, the proof follows from (A147)–(A149). □

Author Contributions

Formal analysis, U.P.; Investigation, U.P.; Methodology, U.P.; Supervision, Y.S.; Writing—original draft, U.P.

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Van der Meulen E.C. Three-terminal communication channels. Adv. Appl. Probab. 1971;3:120–154. doi: 10.2307/1426331. [DOI] [Google Scholar]
  • 2.Kim Y.H. Coding techniques for primitive relay channels; Proceedings of the Allerton Conference on Communication, Control and Computing; Monticello, IL, USA. 26–28 September 2007; pp. 129–135. [Google Scholar]
  • 3.El Gamal A., Kim Y. Network Information Theory. Cambridge University Press; Cambridge, UK: 2011. [Google Scholar]
  • 4.Cover T., Gamal A.E. Capacity theorems for the relay channel. IEEE Trans. Inf. Theory. 1979;25:572–584. doi: 10.1109/TIT.1979.1056084. [DOI] [Google Scholar]
  • 5.Gamal A.E., Zahedi S. Capacity of a class of relay channels with orthogonal components. IEEE Trans. Inf. Theory. 2005;51:1815–1817. [Google Scholar]
  • 6.Xue F. A New Upper Bound on the Capacity of a Primitive Relay Channel Based on Channel Simulation. IEEE Trans. Inf. Theory. 2014;60:4786–4798. doi: 10.1109/TIT.2014.2328591. [DOI] [Google Scholar]
  • 7.Wu X., Özgür A. Cut-set bound is loose for Gaussian relay networks; Proceedings of the Allerton Conference on Communication, Control and Computing; Monticello, IL, USA. 29 September–2 October 2015; pp. 1135–1142. [Google Scholar]
  • 8.Chen Y., Devroye N. Zero-Error Relaying for Primitive Relay Channels. IEEE Trans. Inf. Theory. 2017;63:7708–7715. doi: 10.1109/TIT.2017.2760620. [DOI] [Google Scholar]
  • 9.Wu X., Özgür A. Cut-set bound is loose for Gaussian relay networks. IEEE Trans. Inf. Theory. 2018;64:1023–1037. doi: 10.1109/TIT.2017.2731963. [DOI] [Google Scholar]
  • 10.Wu X., Barnes L.P., Özgür A. “The Capacity of the Relay Channel”: Solution to Cover’s Problem in the Gaussian Case. IEEE Trans. Inf. Theory. 2019;65:255–275. doi: 10.1109/TIT.2018.2876892. [DOI] [Google Scholar]
  • 11.Mondelli M., Hassani S.H., Urbanke R. A New Coding Paradigm for the Primitive Relay Channel; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2018); Talisa Hotel in Vail, CO, USA. 17–22 June 2018; pp. 351–355. [Google Scholar]
  • 12.Ramachandran V. Gaussian degraded relay channel with lossy state reconstruction. AEU Int. J. Electr. Commun. 2018;93:348–353. doi: 10.1016/j.aeue.2018.06.029. [DOI] [Google Scholar]
  • 13.Chen Z., Fan P., Wu D., Xiong K., Letaief K.B. On the achievable rates of full-duplex Gaussian relay channel; Proceedings of the Global Communication Conference (GLOBECOM’2014); Austin, TX, USA. 8–12 December 2014; pp. 4342–4346. [Google Scholar]
  • 14.Kolte R., Özgür A., Gamal A.E. Capacity Approximations for Gaussian Relay Networks. IEEE Trans. Inf. Theory. 2015;61:4721–4734. doi: 10.1109/TIT.2015.2457904. [DOI] [Google Scholar]
  • 15.Jin X., Kim Y. The Approximate Capacity of the MIMO Relay Channel. IEEE Trans. Inf. Theory. 2017;63:1167–1176. doi: 10.1109/TIT.2016.2632625. [DOI] [Google Scholar]
  • 16.Wu X., Barnes L.P., Özgür A. The geometry of the relay channel; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2017); Aachen, Germany. 25–30 June 2017; pp. 2233–2237. [Google Scholar]
  • 17.Lai L., El Gamal H. The relay–eavesdropper channel: Cooperation for secrecy. IEEE Trans. Inf. Theory. 2008;54:4005–4019. doi: 10.1109/TIT.2008.928272. [DOI] [Google Scholar]
  • 18.Yeoh P.L., Yang N., Kim K.J. Secrecy outage probability of selective relaying wiretap channels with collaborative eavesdropping; Proceedings of the Global Communication Conference (GLOBECOM’2015); San Diego, CA, USA. 6–10 December 2015; pp. 1–6. [Google Scholar]
  • 19.Kramer G., van Wijngaarden A.J. On the white Gaussian multiple-access relay channel; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2000); Sorrento, Italy. 25–30 June 2000; p. 40. [Google Scholar]
  • 20.Schein B.E. Ph.D. Thesis. Massachusetts Institute of Technology; Cambridge, MA, USA: 2001. Distributed Coordination in Network Information Theory. [Google Scholar]
  • 21.Rankov B., Wittneben A. Achievable Rate Regions for the Two-way Relay Channel; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2006); Seattle, WA, USA. 9–14 July 2006; pp. 1668–1672. [Google Scholar]
  • 22.Gunduz D., Yener A., Goldsmith A., Poor H.V. The Multiway Relay Channel. IEEE Trans. Inf. Theory. 2013;59:51–63. doi: 10.1109/TIT.2012.2219156. [DOI] [Google Scholar]
  • 23.Maric I., Yates R.D. Forwarding strategies for Gaussian parallel-relay networks; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2004); Chicago, IL, USA. 27 June–2 July 2004; p. 269. [Google Scholar]
  • 24.Kochman Y., Khina A., Erez U., Zamir R. Rematch and forward for parallel relay networks; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2008); Toronto, ON, Canada. 6–11 July 2008; pp. 767–771. [Google Scholar]
  • 25.Awan Z.H., Zaidi A., Vandendorpe L. Secure communication over parallel relay channel. arXiv. 2010 doi: 10.1109/TIFS.2012.2185493.1011.2115 [DOI] [Google Scholar]
  • 26.Xue F., Sandhu S. Cooperation in a Half-Duplex Gaussian Diamond Relay Channel. IEEE Trans. Inf. Theory. 2007;53:3806–3814. [Google Scholar]
  • 27.Kang W., Ulukus S. Capacity of a class of diamond channels; Proceedings of the Allerton Conference on Communication, Control and Computing; Urbana-Champaign, IL, USA. 23–26 September 2008; pp. 1426–1431. [Google Scholar]
  • 28.Chern B., Özgür A. Achieving the Capacity of the N-Relay Gaussian Diamond Network Within logN Bits. IEEE Trans. Inf. Theory. 2014;60:7708–7718. doi: 10.1109/TIT.2014.2360843. [DOI] [Google Scholar]
  • 29.Sigurjónsson S., Kim Y.H. On multiple user channels with state information at the transmitters; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2005); Adelaide, Australia. 4–9 September 2005; pp. 72–76. [Google Scholar]
  • 30.Simeone O., Gündüz D., Shamai S. Compound relay channel with informed relay and destination; Proceedings of the Allerton Conference on Communication, Control and Computing; Monticello, IL, USA. 30 September–2 October 2009; pp. 692–699. [Google Scholar]
  • 31.Zaidi A., Vandendorpe L. Lower bounds on the capacity of the relay channel with states at the source. EURASIP J. Wirel. Commun. Netw. 2009;2009:1–22. doi: 10.1155/2009/634296. [DOI] [Google Scholar]
  • 32.Zaidi A., Kotagiri S.P., Laneman J.N., Vandendorpe L. Cooperative Relaying With State Available Noncausally at the Relay. IEEE Trans. Inf. Theory. 2010;56:2272–2298. doi: 10.1109/TIT.2010.2043782. [DOI] [Google Scholar]
  • 33.Zaidi A., Shamai S., Piantanida P., Vandendorpe L. Bounds on the capacity of the relay channel with noncausal state at the source. IEEE Trans. Inf. Theory. 2013;59:2639–2672. doi: 10.1109/TIT.2012.2229780. [DOI] [Google Scholar]
  • 34.Blackwell D., Breiman L., Thomasian A.J. The capacities of certain channel classes under random coding. Ann. Math. Stat. 1960;31:558–567. doi: 10.1214/aoms/1177705783. [DOI] [Google Scholar]
  • 35.Simon M.K., Alouini M.S. Digital Communication over Fading Channels. Volume 95 John Wiley & Sons; Hoboken, NJ, USA: 2005. [Google Scholar]
  • 36.Shamai S., Steiner A. A broadcast approach for a single-user slowly fading MIMO channel. IEEE Trans. Inf. Theory. 2003;49:2617–2635. doi: 10.1109/TIT.2003.817834. [DOI] [Google Scholar]
  • 37.Abdul Salam A., Sheriff R., Al-Araji S., Mezher K., Nasir Q. Novel Approach for Modeling Wireless Fading Channels Using a Finite State Markov Chain. ETRI J. 2017;39:718–728. doi: 10.4218/etrij.17.0117.0246. [DOI] [Google Scholar]
  • 38.Ozarow L.H., Shamai S., Wyner A.D. Information theoretic considerations for cellular mobile radio. IEEE Trans. Veh. Tech. 1994;43:359–378. doi: 10.1109/25.293655. [DOI] [Google Scholar]
  • 39.Goldsmith A.J., Varaiya P.P. Capacity of fading channels with channel side information. IEEE Trans. Inf. Theory. 1997;43:1986–1992. doi: 10.1109/18.641562. [DOI] [Google Scholar]
  • 40.Caire G., Shamai S. On the capacity of some channels with channel state information. IEEE Trans. Inf. Theory. 1999;45:2007–2019. doi: 10.1109/18.782125. [DOI] [Google Scholar]
  • 41.Zhou S., Zhao M., Xu X., Wang J., Yao Y. Distributed wireless communication system: A new architecture for future public wireless access. IEEE Commun. Mag. 2003;41:108–113. doi: 10.1109/MCOM.2003.1186553. [DOI] [Google Scholar]
  • 42.Xu Y., Lu R., Shi P., Li H., Xie S. Finite-time distributed state estimation over sensor networks with round-robin protocol and fading channels. IEEE Trans. Cybern. 2018;48:336–345. doi: 10.1109/TCYB.2016.2635122. [DOI] [PubMed] [Google Scholar]
  • 43.Kuznetsov A.V., Tsybakov B.S. Coding in a memory with defective cells. Probl. Peredachi Inf. 1974;10:52–60. [Google Scholar]
  • 44.Heegard C., Gamal A.E. On the capacity of computer memory with defects. IEEE Trans. Inf. Theory. 1983;29:731–739. doi: 10.1109/TIT.1983.1056723. [DOI] [Google Scholar]
  • 45.Kuzntsov A.V., Vinck A.J.H. On the general defective channel with informed encoder and capacities of some constrained memories. IEEE Trans. Inf. Theory. 1994;40:1866–1871. doi: 10.1109/18.340461. [DOI] [Google Scholar]
  • 46.Kim Y., Kumar B.V.K.V. Writing on dirty flash memory; Proceedings of the Allerton Conference on Communication, Control and Computing; Monticello, IL, USA. 30 September–3 October 2014; pp. 513–520. [Google Scholar]
  • 47.Bunin A., Goldfeld Z., Permuter H.H., Shamai S., Cuff P., Piantanida P. Key and message semantic-security over state-dependent channels. IEEE Trans. Inf. Forensic Secur. 2018 doi: 10.1109/TIFS.2018.2853108. [DOI] [Google Scholar]
  • 48.Gungor O., Koksal C.E., Gamal H.E. An information theoretic approach to RF fingerprinting; Proceedings of the Asilomar Conference on Signals, Systems and Computers (ACSSC’2013); Pacific Grove, CA, USA. 3–6 November 2013; pp. 61–65. [Google Scholar]
  • 49.Ignatenko T., Willems F.M.J. Biometric security from an information-theoretical perspective. Found. Trends® Commun. Inf. Theory. 2012;7:135–316. doi: 10.1561/0100000051. [DOI] [Google Scholar]
  • 50.Han G., Xiao L., Poor H.V. Two-dimensional anti-jamming communication based on deep reinforcement learning; Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); New Orleans, LA, USA. 5–9 March 2017. [Google Scholar]
  • 51.Xu W., Trappe W., Zhang Y., Wood T. The feasibility of launching and detecting jamming attacks in wireless networks; Proceedings of the ACM International Symposium on Mobile Ad Hoc Networking and Computing; Urbana-Champaign, IL, USA. 25–27 May 2005; pp. 46–57. [Google Scholar]
  • 52.Alnifie G., Simon R. A multi-channel defense against jamming attacks in wireless sensor networks; Proceedings of the ACM Workshop on QoS Security Wireless Mobile Networks; Crete Island, Greece. 22 October 2007; pp. 95–104. [Google Scholar]
  • 53.Padmavathi G., Shanmugapriya D. A survey of attacks, security mechanisms and challenges in wireless sensor networks. arXiv. 20090909.0576 [Google Scholar]
  • 54.Wang T., Liang T., Wei X., Fan J. Localization of Directional Jammer in Wireless Sensor Networks; Proceedings of the 2018 International Conference on Robots & Intelligent System (ICRIS) (ICRIS’2018); Changsha, China. 26–27 May 2018; pp. 198–202. [Google Scholar]
  • 55.Ahlswede R. Elimination of correlation in random codes for arbitrarily varying channels. Z. Wahrscheinlichkeitstheorie Verw. Gebiete. 1978;44:159–175. doi: 10.1007/BF00533053. [DOI] [Google Scholar]
  • 56.Ericson T. Exponential error bounds for random codes in the arbitrarily varying channel. IEEE Trans. Inf. Theory. 1985;31:42–48. doi: 10.1109/TIT.1985.1056995. [DOI] [Google Scholar]
  • 57.Csiszár I., Narayan P. The capacity of the arbitrarily varying channel revisited: Positivity, constraints. IEEE Trans. Inf. Theory. 1988;34:181–193. doi: 10.1109/18.2627. [DOI] [Google Scholar]
  • 58.Ahlswede R. Coloring hypergraphs: A new approach to multi-user source coding, Part 2. J. Comb. 1980;5:220–268. [Google Scholar]
  • 59.Ahlswede R. Arbitrarily varying channels with states sequence known to the sender. IEEE Trans. Inf. Theory. 1986;32:621–629. doi: 10.1109/TIT.1986.1057222. [DOI] [Google Scholar]
  • 60.Jahn J.H. Coding of arbitrarily varying multiuser channels. IEEE Trans. Inf. Theory. 1981;27:212–226. doi: 10.1109/TIT.1981.1056320. [DOI] [Google Scholar]
  • 61.Hof E., Bross S.I. On the deterministic-code capacity of the two-user discrete memoryless Arbitrarily Varying General Broadcast channel with degraded message sets. IEEE Trans. Inf. Theory. 2006;52:5023–5044. doi: 10.1109/TIT.2006.883543. [DOI] [Google Scholar]
  • 62.Winshtok A., Steinberg Y. The arbitrarily varying degraded broadcast channel with states known at the encoder; Proceedings of the 2006 IEEE International Symposium on Information Theory; Seattle, WA, USA. 9–14 July 2006; pp. 2156–2160. [Google Scholar]
  • 63.He X., Khisti A., Yener A. MIMO Multiple Access Channel With an Arbitrarily Varying Eavesdropper: Secrecy Degrees of Freedom. IEEE Trans. Inf. Theory. 2013;59:4733–4745. [Google Scholar]
  • 64.Pereg U., Steinberg Y. The arbitrarily varying degraded broadcast channel with causal side information at the encoder; Proceedings of the 2018 International Zurich Seminar Information Communication (IZS’2018); Aachen, Germany. 25–30 June 2018; pp. 20–24. [Google Scholar]
  • 65.Keresztfalvi T., Lapidoth A. Partially-robust communications over a noisy channel; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2018); Vail, CO, USA. 17–22 June 2018; pp. 2003–2006. [Google Scholar]
  • 66.Gubner J.A. Ph.D. Dissertation. University of Maryland; College Park, MD, USA: 1988. Deterministic Codes for Arbitrarily Varying Multiple-Access Channels. [Google Scholar]
  • 67.Gubner J.A. On the deterministic-code capacity of the multiple-access arbitrarily varying channel. IEEE Trans. Inf. Theory. 1990;36:262–275. doi: 10.1109/18.52472. [DOI] [Google Scholar]
  • 68.Gubner J.A. State constraints for the multiple-access arbitrarily varying channel. IEEE Trans. Inf. Theory. 1991;37:27–35. doi: 10.1109/18.61126. [DOI] [Google Scholar]
  • 69.Gubner J.A. On the capacity region of the discrete additive multiple-access arbitrarily varying channel. IEEE Trans. Inf. Theory. 1992;38:1344–1347. doi: 10.1109/18.144713. [DOI] [Google Scholar]
  • 70.Gubner J.A., Hughes B.L. Nonconvexity of the capacity region of the multiple-access arbitrarily varying channel subject to constraints. IEEE Trans. Inf. Theory. 1995;41:3–13. doi: 10.1109/18.370123. [DOI] [Google Scholar]
  • 71.Ahlswede R., Cai N. Arbitrarily Varying Multiple-Access Channels. Universität Bielefeld; Bielefeld, Germany: 1996. [Google Scholar]
  • 72.Ahlswede R., Cai N. Arbitrarily varying multiple-access channels. I. Ericson’s symmetrizability is adequate, Gubner’s conjecture is true. IEEE Trans. Inf. Theory. 1999;45:742–749. doi: 10.1109/18.749024. [DOI] [Google Scholar]
  • 73.He X., Khisti A., Yener A. MIMO multiple access channel with an arbitrarily varying eavesdropper; Proceedings of the Allerton Conference on Communication, Control and Computing (Allerton’2011); Monticello, IL, USA. 28–30 September 2011; pp. 1182–1189. [Google Scholar]
  • 74.Wiese M., Boche H. The arbitrarily varying multiple-access channel with conferencing encoders. IEEE Trans. Inf. Theory. 2013;59:1405–1416. doi: 10.1109/TIT.2012.2229779. [DOI] [Google Scholar]
  • 75.Nitinawarat S. On the Deterministic Code Capacity Region of an Arbitrarily Varying Multiple-Access Channel Under List Decoding. IEEE Trans. Inf. Theory. 2013;59:2683–2693. doi: 10.1109/TIT.2013.2242514. [DOI] [Google Scholar]
  • 76.MolavianJazi E., Bloch M., Laneman J.N. Arbitrary jamming can preclude secure communication; Proceedings of the Allerton Conference on Communication, Control and Computing; Monticello, IL, USA. 30 September–2 October 2009; pp. 1069–1075. [Google Scholar]
  • 77.Boche H., Schaefer R.F. Capacity results and super-activation for wiretap channels with active wiretappers. IEEE Trans. Inf. Theory. 2013;8:1482–1496. doi: 10.1109/TIFS.2013.2276049. [DOI] [Google Scholar]
  • 78.Aydinian H., Cicalese F., Deppe C. Information Theory, Combinatorics, and Search Theory. Springer; Berlin/Heidelberg, Germany: 2013. Chapter 5. [Google Scholar]
  • 79.Boche H., Schaefer R.F., Poor H.V. On arbitrarily varying wiretap channels for different classes of secrecy measures; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2014); Honolulu, HI, USA. 29 June–4 July 2014; pp. 2376–2380. [Google Scholar]
  • 80.Boche H., Schaefer R.F., Poor H.V. On the continuity of the secrecy capacity of compound and arbitrarily varying wiretap channels. IEEE Trans. Inf. Forensic Secur. 2015;10:2531–2546. doi: 10.1109/TIFS.2015.2465937. [DOI] [Google Scholar]
  • 81.Nötzel J., Wiese M., Boche H. The arbitrarily varying wiretap channel—Secret randomness, stability, and super-activation. IEEE Trans. Inf. Theory. 2016;62:3504–3531. doi: 10.1109/TIT.2016.2550587. [DOI] [Google Scholar]
  • 82.Goldfeld Z., Cuff P., Permuter H.H. Arbitrarily Varying Wiretap Channels with Type Constrained States. IEEE Trans. Inf. Theory. 2016;62:7216–7244. doi: 10.1109/TIT.2016.2619701. [DOI] [Google Scholar]
  • 83.He D., Luo Y. Arbitrarily varying wiretap channel with state sequence known or unknown at the receiver. arXiv. 20171701.02043 [Google Scholar]
  • 84.Boche H., Deppe C. Secure identification for wiretap channels; Robustness, super-additivity and continuity. IEEE Trans. Inf. Forensic Secur. 2018;13:1641–1655. doi: 10.1109/TIFS.2018.2797004. [DOI] [Google Scholar]
  • 85.Pereg U., Steinberg Y. The Arbitrarily Varying Channel Under Constraints with Side Information at the Encoder. IEEE Trans. Inf. Theory. 2019;65:861–887. doi: 10.1109/TIT.2018.2861776. [DOI] [Google Scholar]
  • 86.Pereg U., Steinberg Y. The arbitrarily varying channel under constraints with causal side information at the encoder; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2017); Aachen, Germany. 25–30 June 2017; pp. 2805–2809. [Google Scholar]
  • 87.Csiszár I., Narayan P. Capacity of the Gaussian arbitrarily varying channel. IEEE Trans. Inf. Theory. 1991;37:18–26. doi: 10.1109/18.61125. [DOI] [Google Scholar]
  • 88.Behboodi A., Piantanida P. On the simultaneous relay channel with informed receivers; Proceedings of the IEEE International Symposium on Information Theory (ISIT’2009); Seoul, Korea. 28 June–3 July 2009; pp. 1179–1183. [Google Scholar]
  • 89.Bjelaković I., Boche H., Sommerfeld J. Information Theory, Combinatorics, and Search Theory. Springer; Berlin/Heidelberg, Germany: 2013. Capacity results for arbitrarily varying wiretap channels; pp. 123–144. [Google Scholar]
  • 90.Sion M. On General Minimax Theorems. Pac. J. Math. 1958;8:171–176. doi: 10.2140/pjm.1958.8.171. [DOI] [Google Scholar]
  • 91.Pereg U., Steinberg Y. The arbitrarily varying gaussian relay channel with sender frequency division. arXiv. 20181805.12595 [Google Scholar]
  • 92.Hughes B., Narayan P. Gaussian arbitrarily varying channels. IEEE Trans. Inf. Theory. 1987;33:267–284. doi: 10.1109/TIT.1987.1057288. [DOI] [Google Scholar]
  • 93.Csiszár I., Körner J. Information Theory: Coding Theorems for Discrete Memoryless Systems. 2nd ed. Cambridge University Press; Cambridge, UK: 2011. [Google Scholar]
  • 94.Boyd S., Vandenberghe L. Convex Optimization. Cambridge University Press; Cambridge, UK: 2004. [Google Scholar]
  • 95.Blackwell D., Breiman L., Thomasian A.J. The capacity of a class of channels. Ann. Math. Stat. 1959;30:1229–1241. doi: 10.1214/aoms/1177706106. [DOI] [Google Scholar]
  • 96.Diggavi S.N., Cover T.M. The worst additive noise under a covariance constraint. IEEE Trans. Inf. Theory. 2001;47:3072–3081. doi: 10.1109/18.959289. [DOI] [Google Scholar]
  • 97.Shannon C. A mathematical theory of communication. Bell Syst. Tech. J. 1948;27:379–423, 623–656.. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
  • 98.Cover T.M., Thomas J.A. Elements of Information Theory. 2nd ed. Wiley; Hoboken, NJ, USA: 2006. [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES