Skip to main content
Entropy logoLink to Entropy
. 2021 Feb 24;23(3):265. doi: 10.3390/e23030265

Trade-offs between Error Exponents and Excess-Rate Exponents of Typical Slepian–Wolf Codes

Ran Tamir (Averbuch) 1,*, Neri Merhav 1
PMCID: PMC7995976  PMID: 33668181

Abstract

Typical random codes (TRCs) in a communication scenario of source coding with side information in the decoder is the main subject of this work. We study the semi-deterministic code ensemble, which is a certain variant of the ordinary random binning code ensemble. In this code ensemble, the relatively small type classes of the source are deterministically partitioned into the available bins in a one-to-one manner. As a consequence, the error probability decreases dramatically. The random binning error exponent and the error exponent of the TRCs are derived and proved to be equal to one another in a few important special cases. We show that the performance under optimal decoding can be attained also by certain universal decoders, e.g., the stochastic likelihood decoder with an empirical entropy metric. Moreover, we discuss the trade-offs between the error exponent and the excess-rate exponent for the typical random semi-deterministic code and characterize its optimal rate function. We show that for any pair of correlated information sources, both error and excess-rate probabilities exponential vanish when the blocklength tends to infinity.

Keywords: Slepian–Wolf coding, variable-rate coding, error exponent, excess-rate exponent, typical random code

1. Introduction

As is well known, the random coding error exponent is defined by

Er(R)=limn1nlogEPe(Cn), (1)

where R is the coding rate, Pe(Cn) is the error probability of a codebook Cn, and the expectation is with respect to (w.r.t.) the randomness of Cn across the ensemble of codes. The error exponent of the typical random code (TRC) is defined as [1]

Etrc(R)=limn1nElogPe(Cn). (2)

We believe that the error exponent of the TRC is the more relevant performance metric, as it captures the most likely error exponent of a randomly selected code, as opposed to the random coding error exponent, which is dominated by the relatively poor codes of the ensemble, rather than the channel noise, at relatively low coding rates. In addition, since in random coding analysis, the code is selected at random and remains fixed, it seems reasonable to study the performance of the chosen code itself instead of directly considering the ensemble performance.

To the best of our knowledge, not much is known about TRCs. In [2], Barg and Forney considered TRCs with independently and identically distributed codewords, along with typical linear codes, for the special case of the binary symmetric channel with maximum likelihood (ML) decoding. It was also shown that at a certain range of low rates, Etrc(R) lies between Er(R) and the expurgated exponent, Eex(R). In [3] Nazari et al. provided bounds on the error exponents of TRCs for both discrete memoryless channels (DMC) and multiple-access channels. In a recent article by Merhav [1], an exact single-letter expression has been derived for the error exponent of typical, random, fixed composition codes, over DMCs, and a wide class of (stochastic) decoders, collectively referred to as the generalized likelihood decoder (GLD). Later, Merhav studied error exponents of TRCs for the colored Gaussian channel [4], typical random trellis codes [5], and a Lagrange dual lower bound for the TRC exponent [6]. Large deviations around the TRC exponent were studied in [7].

While originally defined for pure channel coding [1,2,3], the notion of TRCs has natural analogues in other settings as well, such as source coding with side information in the decoder [8]. Typical random Slepian–Wolf (SW) code of a certain variant of the ordinary variable-rate random binning code ensemble is the main theme of this work. The random coding error exponent of SW coding, based on fixed-rate (FR) random binning, was first addressed by Gallager in [9], and improved later on by the expurgated bound in [10] and [11]. Variable-rate (VR) SW coding received less attention in the literature; VR codes under average rate constraint have been studied in [12] and proved to outperform FR codes in terms of error exponents. Optimum trade-offs between the error exponent and the excess-rate exponent in VR coding were analyzed in [13]. Sphere-packing upper bounds for source coding with side information in the FR and VR regimes were studied in [9] and [12], respectively. More works where exponential error bounds in source coding have been studied are [14,15,16,17,18].

It turns out that both the FR and VR ensembles suffer from an intrinsic deficiency, caused by statistical fluctuations in the sizes of the bins that are populated by the relatively small type classes of the source. This fundamental problem of the ordinary ensembles is alleviated in some variant of the ordinary VR ensemble-the semi-deterministic (SD) code ensemble, which has already been proposed and studied in its FR version in [18]. In the SD code ensemble, for source type classes which are exponentially larger than the space of the available bins, we just randomly assign each source sequence into one of the bins, as is done in ordinary random binning. Otherwise, for relatively small type classes, we deterministically order each source sequence into a different bin, which provides a one-to-one mapping. This way, all these relatively small source type classes do not contribute to the probability of error. The main results concerning the SD code are the following:

  1. The random binning error exponent and the error exponent of the TRC are derived in Theorems 1 and 2, respectively, and proved in Theorem 3 to be equal to one another in a few important special cases, which include the matched likelihood decoder, the MAP decoder, and the universal minimum entropy decoder. To the best of our knowledge, this phenomenon has not been seen elsewhere before, since the TRC exponent usually improves upon the random coding exponent. As a byproduct, we are able to provide a relatively simple expression for the TRC exponent.

  2. We prove in Theorem 4 that the error exponent of the TRC under MAP decoding is also attained by two universal decoders: the minimum entropy decoder and the stochastic entropy decoder, which is a GLD with an empirical conditional entropy metric. As far as we know, this result is first of its kind in source coding; in other scenarios, the random coding bound is attained also by universal decoders, but here, we find that the TRC exponent is also universally achievable. Moreover, while the likelihood decoder and the MAP decoder have similar error exponents [19], here we prove a similar result, but for two universal decoders (one stochastic and one deterministic) that share the same metric.

  3. We discuss the trade-offs between the error exponent and the excess-rate exponent for a typical random SD code, similarly to [13], but with a different notion of the excess-rate event, which takes into account the available side information. In Theorem 5, we provide an expression for the optimal rate function that guarantees a required level for the error exponent of the typical random SD code. Analogously, Theorem 6 proposes an expression for the optimal rate function that guarantees a required level for the excess-rate exponent. Furthermore, we show that for any pair of correlated information sources, the typical random SD code attains both exponentially vanishing error and excess-rate probabilities.

The remaining part of the paper is organized as follows. In Section 2, we establish notation conventions. In Section 3, we formalize the model, the coding technique, the main objectives of this work, and we review some background. In Section 4, we provide the main results concerning error exponents and universal decoding in the SD ensemble, and in Section 5, we discuss the trade-offs between the error exponent and the excess-rate exponent.

2. Notation Conventions

Throughout the paper, random variables will be denoted by capital letters, realizations will be denoted by the corresponding lower case letters, and their alphabets will be denoted by calligraphic letters. Random vectors and their realizations will be denoted, respectively, by boldface capital and lower case letters. Their alphabets will be superscripted by their dimensions. Sources and channels will be subscripted by the names of the relevant random variables/vectors and their conditionings, whenever applicable, following the standard notation conventions, e.g., QU, QV|U, and so on. When there is no room for ambiguity, these subscripts will be omitted. For a generic joint distribution QUV={QUV(u,v),uU,vV}, which will often be abbreviated by Q, information measures will be denoted in the conventional manner, but with a subscript Q; that is, HQ(U) is the marginal entropy of U, HQ(U|V) is the conditional entropy of U given V, and IQ(U;V)=HQ(U)HQ(U|V) is the mutual information between U and V. The Kullback–Leibler divergence between two probability distributions, QUV and PUV, is defined as

D(QUVPUV)=(u,v)U×VQUV(u,v)logQUV(u,v)PUV(u,v), (3)

where logarithms, here and throughout the sequel, are understood to be taken to the natural base. The probability of an event E will be denoted by P{E}, and the expectation operator w.r.t. a probability distribution Q will be denoted by EQ[·], where the subscript will often be omitted. For two positive sequences, {an} and {bn}, the notation anbn will stand for equality in the exponential scale, that is, limn(1/n)logan/bn=0. Similarly, an·bn means that lim supn(1/n)logan/bn0, and so on. The indicator function of an event A will be denoted by 𝟙{A}. The notation [t]+ will stand for max{0,t}.

The empirical distribution of a sequence uUn, which will be denoted by P^u, is the vector of relative frequencies, P^u(u), of each symbol uU in u. The type class of uUn, denoted T(u), is the set of all vectors u with P^u=P^u. When we wish to emphasize the dependence of the type class on the empirical distribution P^, we will denote it by T(P^). The set of all types of vectors of length n over U will be denoted by Pn(U), and the set of all possible types over U will be denoted by P(U)=n=1Pn(U). Information measures associated with empirical distributions will be denoted with ‘hats’ and will be subscripted by the sequences from which they are induced. For example, the entropy associated with P^u, which is the empirical entropy of u, will be denoted by H^u(U). Similar conventions will apply to the joint empirical distribution, the joint type class, the conditional empirical distributions and the conditional type classes associated with pairs (and multiples) of sequences of length n. Accordingly, P^uv would be the joint empirical distribution of (u,v)={(ui,vi)}i=1n, T(P^uv) will denote the joint type class of (u,v), T(P^u|v|v) will stand for the conditional type class of u given v, H^uv(U|V) will be the empirical conditional entropy, and so on. Likewise, when we wish to emphasize the dependence of empirical information measures upon a given empirical distribution Q, we denote them using the subscript Q, as described above.

3. Problem Formulation and Background

3.1. Problem Formulation

Let (U,V)={(Ut,Vt)}t=1n be n independent copies of a pair of random variables, (U,V)PUV, taking on values in finite alphabets, U and V, respectively. The vector U will designate the source vector to be encoded and the vector V will serve as correlated side information available to the decoder. In ordinary VR binning, the coding rate is not fixed for every uUn, but depends on its empirical distribution. Let us denote a rate function by R(·), which is a given continuous function from the probability simplex of U to the set of nonnegative reals. In that manner, for every type QUPn(U), all source sequences in T(QU) are randomly partitioned into enR(QU) bins. Every source sequence is encoded by its bin index, denoted by B(u), along with a header that indicates its type index, which requires only a negligible extra rate when n is large enough.

The SD code ensemble is a refinement of the ordinary VR code: for types with HQ(U)R(QU), i.e., type classes which are exponentially larger than the space of available bins, we just randomly assign each source sequence into one out of the enR(QU) bins. For the other types, we deterministically order each member of T(QU) into a different bin. This way, all type classes with HQ(U)<R(QU) do not contribute to the probability of error. The entire binning code of source sequences of blocklength n, i.e., the set {B(u)}uUn, is denoted by Bn. A sequence of SW codes, {Bn}n1, indexed by the block length n, will be denoted by B.

The decoder estimates u based on the bin index B(u), the type index T(u), and the side information sequence v, which is a realization of V. The optimal (MAP) decoder estimates u according to

u^=arg maxuB(u)T(u)P(u,v). (4)

As in [1,20], we consider here the GLD. The GLD estimates u stochastically, using the bin index B(u), the type index T(u), and the side information sequence v, according to the following posterior distribution

PU^=u|v,B(u),T(u)=exp{nf(P^uv)}u˜B(u)T(u)exp{nf(P^u˜v)}, (5)

where P^uv is the empirical distribution of (u,v) and f(·) is a given continuous, real valued functional of this empirical distribution. The GLD provides a unified framework which covers several important special cases, e.g., matched decoding, mismatched decoding, MAP decoding, and universal decoding (similarly to the α-decoders described in [11]). A more detailed discussion is given in [20].

The probability of error is the probability of the event {U^U}. For a given binning code Bn, the probability of error is given by

Pe(Bn)=u,vP(u,v)·𝟙H^u(U)R(P^u)·uB(u)T(u),uuexp{nf(P^uv)}u˜B(u)T(u)exp{nf(P^u˜v)}. (6)

For a given rate function, we derive the random binning exponent of this ensemble, which is defined by

Er(R(·))=limnlogE[Pe(Bn)]n, (7)

and compare it to the TRC exponent, which is

Etrc(R(·))=limnE[logPe(Bn)]n. (8)

Although it is unclear that the limits in (7) and (8) exist a priori, it will be evident from the analyses in Appendix A and Appendix B, respectively.

One way to define the excess-rate probability is as P{R(P^U)R}, where R is some target rate [13]. Due to the availability of side information in the decoder, it makes sense to require a target rate which depends on the pair (u,v). Since the lowest possible compression rate in this setting is given by HP(U|V) [8], given U=u and V=v, it is reasonable to adopt H^uv(U|V) as a reference rate. Hence, an alternative definition of the excess-rate probability of a code Bn, is as per(Bn,R(·),Δ)=P{R(P^U)H^UV(U|V)+Δ}, where Δ>0 is a redundancy threshold. (Note that the entire analysis remains intact if we allow a more general redundancy threshold as Δ=Δ(P^uv). This covers other alternatives for the excess-rate probability, e.g., P{R(P^U)R} or P{R(P^U)αH^U(U)}, α(0,1).) Accordingly, the excess-rate exponent function, achieved by a sequence of codes B, is defined as

Eer(B,R(·),Δ)=lim infn1nlogper(Bn,R(·),Δ). (9)

The main mission is to characterize the optimal trade-off between the error exponent and the excess-rate exponent for the typical random SD code, and the optimal rate function that attains a prescribed value for the error exponent of the typical random SD code.

3.2. Background

In pure channel coding, Merhav [1] has derived a single-letter expression for the error exponent of the typical random fixed composition code:

Etrc(R,QX)=limn1nElogPe(Cn). (10)

In order to present the main result of [1], we define first a few quantities. Consider a DMC, W={W(y|x),xX,yY}, where X and Y are the finite input/output alphabets. Define

α(R,QY)=max{QX˜|Y:IQ(X˜;Y)R,QX˜=QX}{g(QX˜Y)IQ(X˜;Y)}+R, (11)

where the function g(·), which is the decoding metric, is a continuous function that maps joint probability distributions over X×Y to real numbers. Additionally define

Γ(QXX,R)=minQY|XX{D(QY|XW|QX)+IQ(X;Y|X)+[max{g(QXY),α(R,QY)}g(QXY)]+}, (12)

where D(QY|XW|QX) is the conditional divergence between QY|X and W, averaged by QX. A brief intuitive explanation on the term Γ(QXX,R) can be found in [7] (Section 4.1). Having defined the above quantities, the error exponent of the TRC is given by [1]

Etrc(R,QX)=min{QX|X:IQ(X;X)2R,QX=QX}{Γ(QXX,R)+IQ(X;X)R}. (13)

Returning to the SW model, several articles have been written on error exponents for the FR and the VR codes. Here, we mention only those results that are directly relevant to the current work. The random binning and expurgated bounds of the FR ensemble in the SW model are given, respectively, by [11] (Section VI, Theorem 2), [10](Appendix I, Theorem 1)

Erfr(R)=minQUD(QUPU)+Er(QU,PV|U,HQ(U)R), (14)
Eexfr(R)=minQUD(QUPU)+Eex(QU,PV|U,HQ(U)R), (15)

where Er(QU,PV|U,S) and Eex(QU,PV|U,S) are, respectively, the random coding and expurgated bounds associated with the channel PV|U w.r.t. the ensemble of fixed composition code of rate S, whose composition is QU. The exponent function Er(QU,PV|U,S) is given by

Er(QU,PV|U,S)=minQV|U{D(QV|UPV|U|QU)+[IQ(U;V)S]+}, (16)

and Eex(QU,PV|U,S) is given by

Eex(QU,PV|U,S)=min{QU|U:IQ(U;U)S,QU=QU}{EQUU[dPV|U(U,U)]+IQ(U;U)S}, (17)

where

dPV|U(u,u)=logvVPV|U(v|u)PV|U(v|u). (18)

The exact error exponent of VR random binning is given by [13] (Equation (34)):

Ervr(R(·))=minQUVD(QUVPUV)+[R(QU)HQ(U|V)]+. (19)

4. Error Exponents and Universal Decoding

To present some of the results, we need a few more definitions. The minimum conditional entropy (MCE) decoder estimates u, using the bin index B(u), the type index T(u), and the side information vector v, according to

u^=arg minuB(u)T(u)H^uv(U|V). (20)

The stochastic conditional entropy (SCE) decoder is a special case of the GLD with the decoding metric f(P^uv)=H^uv(U|V); i.e., it estimates u according to the following posterior distribution

PU^=u|v,B(u),T(u)=exp{nH^uv(U|V)}u˜B(u)T(u)exp{nH^u˜v(U|V)}. (21)

First, we present random binning error exponents, which are modifications of (19) to this ensemble. Define the expression

E(QUV,R(·))=minQU|VR(QU)HQ(U|V)+[f(QUV)f(QUV)]++ (22)

and the exponent functions:

Er,GLD(R(·))=min{QUV:HQ(U)R(QU)}D(QUVPUV)+E(QUV,R(·)), (23)

and

Er,MAP(R(·))=min{QUV:HQ(U)R(QU)}D(QUVPUV)+[R(QU)HQ(U|V)]+. (24)

The following result is proved in Appendix A.

Theorem 1.

Let R(·) be a given rate function. Then, for the SD ensemble,

  1. Er(R(·))=Er,GLD(R(·)) for the GLD;

  2. Er(R(·))=Er,MAP(R(·)) for the MAP and MCE decoders.

As a matter of fact, a special case of the second part of Theorem 1 has already been proved in [18] for the FR regime, while here, we prove a stronger result, according to which, the MCE decoder attains the same random binning error exponent as the MAP decoder, in the VR coding regime too. The first part of Theorem 1 is completely new; it proposes a single letter expression for the random binning error exponent, for a wide family of stochastic and deterministic decoders. Additionally, note that an analogous result to the first part of Theorem 1 has been proved in [20]. Comparing the expressions in (19) and (24), namely, the random binning error exponents of the ordinary VR and the SD VR ensembles, respectively, we find that they differ at relatively high coding rates, since these minimization problems share the same objective but (24) also has the constraint HQ(U)R(QU). The origin of this constraint is the deterministic coding of the relatively small type classes.

Next, we provide a single-letter expression for the error exponent of the TRCs in this ensemble. We define

γ(R(·),QU,QV)=maxQU˜|V:QU˜=QU,HQ(U˜|V)R(QU˜){f(QU˜V)+HQ(U˜|V)}R(QU˜) (25)

and

Ψ(R(·),QUUV)=max{f(QUV),γ(R(·),QU,QV)}f(QUV)+. (26)

Furthermore, define

Λ(QUU,R(QU))=minQV|UUΨ(R(QU),QUUV)HQ(V|U,U)EQ[logP(V|U)], (27)

and the following exponent function:

Etrc,GLD(R(·))=minQUU:QU=QU,HQ(U)R(QU)Λ(QUU,R(QU))EQ[logP(U)]HQ(U,U)+R(QU). (28)

Then, the following theorem is proved in Appendix B.

Theorem 2.

Let R(·) be a given rate function. Then, for the SD ensemble and the GLD,

Etrc(R(·))=Etrc,GLD(R(·)). (29)

As explained before, an analogous result has already been proved in pure channel coding [1], and one can find a high degree of similarity between the expressions in (25)–(28) and the expressions in Section 3.2. While in channel coding, the coding rate is fixed, here, on the other hand, we allow the rate to depend on the type class of the source. In order to optimize the rate function, we constrain the problem by introducing the excess-rate exponent (9), which is the exponential rate of decay of the probability that the compression rate will be higher than some predefined level. A detailed discussion on optimal rate functions and optimal trade-offs between these two exponents can be found in Section 5.

The definition of the error exponent of the TRC as in (8) should not be taken for granted. The reason for that is the following. It turns out that the definition in (8) and the value of 1nlogPe(Bn) for the highly probable codes in the ensemble may not be the same, and they coincide if and only if the ensemble does not contain both zero error probability codes and positive error probability codes. For example, the FR ensemble in SW coding contains the one-to-one code (which obviously attains Pe(Bn)=0) as long as Rlog|U|, but it is definitely not a typical code, at least when ordinary random binning is considered. Hence, in this case, we conclude that 1nE[logPe(Bn)]=, while the value of 1nlogPe(Bn) for the highly probable codes is still finite. As for the SD code ensemble, the definition in (8) indeed provides the error exponent of the highly probable codes in the ensemble, which is explained by the following reasoning. For any given rate function such that R(QU)<HQ(U) for at least one type class, all the type classes with R(QU)<HQ(U) are encoded by random binning; thus, all the codes in the ensemble have a strictly positive error probability, which implies that the value of 1nlogPe(Bn) concentrates around the error exponent of the TRC, as defined in (8).

The proof of Theorem 2 follows exactly the same lines as the proof of ([1] (Theorem 1)), except for one main modification: when we introduce the type class enumerator N(QUU) (see below) and sum over joint types, the summation set becomes {QUU:QU=QU,HQ(U)R(QU)}, where the constraint HQ(U)R(QU) is due to the indicator function in (6). Afterwards, the analysis of the type class enumerator yields the constraint HQ(U,U)R(QU), which becomes redundant and thus omitted. This constraint is analogous to the constraint IQ(X;X)2R in the minimization of (13). The origin of HQ(U,U)R(QU) is the following. Define

N(QUU)=(u,u)T(QUU)𝟙B(u)=B(u), (30)

which enumerate pairs of source sequences. Then, one of the main steps in the proof of Theorem 2 is deriving the high probability value of N(QUU), which is 0 if HQ(U,U)<R(QU) (a relatively small set of source pair and relatively large number of bins) and en[HQ(U,U)R(QU)] for HQ(U,U)R(QU) (a large set of source sequence pair and a small number of bins). One should note that the analysis of N(QUU) is not trivial, since it is not a binomial random variable; i.e., the enumerator N(QUU) is given by the sum of dependent binary random variables. For a sum N of independent binary random variables, ordinary tools from large deviation theory (e.g., the Chernoff bound) can be invoked for assessing the exponential moments E[Ns], s0, or the large deviation rate function of P{Nenσ}, σIR. For sums of dependent binary random variables, such as N(QUU) in the current problem, this can no longer be done by the same techniques, and it requires more advanced tools (see, e.g., [1,4,5,6]).

It is possible to compare (23) and (28) analytically in the special cases of the matched or the mismatched likelihood decoders and the MCE decoder. In the following theorem, the choice f(QUV)=βEQ[logP˜(U,V)], where P˜(U,V) is a possibly different source distribution than P(U,V), corresponds to a family of stochastic mismatched decoders. We have the following result, the proof of which is given in Appendix D.

Theorem 3.

Consider the SD ensemble and a given rate function R(·). Then,

  1. For a GLD with the decoding metric f(Q)=βEQ[logP˜(U,V)], for a given β>0,
    Etrc,GLD(R(·))=Er,GLD(R(·)). (31)
  2. For the MCE decoder,
    Etrc,MCE(R(·))=Er,MCE(R(·)). (32)

This result is quite surprising at first glance, since one expects the error exponent of the TRC to be strictly better than the random binning error exponent, as in ordinary channel coding at relatively low coding rates [1,2]. This phenomenon is due to the fact that part of the source type classes are deterministically partitioned into bins in a one-to-one fashion, and hence do not affect the probability of error (notice that the constraint HQ(U)R(QU) appears in both the random binning and the TRC exponents, while in the latter, it makes the original constraint HQ(U,U)R(QU) redundant). In the cases of FR or ordinary VR binning, these relatively “thin” type classes dominated the error probability at relatively high binning rates, but now, by encoding them deterministically into the bins; other mechanisms dominate the error event, such as the channel noise (between U and V) or the random binning of the type classes with HQ(U)R(QU). The result of the second part of Theorem 3 is also nontrivial, since it establishes an equality between the error exponent of the TRC and the random binning error exponent, but now for a universal decoder.

Concerning universal decoding, it is already known [21] (Exercise 3.1.6), [13] that the random binning error exponents under optimal MAP decoding in both the FR and VR codes, given by (14) and (19), respectively, are also attained by the MCE decoder. Furthermore, a similar result for the SD ensemble has been proved here in Theorem 1. The natural question that arises is whether the error exponent of the TRC is also universally attainable. The following result, which is proved in Appendix E, provides a positive answer to this question.

Theorem 4.

Consider the SD ensemble and a given rate function R(·). Then, the error exponents of the TRC under the MAP, the MCE, and the SCE decoders are all equal; i.e.,

Etrc,MAP(R(·))=Etrc,MCE(R(·))=Etrc,SCE(R(·)). (33)

Theorem 4 asserts that the error exponent of the typical random SD code is not affected if the optimal MAP decoder is replaced by a certain universal decoder, which must not even be deterministic. While the left hand equality in (33) follows immediately from the results of Theorems 1 and 3, the right hand equality in (33) is far less trivial, since the SCE decoder is both universal and stochastic, and hence, its TRC exponent is expected to be inferior w.r.t. the TRC exponent under MAP decoding, but nevertheless, they turn out to be equal. Comparing to channel coding, it has been recently proved in [22] that the error exponent of the typical random fixed composition code (given in (13)) is the same for the ML and the maximum mutual information decoder, but on the other hand, numerical evidence shows that a GLD which is based on an empirical mutual information metric attains a strictly lower exponent.

5. Optimal Trade-off Functions

In this section, we study the optimal trade-off between the threshold Δ, the error exponent of the TRC, and the excess-rate exponent. Since both exponents depend on the rate function, we wish to characterize rate functions that are optimal w.r.t. this trade-off. Since a single-letter characterization of the error exponent of the TRC has already been given in (28), we next provide a single-letter expression for the excess-rate exponent. Define the following exponent function:

Eer(R(·),Δ)=min{QUV:R(QU)HQ(U|V)+Δ}D(QUVPUV). (34)

Then, we have the following.

Proposition 1.

Fix Δ>0 and let R(·) be any rate function. Then,

Eer(B,R(·),Δ)=Eer(R(·),Δ). (35)

Proof. 

The excess-rate probability is given by:

P{R(P^U)H^UV(U|V)+Δ}
=QUV𝟙{R(QU)HQ(U|V)+Δ}·P{(U,V)T(QUV)} (36)
{QUV:R(QU)HQ(U|V)+Δ}expnD(QUVPUV) (37)
expn·min{QUV:R(QU)HQ(U|V)+Δ}D(QUVPUV), (38)

which proves the desired result. □

Since Proposition 1 is proved by the method of types [21], we conclude that the excess-rate event is dominated by one specific type class T(QUV), whose respective rate R(QU) has been chosen too large w.r.t. the value of HQ(U|V)+Δ. One extreme case is when the rate function is given by HQ(U), which obviously provides a one-to-one mapping, since the size of each T(QU) is upper-bounded by enHQ(U). In this case, the probability of error is zero, while the excess-rate probability is one, at least when Δ is not too large. In Section 5.2, we prove that the optimal rate function is indeed upper-bounded by HQ(U), but can also be strictly smaller, especially when the requirement on the error exponent is not too stringent.

One way to explore the trade-off between the error exponent of the TRC and the excess-rate exponent, which will be presented in Section 5.1, is to require the excess-rate exponent to exceed some value Er>0, then solve Eer(R(·),Δ)Er for an optimal rate function R*(QU), and then to substitute this optimal rate function back into the error exponents in (24) and (28) to give expressions for the optimal trade-off function Ee(Er,Δ). In Section 5.2, we present an alternative option to characterize this trade-off, which is to require the error exponent of the TRC to exceed some value Ee>0, to solve Ee(R(·))Ee in order to extract an optimal rate function, and then to substitute it back into the excess-rate exponent in (34) to provide an expression for the optimal trade-off function Eer(Ee,Δ).

5.1. Constrained Excess-Rate Exponent

Relying on the exponent function in (34), the following theorem proposes a rate function, whose optimality is proved in Appendix F.

Theorem 5.

Let Er>0 be fixed. Then, the constraint Eer(R(·),Δ)Er implies that

R(QU)J(QU,Er,Δ)=min{QV|U:D(QUVPUV)Er}HQ(U|V)+Δ. (39)

This means that we have a dichotomy between two kinds of source types. Each type class that is associated with an empirical distribution that is relatively close to the source distribution, i.e., when D(QUVPUV)Er for some QV|U, is partitioned into enJ(QU,Er,Δ) bins, and the rest of the type classes, those that are relatively distant from PU, are encoded by a one-to-one mapping. Two extreme cases should be considered here. First, when Er is relatively small, then only the types closest to PU are encoded with a rate approximately HP(U|V)+Δ, which can be made arbitrarily close to the SW limit [8], and each a–typical source sequence is allocated with n·log2|U| bits. This coding scheme is the one related to VR coding with an average rate constraint, like the one discussed in [12]. Second, when Er is extremely large, then each type class is encoded to exp{nΔ} bins, which is equivalent to FR coding.

Following the first part of Theorem 3, let us denote the error exponent of the TRC under MAP decoding by Ee(·). Upon substituting the optimal rate function of Theorem 5 back into (24) and (28) and using the fact that Ee(·) is monotonically increasing, we find that the optimal trade-off function for the typical random SD code is given by

Ee(Er,Δ)=minQUV:HQ(U)J(QU)D(QUVPUV)+[J(QU)HQ(U|V)]+, (40)

or, alternatively,

Ee(Er,Δ)=minQUU:QU=QU,HQ(U)J(QU)Λ(QUU,J(QU))EQ[logP(U)]HQ(U,U)+J(QU), (41)

where J(QU)=J(QU,Er,Δ) is given in (39). The dependence of Ee(Er,Δ) on Er is as follows. Let QUU*(Δ) and QV|U* be the respective minimizers of the problems which are similar to (39) and (41), except that the constraint D(QUVPUV)Er is removed from (39). Furthermore, let QU*(Δ) be the marginal distribution of QUU*(Δ). Now, when Er is sufficiently large, i.e., when ErD(QU*(Δ)×QV|U*PUV), Ee(Er,Δ) reaches a plateau and is the lowest possible. It follows from the fact that the stringent requirement on the excess-rate forces the encoder to encode each type class QU to its target rate Δ, thus all of them affect the error event. Otherwise, when Er<D(QU*(Δ)×QV|U*PUV), the constraint D(QUVPUV)Er is active and Ee(Er,Δ) is a monotonically nonincreasing function of Er. The reason for that is the fact that as Er decreases, more and more type classes are encoded with n·log2|U| bits, and hence do not contribute to the error event. When Er=0, necessarily QU=PU, only the typical set is encoded, and Ee(0,Δ) is the highest possible. In this case, J(QU)=HP(U|V)+Δ and the constraint set in (41) becomes empty when Δ>IP(U;V), and then Ee(0,Δ)=.

5.2. Constrained Error Exponent

Based on (24), the following theorem proposes a rate function, whose optimality is proved in Appendix G.

Theorem 6.

Let Ee>0 be fixed. Then, the constraint Ee(R(·))Ee implies that

R(QU)Ω(QU,Ee)=minHQ(U),G(QU,Ee), (42)

where,

G(QU,Ee)=max{QV|U:D(QUVPUV)Ee}{HQ(U|V)+EeD(QUVPUV)}. (43)

The dependence of G(QU,Ee) on Ee is as follows. For any given QU, let Q˜V|U be the minimizer of D(QUVPUV). Then, as long as Ee<D(QU×Q˜V|UPUV), the constraint set in (43) is empty, and R(QU) can vanish, which practically means that in this range, the entire type class T(QU) can be totally ignored, while still achieving PeenEe. Only for the unique type QU=PU, G(PU,Ee)>0 for all Ee0, and specifically, we find that G(PU,0)=HP(U|V). Furthermore, let QV|U* be the maximizer in the unconstrained problem

maxQV|UHQ(U|V)D(QUVPUV). (44)

Then, as long as Ee[D(QU×Q˜V|UPUV),D(QU×QV|U*PUV)), G(QU,Ee) is a monotonically nondecreasing function of Ee. When EeD(QU×QV|U*PUV), the maximization in (43) reaches its unconstrained optimum, and G(QU,Ee) increases without bound in an affine fashion as Ee+HQ*(U|V)D(QU×QV|U*PUV). As can be seen in (42), Ω(QU,Ee) finally reaches a plateau at the level of HQ(U).

Upon substituting Ω(QU,Ee) back into (34) and using the fact that Eer(·,Δ) is monotonically nonincreasing, we find that the trade-off function is given by

Eer(Ee,Δ)=min{QUV:Ω(QU,Ee)HQ(U|V)+Δ}D(QUVPUV). (45)

Since Ω(QU,Ee) is monotonically nondecreasing in Ee for every QU, Eer(Ee,Δ) is monotonically nonincreasing in Ee, which is not very surprising. The dependence of Eer(Ee,Δ) on Ee and Δ is as follows. At Ee=0, notice that Ω(QU,0)= for any QUPU while Ω(PU,0)=HP(U|V). Thus, Eer(0,Δ)=0 as long as Δ=0, and it follows from the monotonicity that Eer(Ee,0)=0 everywhere. Otherwise, if Δ>0, {QUV:Ω(QU,Ee)HQ(U|V)+Δ} is empty as long as Ee<Ee*(Δ), where an expression for Ee*(Δ) can be found by solving

maxQUV{Ω(QU,Ee)HQ(U|V)}Δ, (46)

and then Eer(Ee,Δ)= in this range. In the other extreme case of a very large Ee, Ω(QU,Ee) reaches a plateau at a level of HQ(U). Then, if ΔHP(U)HP(U|V)=IP(U;V), Eer(Ee,Δ) reaches zero for a sufficiently large Ee. Else, if Δ>IP(U;V), Eer(Ee,Δ) reaches a strictly positive plateau, given by

min{QUV:IQ(U;V)Δ}D(QUVPUV), (47)

which is a monotonically nondecreasing function of Δ. Particularly, it means that in this range, the typical random SD code attains both an exponentially vanishing excess-rate probability and Pe0.

It is interesting to relate this to the expurgated bound of the FR code in the SW model, which is given by (15). Comparing Eexfr(R) and Ee(,Δ) analytically is rather difficult. Thus, we examined these two exponent functions numerically. Consider the case of a double binary source with alphabets U=V={0,1}, and joint probabilities given by PUV(0,0)=0.75, PUV(0,1)=0.1, PUV(1,0)=0, and PUV(1,1)=0.15. We already mentioned before, that in the special case of Er=, the rate function is given by the threshold Δ, hence we choose Δ=R in order to have a fair comparison. Graphs of the functions Eexfr(R) and Ee(,R) are presented in Figure 1.

Figure 1.

Figure 1

Graphs of the functions Eexfr(R) and Ee(,R).

As can be seen in Figure 1, both Eexfr(R) and Ee(,R) tend to infinity as R tends to log20.693. For relatively high binning rates, Eexfr(R) is strictly higher than Ee(,R), which can be explained in the following way: Referring to the analogy between SW coding and channel coding, one can think of each bin as containing a channel code. In general, a channel code behaves well if it does not contain pairs of relatively “close” codewords. Since we randomly assign the source vectors into the bins (even if the populations of the bins are totally equal, which can be attained by randomly partitioning each type class into exp{nR} subsets), it is reasonable to assume that some bins will contain relatively bad codebooks. On the other hand, in the expurgated SW code [11], each type class T(QU) is partitioned into exp{nR} “balanced” subsets in some sense (referring to the enumerators N(QUU) in (30), they are equally populated in all of the bins), such that the codebooks contained in the bins have approximately equal error probabilities. Moreover, we conclude from (15) that each bin contains a codebook with a quality of an expurgated channel code. This code is certainly better than the TRCs in the SD ensemble.

In channel coding, it is known [23] that the random Gilbert–Varshamov ensemble has an exact random coding error exponent which is as high as the maximum between (16) and (17). In SW source coding, on the other hand, it seems to be a more challenging problem to define an ensemble, such that the error exponent of its TRCs is as high as Eexfr(R) of (15). Since the gap between Eexfr(R) and Ee(,R) is not necessarily very significant, as can be seen in Figure 1, we conclude that the SD ensemble may be more attractive because the amount of computations needed for drawing a code from it are much lower than the amount of computations required for having an expurgated SW code. In addition, it is important to note that the probability of drawing a SD code with an exponent much lower than Ee(,R) decays exponentially fast, in analogy to the result in pure channel coding [7].

Appendix A. Proof of Theorem 1

By definition, we have

E[Pe(Bn)]=EuB(U),uUexp{nf(P^uV)}u˜B(U)exp{nf(P^u˜V)}. (A1)

Step 1: Averaging Over the Random Code

We first condition on the true source sequences (U=u,V=v) and take the expectation only w.r.t. the random binning. We get

E[Pe(Bn)|u,v]
=EuB(u),uuexp{nf(P^uv)}exp{n·f(P^uv)}+uB(u),uuexp{nf(P^uv)} (A2)
=01PuB(u),uuexp{nf(P^uv)}exp{n·f(P^uv)}+uB(u),uuexp{nf(P^uv)}sds (A3)
=0nenξ·PuB(u),uuexp{nf(P^uv)}exp{n·f(P^uv)}+uB(u),uuexp{nf(P^uv)}enξdξ (A4)
=0nenξ·P(1enξ)uB(u),uuexp{nf(P^uv)}enξexp{nf(P^uv)}dξ (A5)
0nenξ·PuB(u),uuexp{nf(P^uv)}exp{n[f(P^uv)ξ]}dξ, (A6)

where (A4) follows by changing the integration variable in (A3) according to s=enξ. Define

Nu,v(QU|V)=uB(u),uu𝟙{uT(QU|V|v)}, (A7)

such that the probability in (A6) is given by

PuB(u),uuexp{nf(P^uv)}exp{n[f(P^uv)ξ]}
=PQU|VNu,v(QU|V)exp{nf(QUV)}exp{n[f(P^uv)ξ]} (A8)
PmaxQU|VNu,v(QU|V)exp{nf(QUV)}exp{n[f(P^uv)ξ]} (A9)
=PQU|VNu,v(QU|V)exp{nf(QUV)}exp{n[f(P^uv)ξ]} (A10)
QU|VPNu,v(QU|V)exp{n[f(P^uv)f(QUV)ξ]}, (A11)

where QUV=QU|V×P^v. Let us denote B0=f(P^uv)f(QUV). Now, given u and v, Nu,v(QU|V) is a binomial sum of |T(QU|V|v)|enHQ(U|V) trials and success rate of the exponential order of enR(QU). Therefore, using the techniques of [24] (Section 6.3),

1nlogPNu,v(QU|V)exp{n[B0ξ]}
=R(QU)HQ(U|V)+HQ(U|V)R(QU)+B0ξHQ(U|V)R(QU)+<B0ξ (A12)
=R(QU)HQ(U|V)+ξB0HQ(U|V)R(QU)+ξ<B0HQ(U|V)R(QU)+, (A13)

and so,

0enξ·PNu,v(QU|V)exp{n[B0ξ]}dξ
B0[HQ(U|V)R(QU)]++enξ·en[R(QU)HQ(U|V)]+dξ (A14)
expn[R(QU)HQ(U|V)]++B0[HQ(U|V)R(QU)]++ (A15)
=expnR(QU)HQ(U|V)+[B0]+R(QU)HQ(U|V)R(QU)HQ(U|V)+B0+R(QU)<HQ(U|V) (A16)
=expnR(QU)HQ(U|V)+[B0]++R(QU)HQ(U|V)R(QU)HQ(U|V)+[B0]++R(QU)<HQ(U|V) (A17)
=expn·R(QU)HQ(U|V)+[B0]++. (A18)

Finally, we have that

QU|V0enξ·PNu,v(QU|V)exp{n[B0ξ]}dξ (A19)
maxQU|Vexpn·R(QU)HQ(U|V)+[B0]++ (A20)
=expn·minQU|VR(QU)HQ(U|V)+[B0]++, (A21)

thus,

E(u,v)=minQU|VR(QU)HQ(U|V)+[f(P^uv)f(QUV)]++. (A22)

Step 2: Averaging Over U and V

Notice that the exponent function E(u,v) depends on (u,v) only via the empirical distribution P^uv. Averaging over the source and the side information sequences, now yields

EPe(Bn)=u,vP(u,v)·𝟙H^u(U)R(P^u)·expn·E(P^uv) (A23)
{QUV:HQ(U)R(QU)}en·D(QUVPUV)·expn·E(QUV) (A24)
expn·min{QUV:HQ(U)R(QU)}D(QUVPUV)+E(QUV), (A25)

which proves the first point of Theorem 1.

Step 3: Moving from Stochastic to Deterministic Decoding

In order to transform the GLD into the general deterministic decoder of

u^=arg maxuB(u)T(u)f(P^uv), (A26)

we just have to multiply f(·), in

E(QUV)=minQU|VR(QU)HQ(U|V)+[f(QUV)f(QUV)]++, (A27)

by β0, and then let β. We find that the overall error exponent of the SD ensemble with the general deterministic decoder of (A26) is given by

E(P)=min{QUV:HQ(U)R(QU)}D(QUVPUV)+E˜(QUV), (A28)

where,

E˜(QUV)=min{QU|V:f(QUV)f(QUV)}R(QU)HQ(U|V)+. (A29)

Step 4: A Fundamental Limitation on the Error Exponent

Note that the minimum in (A29) can be upper–bounded by choosing a specific distribution in the feasible set. In (A29), we take QU|V=QU|V and then

E˜(QUV)R(QU)HQ(U|V)+. (A30)

Hence, the overall error exponent is upper–bounded as

E(P)min{QUV:HQ(U)R(QU)}D(QUVPUV)+R(QU)HQ(U|V)+. (A31)

Step 5: An Optimal Universal Decoder

We prove that the upper bound of (A31) is attainable by choosing the universal decoding metric f(QUV)=HQ(U|V). Now, we get for (A29)

E˜(QUV)=min{QU|V:f(QUV)f(QUV)}R(QU)HQ(U|V)+ (A32)
=min{QU|V:HQ(U|V)HQ(U|V)}R(QU)HQ(U|V)+ (A33)
=R(QU)HQ(U|V)+, (A34)

which completes the proof of Theorem 1.

Appendix B. Proof of Theorem 2

Lower Bound on the Error Exponent

Our starting point is the following inequality, for any ρ>0,

E[logPe(Bn)]logE[Pe(Bn)]1/ρρ, (A35)

which is due to the following considerations. First, for a positive random variable X, the function

f(ρ)=logEX1/ρρ (A36)

is monotonically decreasing, and second, by L’Hospital’s rule,

limρlogEX1/ρρ=E[logX]. (A37)

Recall that the error probability is given by

Pe(Bn)=u,vP(u,v)·𝟙H^u(U)R(P^u)·uB(u)T(u),uuexp{nf(P^uv)}u˜B(u)T(u)exp{nf(P^u˜v)}. (A38)

Let

Zu(v)=u˜B(u)T(u),u˜uexp{nf(P^u˜v)}, (A39)

fix ϵ>0 arbitrarily small, and for every uUn and vVn, define the set

Bϵ(u,v)=Bn:Zu(v)exp{nα(R+ϵ,P^u,P^v)}. (A40)

Following the result of [20] (Appendix B), we prove the following modification in Appendix C.

Lemma A1.

Let ϵ>0 be arbitrarily small. Then, for every uUn and vVn,

PZu(v)exp{nα(R+ϵ,P^u,P^v)}exp{enϵ+nϵ+1}. (A41)

Thus, by the union bound,

PuUnvVnBϵ(u,v)=PBϵuUnvVnPBϵ(u,v) (A42)
uUnvVnexp{enϵ+nϵ+1} (A43)
=|U×V|n·exp{enϵ+nϵ+1}, (A44)

which still decays double–exponentially fast. Recall that Q={QUU:QU=QU}. Then, for any ρ1

EPe(Bn)1/ρ
=EPe(Bn)1/ρ·𝟙{Bϵc}+EPe(Bn)1/ρ·𝟙{Bϵ}Eu,vP(u,v)𝟙H^u(U)R(P^u)uB(u)T(u),uuexp{nf(P^uv)}exp{nf(P^uv)}+Zu(v)1/ρ𝟙{Bϵc} (A45)
+P{Bϵ}Eu,vuB(u)T(u),uuP(u,v)𝟙H^u(U)R(P^u) (A46)
×min1,exp{nf(P^uv)}exp{nf(P^uv)}+exp{nα(R+ϵ,P^u,P^v)}1/ρ
+|U×V|n·exp{enϵ+nϵ+1}EuuB(u)T(u),uuP(u)𝟙H^u(U)R(P^u) (A47)
×vP(v|u)expn·max{f(P^uv),α(R+ϵ,P^u,P^v)}f(P^uv)+1/ρ (A48)
EuuB(u)T(u),uuP(u)·𝟙H^u(U)R(P^u)·expn·Λ(P^uu,R+ϵ)1/ρ (A49)
=E{QUUQ:HQ(U)R(QU)}N(QUU)·enEQ[logP(U)]·expn·Λ(QUU,R+ϵ)1/ρ (A50)
{QUUQ:HQ(U)R(QU)}EN(QUU)1/ρ·en(EQ[logP(U)])/ρ·expn·Λ(QUU,R+ϵ)/ρ, (A51)

where (A47) is due to Lemma A1, (A49) is by the method of types and the definition of Λ(QUU,R) in (27), and in (A50) we used the definition of N(QUU) in (30). Therefore, our next task is to evaluate the 1/ρ–th moment of N(QUU). Let us define

Nu(QU|U)=uT(QU|U|u)𝟙B(u)=B(u). (A52)

For a given ρ1, let s[1,ρ]. Then,

EN(QUU)1/ρ=EuT(QU)Nu(QU|U)1/ρ (A53)
=EuT(QU)Nu(QU|U)1/ss/ρ (A54)
EuT(QU)Nu(QU|U)1/ss/ρ (A55)
EuT(QU)Nu(QU|U)1/ss/ρ (A56)
=uT(QU)ENu(QU|U)1/ss/ρ, (A57)

where (A56) follows from Jensen’s inequality. Now, Nu(QU|U) is a binomial random variable with |T(QU|U|u)|enHQ(U|U) trials and success rate which is of the exponential order of enR. We have that [24](Section 6.3)

ENu(QU|U)1/sexp{n[HQ(U|U)R]/s}HQ(U|U)Rexp{n[HQ(U|U)R]}HQ(U|U)<R, (A58)

and so,

EN(QUU)1/ρenHQ(U)·s/ρ·ENu(QU|U)1/ss/ρ (A59)
enHQ(U)·s/ρ·exp{n[HQ(U|U)R]/ρ}HQ(U|U)Rexp{n[HQ(U|U)R]s/ρ}HQ(U|U)<R (A60)
=exp{n[HQ(U)·s+HQ(U|U)R]/ρ}HQ(U|U)Rexp{n[HQ(U)+HQ(U|U)R]s/ρ}HQ(U|U)<R (A61)
=exp{n[HQ(U)·s+HQ(U|U)R]/ρ}HQ(U|U)Rexp{n[HQ(U,U)R]s/ρ}HQ(U|U)<R. (A62)

After optimizing over s, we get

1nlogEN(QUU)1/ρ
min1sρHQ(U)·s+HQ(U|U)R/ρHQ(U|U)RHQ(U,U)Rs/ρHQ(U|U)<R,HQ(U,U)RHQ(U,U)Rs/ρHQ(U|U)<R,HQ(U,U)<R (A63)
=HQ(U)+HQ(U|U)R/ρHQ(U|U)RHQ(U,U)R/ρHQ(U|U)<R,HQ(U,U)RHQ(U,U)Rρ/ρHQ(U|U)<R,HQ(U,U)<R (A64)
=HQ(U,U)R/ρHQ(U,U)RHQ(U,U)RHQ(U,U)<R, (A65)

which gives, after raising to the ρ–th power,

EN(QUU)1/ρρexp{nHQ(U,U)R}HQ(U,U)Rexp{nHQ(U,U)R·ρ}HQ(U,U)<R (A66)
=exp{n([HQ(U,U)R]+ρ[RHQ(U,U)]+)}. (A67)

Let us denote F(Q,R,ρ)=[HQ(U,U)R]+ρ[RHQ(U,U)]+. Continuing now from (A51),

EPe(Bn)1/ρρ
·{QUUQ:HQ(U)R(QU)}EN(QUU)1/ρ·en[EQlogP(U)]/ρ·expn·Λ(QUU,R+ϵ)/ρρ (A68)
{QUUQ:HQ(U)R(QU)}EN(QUU)1/ρρ·enEQlogP(U)·expn·Λ(QUU,R+ϵ) (A69)
{QUUQ:HQ(U)R(QU)}exp{n(F(Q,R,ρ)+EQ[logP(U)]Λ(QUU,R+ϵ))} (A70)
expn·min{QUUQ:HQ(U)R(QU)}(Λ(QUU,R+ϵ)F(Q,R,ρ)EQ[logP(U)]). (A71)

where (A70) follows from (A67). Finally, it follows by (A35) that

lim infn1nE[logPe(Bn)]
lim infn1nlogE[Pe(Bn)]1/ρρ (A72)
min{QUUQ:HQ(U)R(QU)}(Λ(QUU,R+ϵ)F(Q,R,ρ)EQ[logP(U)]). (A73)

Letting ρ grow without bound yields that

lim infn1nE[logPe(Bn)]min{QUUQ:HQ(U)R(QU),HQ(U,U)R(QU)}(Λ(QUU,R+ϵ)HQ(U,U)
+R(QU)EQ[logP(U)]) (A74)
=min{QUUQ:HQ(U)R(QU)}(Λ(QUU,R+ϵ)HQ(U,U)+R(QU)EQ[logP(U)]). (A75)

Due to the arbitrariness of ϵ>0, we have proved that

lim infn1nE[logPe(Bn)]min{QUUQ:HQ(U)R(QU)}(Λ(QUU,R)HQ(U,U)+R(QU)EQ[logP(U)]). (A76)

completing half of the proof of Theorem 2.

Upper Bound on the Error Exponent

Consider a joint distribution QUU, that satisfies HQ(U,U)>R, and define the event E(QUU)={Bn:N(QUU)<exp{n[HQ(U,U)Rϵ]}}. We want to show that P{E(QUU)} is small. Consider the following:

P{E(QUU)}=P{N(QUU)<exp{n[HQ(U,U)Rϵ]}} (A77)
=P{N(QUU)<enϵ·E{N(QUU)}} (A78)
=PN(QUU)E{N(QUU)}1<(1enϵ) (A79)
PN(QUU)E{N(QUU)}E{N(QUU)}2>(1enϵ)2 (A80)
Var{N(QUU)}(1enϵ)2·E2{N(QUU)}. (A81)

Let us use the shorthand notations I(u,u)=𝟙B(u)=B(u), K=|T(QUU)|, and p=enR. Concerning the variance of N(QUU), we have the following

Var{N(QUU)}
=E{N2(QUU)}E2{N(QUU)} (A82)
=E(u,u)T(QUU)I(u,u)×(u˜,u^)T(QUU)I(u˜,u^)(Kp)2 (A83)
=(u,u)T(QUU)(u˜,u^)T(QUU)EI(u,u)I(u˜,u^)(Kp)2 (A84)
=(u,u)T(QUU)EI2(u,u)+(u,u),(u˜,u^)T(QUU)(u,u)(u˜,u^)EI(u,u)I(u˜,u^)(Kp)2 (A85)
=Kp+K(K1)p2(Kp)2 (A86)
=Kp(1p) (A87)
exp{n[HQ(U,U)R]}, (A88)

and hence,

P{E(QUU)}·exp{n[HQ(U,U)R]}exp{n[2HQ(U,U)2R]} (A89)
=exp{n[HQ(U,U)R]}, (A90)

which decays to zero since we have assumed that HQ(U,U)>R. Furthermore, if HQ(U,U)R+ϵ, then P{E(QUU)} tends to zero at least as fast as enϵ. Now, for a given ϵ>0, and a given joint type QUUV, such that HQ(U,U)R+ϵ, let us define

Zuu(v)=u˜B(u)T(u),u˜u,uexp{nf(P^u˜v)}, (A91)

and

Gn(QUUV)={Bn:(u,u)T(QUU)𝟙{B(u)=B(u)}×vT(QV|UU|u,u)𝟙Zuu(v)en[α(R2ϵ,QU,QV)+ϵ]exp{n[HQ(U,U)R3ϵ/2]}·|T(QV|UU|u,u)|}, (A92)

where (u,u) in the expression |T(QV|UU|u,u)| should be understood as any pair of source sequences in T(QUU). Next, we define

Gn={QUUV:HQ(U,U)R+ϵ}[Gn(QUUV)Ec(QUU)]. (A93)

We start by proving that P{Gn}1 as n, or equivalently, that P{Gnc}0 as n. Now,

P{Gnc}=P{QUUV:HQ(U,U)R+ϵ}[Gnc(QUUV)E(QUU)] (A94)
{QUUV:HQ(U,U)R+ϵ}PGnc(QUUV)E(QUU) (A95)
={QUUV:HQ(U,U)R+ϵ}[PE(QUU)+PGnc(QUUV)Ec(QUU)]. (A96)

The last summation contains a polynomial number of terms. If we prove that the summand tends to zero exponentially with n, then P{Gnc}0 as n. The first term in the summand, PE(QUU), has already been proved to be upper bounded by enϵ. Concerning the second term, we have the following

PGnc(QUUV)Ec(QUU)=P[(u,u)T(QUU)𝟙{B(u)=B(u)}·vT(QV|UU|u,u)𝟙Zuu(v)en[α(R2ϵ,QU,QV)+ϵ]<exp{n[HQ(U,U)R3ϵ/2]}·|T(QV|UU|u,u)|, (A97)
N(QUU)exp{n[HQ(U,U)Rϵ]}]=P[(u,u)T(QUU)𝟙{B(u)=B(u)}·vT(QV|UU|u,u)𝟙Zuu(v)>en[α(R2ϵ,QU,QV)+ϵ]>[N(QUU)exp{n[HQ(U,U)R3ϵ/2]}]·|T(QV|UU|u,u)|, (A98)
N(QUU)exp{n[HQ(U,U)Rϵ]}]P[(u,u)T(QUU)𝟙{B(u)=B(u)}·vT(QV|UU|u,u)𝟙Zuu(v)>en[α(R2ϵ,QU,QV)+ϵ]>[exp{n[HQ(U,U)Rϵ]}exp{n[HQ(U,U)R3ϵ/2]}]·|T(QV|UU|u,u)|, (A99)
N(QUU)exp{n[HQ(U,U)Rϵ]}]P[(u,u)T(QUU)𝟙{B(u)=B(u)}·vT(QV|UU|u,u)𝟙Zuu(v)>en[α(R2ϵ,QU,QV)+ϵ]>
[exp{n[HQ(U,U)Rϵ]}exp{n[HQ(U,U)R3ϵ/2]}]·|T(QV|UU|u,u)|] (A100)
E(u,u)T(QUU)𝟙{B(u)=B(u)}·vT(QV|UU|u,u)𝟙Zuu(v)>en[α(R2ϵ,QU,QV)+ϵ][exp{n[HQ(U,U)Rϵ]}exp{n[HQ(U,U)R3ϵ/2]}]·|T(QV|UU|u,u)| (A101)
·|T(QUU)|·|T(QV|UU|u,u)|·PB(u)=B(u),Zuu(v)>en[α(R2ϵ,QU,QV)+ϵ]exp{n[HQ(U,U)Rϵ]}·|T(QV|UU|u,u)| (A102)
exp{nHQ(U,U)}·PB(u)=B(u)·PZuu(v)>en[α(R2ϵ,QU,QV)+ϵ]exp{n[HQ(U,U)Rϵ]} (A103)
=enϵ·PZuu(v)>en[α(R2ϵ,QU,QV)+ϵ], (A104)

where (A99) follows by using the second event N(QUU)exp{n[HQ(U,U)Rϵ]} to increase the first event inside the probability in (A98), (A100) is true since the second event in (A99) was omitted, (A101) follows from Markov’s inequality, and (A103) is due to the independence between the two events inside the probability in (A102). As for the probability in (A104),

PZuu(v)>en[α(R2ϵ,QU,QV)+ϵ]
=PQU|VN(QUV)enf(QUV)>en[α(R2ϵ,QU,QV)+ϵ] (A105)
maxQU|VPN(QUV)>exp{n[α(R2ϵ,QU,QV)+ϵf(QUV)]} (A106)
enE, (A107)

where N(QUV) is the number of source sequences within B(u), other than u and u, that fall in the conditional type class T(QU|V|v), which is a binomial random variable with enHQ(U|V)2 trials and success rate of exponential order enR, and hence,

E=minQU|V[RHQ(U|V)]+f(QUV)+[HQ(U|V)R]+α(R2ϵ,QU,QV)+ϵf(QUV)+[HQ(U|V)R]+<α(R2ϵ,QU,QV)+ϵ (A108)
=min{QU|V:f(QUV)+[HQ(U|V)R]+α(R2ϵ,QU,QV)+ϵ}[RHQ(U|V)]+. (A109)

By definition of the function α(R,QU,QV), the set {QU|V:f(QUV)+[HQ(U|V)R]+α(R2ϵ,QU,QV)+ϵ} is a subset of {QU|V:HQ(U|V)R2ϵ}. Thus,

Emin{QU|V:HQ(U|V)R2ϵ}[RHQ(U|V)]+2ϵ, (A110)

and hence, PZuu(v)>en[α(R2ϵ,QU,QV)+ϵ]·e2nϵ, which provides

PGnc(QUUV)Ec(QUU)·enϵ·e2nϵ=enϵ, (A111)

which proves that P{Gn}1 as n. Now, for a given BnGn(QUUV), we define the set

K(Bn,QUUV)={(u,u,v):Zuu(v)exp{n[α(R2ϵ,QU,QV)+ϵ]}}, (A112)

and

K(Bn,QUUV|u,u)={v:(u,u,v)K(Bn,QUUV)}. (A113)

Then, by definition, for any BnGn(QUUV),

(u,u)T(QUU)𝟙{B(u)=B(u)}·|T(QV|UU|u,u)K(Bn,QUUV|u,u)||T(QV|UU|u,u)|exp{n[HQ(U,U)R3ϵ/2]}, (A114)

where we have used the fact that T(QV|UU|u,u) has exponentially the same cardinality for all (u,u)T(QUU). Wrapping all up, we get that for any BnGn,

Pe(Bn)
=u,vP(u,v)𝟙H^u(U)R(P^u)uB(u)T(u),uuexp{nf(P^uv)}exp{nf(P^uv)}+exp{nf(P^uv)}+Zuu(v){QUU:HQ(U,U)R+ϵ,HQ(U)R(QU)}(u,u)T(QUU)𝟙{B(u)=B(u)}·exp{nEQlogP(U)}×QV|UUvT(QV|UU|u,u)K(Bn,QUUV|u,u)exp{nEQlogP(V|U)} (A115)
×exp{nf(QUV)}exp{nf(QUV)}+exp{nf(QUV)}+Zuu(v){QUU:HQ(U,U)R+ϵ,HQ(U)R(QU)}(u,u)T(QUU)𝟙{B(u)=B(u)}·exp{nEQlogP(U)}×QV|UUvT(QV|UU|u,u)K(Bn,QUUV|u,u)exp{nEQlogP(V|U)} (A116)
×exp{nf(QUV)}exp{nf(QUV)}+exp{nf(QUV)}+exp{n[α(R2ϵ,QU,QV)+ϵ]}{QUU:HQ(U,U)R+ϵ,HQ(U)R(QU)}(u,u)T(QUU)𝟙{B(u)=B(u)}×QV|UU|T(QV|UU|u,u)K(Bn,QUUV|u,u)||T(QV|UU|u,u)|·|T(QV|UU|u,u)|·enEQlogP(U,V) (A117)
×exp{n·[max{f(QUV),α(R2ϵ,QU,QV)+ϵ}f(QUV)]+}{QUUV:HQ(U,U)R+ϵ,HQ(U)R(QU)}(u,u)T(QUU)𝟙{B(u)=B(u)}×|T(QV|UU|u,u)K(Bn,QUUV|u,u)||T(QV|UU|u,u)|·enHQ(V|U,U)·enEQlogP(U,V) (A118)
×exp{n·[max{f(QUV),α(R2ϵ,QU,QV)+ϵ}f(QUV)]+}{QUUV:HQ(U,U)R+ϵ,HQ(U)R(QU)}exp{n[HQ(U,U)R3ϵ/2]}·enHQ(V|U,U) (A119)
×enEQlogP(U,V)·exp{n·[max{f(QUV),α(R2ϵ,QU,QV)+ϵ}f(QUV)]+}expn·min{QUUV:HQ(U,U)R+ϵ,HQ(U)R(QU)}{HQ(U,U)+R+3ϵ/2HQ(V|U,U) (A120)
EQ[logP(U,V)]+[max{f(QUV),α(R2ϵ,QU,QV)+ϵ}f(QUV)]+} (A121)
=exp{nEtrc(R,ϵ)}, (A122)

where (A117) follows from the definition of the set K(Bn,QUUV|u,u) in (A113) and (A120) is due to (A114). Consider the following:

E1nlogPe(Bn)
=BnP{Bn}1nlogPe(Bn) (A123)
=BnGnP{Bn}1nlogPe(Bn)+BnGncP{Bn}1nlogPe(Bn) (A124)
BnGnP{Bn}1nlogenEtrc(R,ϵ)+BnGncP{Bn}1nlogenEsp(R) (A125)
=P{Gn}Etrc(R,ϵ)+P{Gnc}Esp(R), (A126)

which implies that

lim supnE1nlogPe(Bn)Etrc(R,ϵ). (A127)

It follows from the arbitrariness of ϵ that

lim supnE1nlogPe(Bn)min{QUUV:HQ(U,U)R,HQ(U)R(QU)}{HQ(U,U)+RHQ(V|U,U)
EQ[logP(U,V)]+[max{f(QUV),α(R,QU,QV)}f(QUV)]+}=min{QUUV:HQ(U)R(QU)}{HQ(U,U)+RHQ(V|U,U) (A128)
EQ[logP(U,V)]+[max{f(QUV),α(R,QU,QV)}f(QUV)]+} (A129)
=min{QUUQ:HQ(U)R(QU)}{Λ(QUU,R)HQ(U,U)+R(QU)EQ[logP(U)]}, (A130)

which completes the proof of Theorem 2.

Appendix C. Proof of Lemma A1

Let N(T(QU|V|v),B(u)) be defined as

N(T(QU|V|v),B(u))=uT(QU|V|v)𝟙B(u)=B(u). (A131)

First, note that

Zu(v)=u˜B(u)T(u),u˜uexp{nf(P^u˜v)}=QU|VS(P^u,P^v)N(T(QU|V|v),B(u))enf(QUV), (A132)

where S(P^u,P^v)={QU|V:(P^v×QU|V)U=P^u}. Thus, taking the randomness of {B(u)}uUn into account,

PZv(u)exp{nα(R+ϵ,P^u,P^v)}
=PQU|VS(P^u,P^v)N(T(QU|V|v),B(u))enf(QUV)exp{nα(R+ϵ,P^u,P^v)} (A133)
PmaxQU|VS(P^u,P^v)N(T(QU|V|v),B(u))enf(QUV)exp{nα(R+ϵ,P^u,P^v)} (A134)
=PQU|VS(P^u,P^v)N(T(QU|V|v),B(u))enf(QUV)exp{nα(R+ϵ,P^u,P^v)} (A135)
=PQU|VS(P^u,P^v)N(T(QU|V|v),B(u))exp{n[α(R+ϵ,P^u,P^v)f(QUV)]}. (A136)

Now, N(T(QU|V|v),B(u)) is a binomial random variable with |T(QU|V|v)|enHQ(U|V) trials and success rate which is of the exponential order of enR. We prove that by the very definition of the function α(R+ϵ,P^u,P^v), there must exist some conditional distribution QU|V*S(P^u,P^v) such that for QUV*=P^v×QU|V*, the two inequalities HQ*(U|V)R+ϵ and HQ*(U|V)Rϵα(R+ϵ,P^u,P^v)f(QUV*) hold. To show that, we assume conversely, i.e., that for every conditional distribution QU|VS(P^u,P^v), which defines QUV=P^v×QU|V, either HQ(U|V)<R+ϵ or HQ(U|V)Rϵ<α(R+ϵ,P^u,P^v)f(QUV), which means that for every distribution QU|VS(P^u,P^v)

HQ(U|V)ϵ<max{R,R+α(R+ϵ,P^u,P^v)f(QUV)} (A137)
=R+[α(R+ϵ,P^u,P^v)f(QUV)]+. (A138)

Writing it slightly differently, for every QU|VS(P^u,P^v) there exists some real number t[0,1] such that

HQ(U|V)ϵ<R+t[α(R+ϵ,P^u,P^v)f(QUV)], (A139)

or equivalently,

α(R+ϵ,P^u,P^v)>maxQU|VS(P^u,P^v)mint[0,1]f(QUV)+HQ(U|V)Rϵt (A140)
=maxQU|VS(P^u,P^v)f(QUV)+HQ(U|V)RϵHQ(U|V)R+ϵHQ(U|V)<R+ϵ (A141)
=max{QU|VS(P^u,P^v):HQ(U|V)R+ϵ}[f(QUV)+HQ(U|V)]Rϵ (A142)
α(R+ϵ,P^u,P^v), (A143)

which is a contradiction. Let the conditional distribution QU|V* be as defined above. Then,

PQU|VS(P^u,P^v)N(T(QU|V|v),B(u))exp{n[α(R+ϵ,P^u,P^v)f(QUV)]} (A144)
PN(T(QU|V*|v),B(u))exp{n[α(R+ϵ,P^u,P^v)f(QUV*)]}. (A145)

Now, we know that both of the inequalities HQ*(U|V)R+ϵ and HQ*(U|V)Rϵα(R+ϵ,P^u,P^v)f(QUV*) hold. By the Chernoff bound, the probability of (A145) is upper bounded by

expenHQ*(U|V)D(eanebn), (A146)

where a=HQ*(U|V)+f(QUV*)α(R+ϵ,P^u,P^v) and b=R, and where D(αβ), for α,β[0,1], is the binary divergence function, that is

D(αβ)=αlogαβ+(1α)log1α1β. (A147)

Since abϵ, the binary divergence is lower bounded as follows [24](Section 6.3):

D(eanebn)ebn1e(ab)n[1+n(ab)] (A148)
enR[1enϵ(1+nϵ)], (A149)

where in the second inequality, we invoked the decreasing monotonicity of the function f(t)=(1+t)et for t0. Finally, we get that

PN(T(QU|V*|v),B(u))exp{n[α(R+ϵ,P^u,P^v)f(QUV*)]} (A150)
expenHQ*(U|V)·enR[1enϵ(1+nϵ)] (A151)
expenϵ[1enϵ(1+nϵ)] (A152)
=expenϵ+nϵ+1. (A153)

This completes the proof of Lemma A1.

Appendix D. Proof of Theorem 3

By definition of the error exponents, it follows that Etrc,GLD(R(·))Er,GLD(R(·)). We now prove the other direction. The expression in (28) can also be written as

Etrc,GLD(R(·))
=minQUU:QU=QU,HQ(U)R(QU)Λ(QUU,R(QU))EQ[logP(U)]HQ(U,U)+R(QU)=minQUU:QU=QU,HQ(U)R(QU)minQV|UUΨ(R(QU),QUUV)HQ(V|U,U)EQ[logP(V|U)] (A154)
EQ[logP(U)]HQ(U,U)+R(QU) (A155)
=minQUUV:QU=QU,HQ(U)R(QU)Ψ(R(QU),QUUV)HQ(U,U,V)EQ[logP(U,V)]+R(QU) (A156)
=minQUUV:QU=QU,HQ(U)R(QU)Ψ(R(QU),QUUV)+D(QUVPUV)HQ(U|U,V)+R(QU)=minQD(QUVPUV)+R(QU)HQ(U|U,V) (A157)
+max{f(QUV),γ(R(QU),QU,QV)}f(QUV)+, (A158)

with the set Q given by Q={QUUV:QU=QU,HQ(U)R(QU)}, and where,

γ(R(·),QU,QV)=maxQU˜|V:QU˜=QU,HQ(U˜|V)R(QU˜){f(QU˜V)+HQ(U˜|V)}R(QU). (A159)

We upper–bound the minimum in (A158) by decreasing the feasible set; we add to Q the constraint that UVU form a Markov chain in that order and denote the new feasible set by Q˜. We get that

Etrc,GLD(R(·))minQ˜D(QUVPUV)+R(QU)HQ(U|U,V)
+max{f(QUV),γ(R(QU),QU,QV)}f(QUV)+=minQ˜D(QUVPUV)+R(QU)HQ(U|V) (A160)
+max{f(QUV),γ(R(QU),QU,QV)}f(QUV)+=min{QUV:HQ(U)R(QU)}D(QUVPUV)+minQU|VQ^{R(QU)HQ(U|V) (A161)
+max{f(QUV),γ(R(QU),QU,QV)}f(QUV)+}, (A162)

where Q^={QU|V:QU=QU}. In order to upper–bound the inner minimum in (A162), we split into two cases, according to the maximum between f(QUV) and γ(R(QU),QU,QV). This is legitimate when the inner minimum and this maximum can be interchanged, which is possible at least in the special cases of the matched/mismatched decoding metrics f(Q)=βEQ[logP˜(U,V)] for some β>0, since if f(Q) is linear, then the entire expression inside the inner minimum in (A162) is convex in QU|V. On the one hand, if the maximum is given by f(QUV), then the inner minimum in (A162) is just

minQU|VQ^R(QU)HQ(U|V)+f(QUV)f(QUV)+. (A163)

On the other hand, if the maximum is given by γ(R(QU),QU,QV), let Q*=QU˜|V* be the maximizer in (A159), and then

minQU|VQ^R(QU)HQ(U|V)+γ(R(QU),QU,QV)f(QUV)+
=minQU|VQ^R(QU)HQ(U|V)+f(QU˜V*)+HQ*(U˜|V)R(QU)f(QUV)+ (A164)
R(QU)HQ*(U|V)+f(QU˜V*)+HQ*(U˜|V)R(QU)f(QUV*)+ (A165)
=R(QU)HQ*(U|V)+HQ*(U˜|V)R(QU)+ (A166)
=R(QU)HQ*(U|V)+HQ*(U˜|V)R(QU) (A167)
=0, (A168)

where (A165) is because we choose QU|V*=QU˜|V* instead of minimizing over all QU|VQ^ and (A167) is true since HQ*(U˜|V)R(QU) by the definition of γ(R(QU),QU,QV). Combining (A163) and (A168), we find that (A162) is upper–bounded by

Etrc,GLD(R(·))minQUV:HQ(U)R(QU)D(QUVPUV)
+maxminQU|VQ^R(QU)HQ(U|V)+f(QUV)f(QUV)+,0=minQUV:HQ(U)R(QU)D(QUVPUV) (A169)
+minQU|VQ^R(QU)HQ(U|V)+f(QUV)f(QUV)++=minQUV:HQ(U)R(QU)D(QUVPUV) (A170)
+minQU|VQ^R(QU)HQ(U|V)+f(QUV)f(QUV)++ (A171)
=Er,GLD(R(·)), (A172)

which proves the first point of the theorem. Moving forward, consider the following:

Etrc,MAP(R(·))=(a)Er,MAP(R(·))=(b)Er,MCE(R(·))(c)Etrc,MCE(R(·))(d)Etrc,MAP(R(·)), (A173)

where (a) follows from the first point in this theorem by using the matched decoding metric f(Q)=βEQ[logP(U,V)] and letting β. Equality (b) is due to the second point of Theorem 1, which ensures that the random binning error exponents of the MAP and the MCE decoders are equal. Passage (c) is thanks to the fact that for any decoder, the error exponent of the typical random code is always at least as high as the random coding error exponent and (d) is due to the fact that the MAP decoder is optimal. Finally, the leftmost and the rightmost sides of (A173) are the same, which implies that passages (c) and (d) must hold with equalities. The equality in passage (c) concludes the second point of the theorem.

Appendix E. Proof of Theorem 4

The left equality in (33) is implied by the proved equality in passage (d) in (A173). In order to prove the right equality in (33), first note that Etrc,SCE(R(·))Etrc,MAP(R(·)) by the optimality of the MAP decoder. For the other direction, consider the universal decoding metric of f(QUV)=HQ(U|V). Then, trivially,

γ(R(·),QU,QV)=maxQU˜|V:QU˜=QU,HQ(U˜|V)R(QU˜){f(QU˜V)+HQ(U˜|V)}R(QU)=R(QU), (A174)

and

Ψ(R(·),QUUV)=max{f(QUV),γ(R(·),QU,QV)}f(QUV)+ (A175)
=max{HQ(U|V),R(QU)}+HQ(U|V)+ (A176)
=HQ(U|V)min{HQ(U|V),R(QU)}+ (A177)
HQ(U|U,V)min{HQ(U|V),R(QU)}+. (A178)

We have the following

Etrc,SCE(R(·))=minQD(QUVPUV)+R(QU)HQ(U|U,V)
+max{f(QUV),γ(R(QU),QU,QV)}f(QUV)+minQD(QUVPUV)+R(QU)HQ(U|U,V) (A179)
+HQ(U|U,V)min{HQ(U|V),R(QU)}+ (A180)
=minQD(QUVPUV)min{HQ(U|V),HQ(U|U,V),R(QU)}+R(QU) (A181)
minQD(QUVPUV)min{HQ(U|V),HQ(U),R(QU)}+R(QU) (A182)
=min{QUV:HQ(U)R(QU)}D(QUVPUV)min{HQ(U|V),HQ(U),R(QU)}+R(QU) (A183)
=min{QUV:HQ(U)R(QU)}D(QUVPUV)min{HQ(U|V),R(QU)}+R(QU) (A184)
=min{QUV:HQ(U)R(QU)}D(QUVPUV)+max{R(QU)HQ(U|V),0} (A185)
=min{QUV:HQ(U)R(QU)}D(QUVPUV)+[R(QU)HQ(U|V)]+ (A186)
=Etrc,MAP(R(·)), (A187)

which completes the proof of the theorem.

Appendix F. Proof of Theorem 5

We start by writing the expression in (34) in a slightly different way using

min{Q:g(Q)0}f(Q)=minQsups0{f(Q)+s·g(Q)}:

Eer(R(·),Δ)=min{QUV:R(QU)HQ(U|V)+Δ}D(QUVPUV) (A188)
=minQUVsupσ0{D(QUVPUV)+σ·(HQ(U|V)+ΔR(QU))}. (A189)

Now, the requirement Eer(R(·),Δ)Er is equivalent to

minQUVsupσ0{D(QUVPUV)+σ·(HQ(U|V)+ΔR(QU))}Er (A190)

or,

QUV,σ0,D(QUVPUV)+σ·(HQ(U|V)+ΔR(QU))Er (A191)

or,

QU,QV|U,σ0,R(QU)HQ(U|V)+Δ+D(QUVPUV)Erσ (A192)

or that for any QUP(U),

R(QU)minQV|Usupσ0HQ(U|V)+Δ+D(QUVPUV)Erσ (A193)
=minQV|UHQ(U|V)+ΔD(QUVPUV)ErD(QUVPUV)>Er (A194)
=min{QV|U:D(QUVPUV)Er}HQ(U|V)+Δ, (A195)

with the understanding that a minimum over an empty set equals infinity.

Appendix G. Proof of Theorem 6

It follows by the identities min{Q:g(Q)0}f(Q)=minQsups0{f(Q)+s·g(Q)} and A+=maxμ[0,1]μA that (24) can also be written as

Ee(R(·))=minQUminQV|Umaxμ[0,1]supσ0{D(QUVPUV)+μ·(R(QU)HQ(U|V))+σ·(R(QU)HQ(U))}, (A196)

such that Ee(R(·))Ee is equivalent to

QU,QV|U,μ[0,1],σ0:D(QUVPUV)+μ·(R(QU)HQ(U|V))+σ·(R(QU)HQ(U))Ee, (A197)

or,

QU,QV|U,μ[0,1],σ0:R(QU)μ·HQ(U|V)+σ·HQ(U)+EeD(QUVPUV)μ+σ, (A198)

or that for any QUP(U),

R(QU)maxQV|Uminμ[0,1]infσ0μ·HQ(U|V)+σ·HQ(U)+EeD(QUVPUV)μ+σ (A199)
=maxQV|Uminμ[0,1]minHQ(U),HQ(U|V)+EeD(QUVPUV)μ (A200)
=maxQV|UminHQ(U),minμ[0,1]HQ(U|V)+EeD(QUVPUV)μ (A201)
=maxQV|Umin{HQ(U),HQ(U|V)+EeD(QUVPUV)}EeD(QUVPUV)Ee<D(QUVPUV) (A202)
=max{QV|U:D(QUVPUV)Ee}min{HQ(U),HQ(U|V)+EeD(QUVPUV)} (A203)
=minHQ(U),max{QV|U:D(QUVPUV)Ee}{HQ(U|V)+EeD(QUVPUV)}, (A204)

and the proof is complete.

Author Contributions

Both authors contributed equally to this research. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Israel Science Foundation (ISF) grant number 137/18.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Merhav N. Error exponents of typical random codes. IEEE Trans. Inf. Theory. 2018;64:6223–6235. doi: 10.1109/TIT.2018.2834503. [DOI] [Google Scholar]
  • 2.Barg A., Forney G.D., Jr. Random codes: Minimum distances and error exponents. IEEE Trans. Inf. Theory. 2003;48:2568–2573. doi: 10.1109/TIT.2002.800480. [DOI] [Google Scholar]
  • 3.Nazari A., Anastasopoulos A., Pradhan S.S. Error exponent for multiple–access channels: Lower bounds. IEEE Trans. Inf. Theory. 2014;60:5095–5115. doi: 10.1109/TIT.2014.2331952. [DOI] [Google Scholar]
  • 4.Merhav N. Error exponents of typical random codes for the colored Gaussian channel. IEEE Trans. Inf. Theory. 2019;65:8164–8179. doi: 10.1109/TIT.2019.2917175. [DOI] [Google Scholar]
  • 5.Merhav N. Error exponents of typical random trellis codes. IEEE Trans. Inf. Theory. 2020;66:2067–2077. doi: 10.1109/TIT.2019.2942624. [DOI] [Google Scholar]
  • 6.Merhav N. A Lagrange–dual lower bound to the error exponent of the typical random code. IEEE Trans. Inf. Theory. 2020;66:3456–3464. doi: 10.1109/TIT.2019.2963392. [DOI] [Google Scholar]
  • 7.Tamir (Averbuch) R., Merhav N., Weinberger N., Guillén i Fàbregas A. Large deviations behavior of the logarithmic error probability of random codes. IEEE Trans. Inf. Theory. 2020;66:6635–6659. doi: 10.1109/TIT.2020.2995136. [DOI] [Google Scholar]
  • 8.Slepian D., Wolf J. Noiseless coding of correlated information sources. IEEE Trans. Inf. Theory. 1973;19:471–480. doi: 10.1109/TIT.1973.1055037. [DOI] [Google Scholar]
  • 9.Gallager R.G. Source coding with side information and universal coding. LIDS-P-937. [(accessed on 20 February 2021)]; Available online: http://web.mit.edu/gallager/www/papers/paper5.pdf.
  • 10.Ahlswede R., Dueck G. Good codes can be produced by a few permutations. IEEE Trans. Inf. Theory. 1982;28:430–443. doi: 10.1109/TIT.1982.1056514. [DOI] [Google Scholar]
  • 11.Csiszár I., Körner J. Graph decomposition: A new key to coding theorems. IEEE Trans. Inf. Theory. 1981;27:5–12. doi: 10.1109/TIT.1981.1056281. [DOI] [Google Scholar]
  • 12.Chen J., He D.-K., Jagmohan A., Lastras-Montaño L.A. On the Reliability Function of Variable-Rate Slepian-Wolf Coding. Entropy. 2017;19:389. doi: 10.3390/e19080389. [DOI] [Google Scholar]
  • 13.Weinberger N., Merhav N. Optimum Tradeoffs Between the Error Exponent and the Excess-Rate Exponent of Variable-Rate Slepian–Wolf Coding. IEEE Trans. Inf. Theory. 2015;61:2165–2190. doi: 10.1109/TIT.2015.2405537. [DOI] [Google Scholar]
  • 14.Csiszár I. Linear codes for sources and source networks: Error exponents, universal coding. IEEE Trans. Inf. Theory. 1982;28:585–592. doi: 10.1109/TIT.1982.1056524. [DOI] [Google Scholar]
  • 15.Csiszár I., Körner J. Towards a general theory of source networks. IEEE Trans. Inf. Theory. 1980;26:155–165. doi: 10.1109/TIT.1980.1056166. [DOI] [Google Scholar]
  • 16.Kelly B.G., Wagner A.B. Improved Source Coding Exponents via Witsenhausen’s Rate. IEEE Trans. Inf. Theory. 2011;57:5615–5633. doi: 10.1109/TIT.2011.2162178. [DOI] [Google Scholar]
  • 17.Kelly B.G., Wagner A.B. Reliability in Source Coding With Side Information. IEEE Trans. Inf. Theory. 2012;58:5086–5111. doi: 10.1109/TIT.2012.2201346. [DOI] [Google Scholar]
  • 18.Oohama Y., Han T. Universal coding for the Slepian-Wolf data compression system and the strong converse theorem. IEEE Trans. Inf. Theory. 1994;40:1908–1919. doi: 10.1109/18.340465. [DOI] [Google Scholar]
  • 19.Liu J., Cuff P., Verdú S. On α–decodability and α–likelihood decoder; Proceedings of the 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton); Monticello, IL, USA. 3–6 October 2017. [Google Scholar]
  • 20.Merhav N. The generalized stochastic likelihood decoder: Random coding and expurgated bounds. IEEE Trans. Inf. Theory. 2017;63:5039–5051. doi: 10.1109/TIT.2017.2689787. See also a correction at IEEE Trans. Inf. Theory2017, 63, 6827–6829. [DOI] [Google Scholar]
  • 21.Csiszár I., Körner J. Information Theory: Coding Theorems for Discrete Memoryless Systems. Cambridge University Press; Cambridge, UK: 2011. [Google Scholar]
  • 22.Tamir (Averbuch) R., Merhav N. The MMI decoder is asymptotically optimal for the typical random code and for the expurgated code. arXiv. 20202007.12225 [Google Scholar]
  • 23.Somekh-Baruch A., Scarlett J., Guillén i Fàbregas A. Generalized Random Gilbert-Varshamov Codes. IEEE Trans. Inf. Theory. 2019;65:3452–3469. doi: 10.1109/TIT.2019.2902387. [DOI] [Google Scholar]
  • 24.Merhav N. Statistical physics and information theory. Found. Trends Commun. Inf. Theory. 2009;6:1–212. [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES