Skip to main content
Entropy logoLink to Entropy
. 2022 Nov 19;24(11):1695. doi: 10.3390/e24111695

Are Guessing, Source Coding and Tasks Partitioning Birds of A Feather?

M Ashok Kumar 1,*, Albert Sunny 2, Ashish Thakre 3, Ashisha Kumar 3, G Dinesh Manohar 4
Editors: Nicusor Minculete, Shigeru Furuichi
PMCID: PMC9689969  PMID: 36421550

Abstract

This paper establishes a close relationship among the four information theoretic problems, namely Campbell source coding, Arikan guessing, Huleihel et al. memoryless guessing and Bunte and Lapidoth tasks’ partitioning problems in the IID-lossless case. We first show that the aforementioned problems are mathematically related via a general moment minimization problem whose optimum solution is given in terms of Renyi entropy. We then propose a general framework for the mismatched version of these problems and establish all the asymptotic results using this framework. The unified framework further enables us to study a variant of Bunte–Lapidoth’s tasks partitioning problem which is practically more appealing. In addition, this variant turns out to be a generalization of Arıkan’s guessing problem. Finally, with the help of this general framework, we establish an equivalence among all these problems, in the sense that, knowing an asymptotically optimal solution in one problem helps us find the same in all other problems.

Keywords: guessing, source coding, tasks partitioning, Shannon entropy, Rényi entropy, Kullback–Leibler divergence, relative α-entropy, Sundaresan’s divergence

MSC: 94A15, 94A17, 94A50, 62B10

1. Introduction

The concept of entropy is very central to information theory. In source coding, the expected number of bits required (per letter) to encode a source with (finite) alphabet set X and probability distribution P is the Shannon entropy H(P):=xXP(x)logP(x). If the compressor does not know the true distribution P, but assumes a distribution Q (mismatch), then the number of bits required for compression is H(P)+I(P,Q), where

I(P,Q):=xXP(x)logQ(x)+xXP(x)logP(x), (1)

is the entropy of P relative to Q (or the Kullback–Leibler divergence). In his seminal paper, Shannon [1] argued that H(P) can also be regarded as a measure of uncertainty. Subsequently, Rényi [2] introduced an alternate measure of uncertainty, now known as Rényi entropy of order α, as

Hα(P):=11αlogxXP(x)α, (2)

where α>0 and α1. Rényi entropy can also be regarded as a generalization of the Shannon entropy as limα1Hα(P)=H(P). Refer Aczel and Daroczy [3] and the references therein for an extensive study of various measures of uncertainty and their characterizations.

In 1965, Campbell [4] gave an operational meaning to Rényi entropy. He showed that, instead of expected code lengths, if one minimizes the cumulants of code lengths, then the optimal cumulant is Rényi entropy Hα(P), where α=1/(1+ρ) with ρ being the order of the cumulant. He also showed that the optimal cumulant can be achieved by encoding sufficiently long sequences of symbols. Sundaresan (Theorem 8 of [5]) (c.f. Blumer and McElice [6]) showed that, in the mismatched case, the optimal cumulant is Hα(P)+Iα(P,Q), where

Iα(P,Q):=α1αlogxP(x)Q(x)αyQ(y)αα1αHα(P) (3)

is called α-entropy of P relative to Q or Sundaresan’s divergence [7]. Hence, Iα(P,Q) can be interpreted as the penalty for not knowing the true distribution. The first term in (3) is sometimes called the Renyi cross-entropy and is analogous to the first term of (1). Iα(P,Q)0 with equality if and only if P=Q. Iα-divergence can also be regarded as a generalization of the Kullback–Leibler divergence as limα1Iα(P,Q)=I(P,Q). Refer to [5,8,9] for detailed discussions on the properties of Iα. Lutwak et al. also independently identified Iα in the context of maximum Rényi entropy and called it an α-Rényi relative entropy (Equation (4) of [10]). Iα, for α>1, also arises in robust inference problems (see [11] and the references therein).

In [12], Massey studied a guessing problem where one is interested in the expected number of guesses required to guess a random variable X that assumes values from an infinite set, and found a lower bound in terms of Shannon entropy. Arıkan [13] studied it for a finite alphabet set and showed that Rényi entropy arises as the optimal solution in minimizing moments of the number of guesses. Subsequently, Sundaresan [5] showed that the penalty in guessing according to a distribution Q when the true distribution is P, is given by Iα(P,Q). It is interesting to note that guesswork has also been studied from a large deviations point of view [14,15,16,17,18]. Bunte and Lapidoth [8] studied a problem on partitioning of tasks and showed that Rényi entropy and Sundaresan’s divergence play a similar role in the optimal number of tasks performed. We propose, in this paper, a variant of this problem where the tasks in each subset of the partition are performed according to the decreasing order of probabilities. We show that Rényi antropy and Sundaresan’s divergences arise as optimal solutions in this problem too. Huleihel et al. [19,20] studied the memoryless guessing problem, a variant of Arıkan’s guessing problem, with i.i.d. (independent and identically distributed) guesses and showed that the minimum attainable factorial moments of number of guesses is the Rényi entropy. We show, in this paper, that the minimum factorial moment in the mismatched case is measured by the Sundaresan’s divergence.

We observe that, in all these problems, the objective is to minimize usual moments or factorial moments of random variables, and Rényi entropy and Sundaresan’s divergence arise in optimal solutions. The relationship between source coding and guessing is well-known in the literature. Arıkan and Merhav established a close relationship between lossy source coding and guessing with distortion using large deviation techniques [14,21]. The same for the lossless case was done by Hanawal and Sundaresan [17]. In this paper, we establish a general framework for all the five problems in the IID-lossless case. We then use this to establish upper and lower bounds for the mismatched version of these problems. This helps us find an equivalence among all these problems, in the sense that knowing an asymptotically optimal solution in one problem helps us find the same in all other problems.

Our Contributions in the Paper:

  • (a)

    a general framework for the problems on source coding, guessing and tasks partitioning;

  • (b)

    lower and upper bounds for the general framework of these problems both in matched and mis-matched cases;

  • (c)

    a unified approach to derive bounds for the mismatched version of these problems;

  • (d)

    a generalized tasks partitioning problem; and

  • (e)

    establishing operational commonality among the problems.

Organisation of the Paper:

In Section 2, we present our unified framework, and find conditions under which lower and upper bounds are attained. In Section 3, we present four well-known information-theoretic problems, namely, Campbell’s source coding, Arıkan’s guessing, Huleihel et al.’s memoryless guessing, and Bunte–Lapidoth’s tasks partitioning, and re-establish and refine major results pertaining to these problems. In Section 4, we propose and solve a generalized tasks partitioning problem. In Section 5, we establish a connection among the aforementioned problems. Finally, we summarize and conclude the paper in Section 6.

2. A General Minimization Problem

In this section, we present a general minimization problem whose optimum solution evaluates to Rényi entropy. We will later show that all problems stated in Section 3 are particular instances of this general problem.

 Proposition 1. 

Let ψ:X(0,) be such that xXψ(x)1k for some k>0. Then, for ρ(1,0)(0,),

1ρlogEP[ψ(X)ρ]Hα(P)logk, (4)

where EP[·] denotes the expectation with respect to probability distribution P on X, Hα(P) is the Rényi entropy of order α, and α:=α(ρ)=1/(1+ρ). The lower bound is achieved if and only if

ψ(x)1=k·P(x)α/ZP,αforxX, (5)

where ZP,α:=xXP(x)α.

 Proof. 

Observe that

sgn(ρ)xXP(x)ψ(x)ρ=sgn(ρ)xXP(x)αψ(x)1P(x)αρ(a)sgn(ρ)xP(x)α·xψ(x)1xP(x)αρ=sgn(ρ)xP(x)α1+ρ·xψ(x)1ρ(b)sgn(ρ)xP(x)α1+ρ·kρ,

where (a) is due to the generalised log-sum inequality (Equation (4.1) of [22]) applied to the function f(x)=sgn(ρ)·xρ; and (b) follows from the hypothesis that xψ(x)1k. By taking log and then dividing by ρ, we obtain (4). Equality holds in (a) if and only if ψ(x)1=νP(x)α for some constant ν and in (b) if and only if xψ(x)1=k. This completes the proof. □

The left-side of (4) is called normalised cumulant of ψ(X) of order ρ. The measure P(α)(x):=P(x)α/ZP,α in (5) that attains the lower bound in (4) is called an α-scaled measure or escort measure of P. This measure also arises in robust inference (Equation (7) of [11]) and statistical physics [23]. The above proposition can also be proved using a variational formula as follows. By a version of Donsker–Varadhan variational formula (Propostion 4.5.1 of [24]), for any real-valued f on X, we have

logEP[2f(X)]=maxQ{EQ[f(X)]D(QP)}, (6)

where the max is over all probability distributions Q on X. Taking ρ>0 and f(x)=ρlogψ(x) in (6), we have

logEP[ψ(X)ρ]=maxQ{ρEQ[logψ(X)]D(QP)}=maxQρxQ(x)logψ(x)1Q(x)Q(x)D(QP)=maxQρH(Q)ρxQ(x)logψ(x)1Q(x)D(QP)(a)maxQρH(Q)ρlogxψ(x)1D(QP)(b)ρmaxQH(Q)1ρD(QP)ρlogk,

where (a) is by the log-sum inequality (Equation (4.1) of [22]) and (b) is by applying the constraint xψ(x)1k. For ρ(1,0), the inequalities in (a) and (b) are reversed, and the last max is replaced by min. Hence, (4) follows as the last max is equal to Hα(P) by (Theorem 1 of [25]). Equality in (a) and (b) holds if and only if ψ(x)1=k·Q(x). In addition, the last max is attained when Q(x)=P(x)α/ZP,αforxX. This completes the proof. The following is the analogous one for Shannon entropy.

 Proposition 2. 

Let ψ:X(0,) be such that xXψ(x)1k. Then,

EPlogψ(X)H(P)logk. (7)

Equality in (7) is achieved if and only if ψ(x)1=k·P(x)xX.

Proof. 

EPlogψ(X)=xP(x)logP(x)·ψ(x)1P(x)=H(P)xP(x)logψ(x)1P(x)H(P)logxψ(x)1H(P)logk,

where the penultimate inequality is due to the log-sum inequality. Equality holds in both inequalities if and only if ψ(x)1=k·P(x)xX. □

It is interesting to note that logEP[ψ(X)ρ]/ρEP[log(ψ(X))] and Hα(P)H(P) as ρ0 in (4). We now extend Propositions 1 and 2 to sequences of random variables. Let Xn be the set of all n-length sequences of elements of X, and Pn be the n-fold product distribution of P on Xn, that is, for xn:=x1,,xnXn, Pn(xn)=i=1nP(xi).

 Corollary 1. 

Given any n1, if ψn:Xn[0,) is such that xnXnψn(xn)1 kn for some kn>0, then

  • (a)
    For ρ(1,0)(0,),
    lim infn1nρlogEPn[ψn(Xn)ρ]Hα(P)lim supnlogknn.
  • (b)
    lim infn1nEPn[logψn(Xn)]H(P)lim supnlogknn,

where EPn[·] denotes the expectation with respect to probability distribution Pn on Xn.

 Proof. 

It is easy to see that Hα(Pn)=nHα(P) and H(Pn)=nH(P). Applying Propositions 1 and 2, dividing throughout by n and taking lim infn, the results follow. □

A General Framework for Mismatched Cases

In this sub-section, we establish a unified approach for cases when there is mismatch between assumed and true distributions.

 Proposition 3. 

Let ρ>1, α=1/(1+ρ), and Q be a probability distribution on X. For n1, let Qn be the n-fold product distribution of Q on Xn. If ψn:Xn(0,) is such that

ψn(xn)cn·ZQn,αQn(xn)α (8)

for some cn>0, then

  • (a)
    for ρ0, we have
    sgn(ρ)·EPn[ψn(Xn)ρ]sgn(ρ)·2nρ[Hα(P)+Iα(P,Q)+n1logcn],
  • (b)
    for ρ0, we have
    lim supn1nρlogEPn[ψn(Xn)ρ]Hα(P)+Iα(P,Q)+lim supn1nlogcn,
  • (c)
    for ρ=0, we have
    EPn[logψn(Xn)]n[H(P)+I(P,Q)+n1logcn],
  • (d)
    for ρ=0, we have
    lim supn1nEPn[logψn(Xn)]H(P)+I(P,Q)+lim supn1nlogcn.

 Proof. 

Part (a): From (8), we have

sgn(ρ)·EPn[ψn(Xn)ρ]=sgn(ρ)·xnXnPn(xn)ψn(xn)ρsgn(ρ)·cnρZQn,αρxnXnPn(xn)Qn(xn)αρ=sgn(ρ)·2ρ[Hα(Pn)+Iα(Pn,Qn)+logcn]=sgn(ρ)·2nρ[Hα(P)+Iα(P,Q)+n1logcn], (9)

where the penultimate equality holds from the definition of Iα, and the last one holds because Hα(Pn)=nHα(P), Iα(Pn,Qn)=nIα(P,Q).

Part (b): Taking log, dividing throughout by nρ, and then applying lim sup successively on both sides of (9), the result follows.

Part (c): When ρ=0, we have α=1 and (8) becomes ψn(xn)cnQn(xn). Hence,

EPn[logψn(Xn)]=xnXnPn(xn)logψn(xn)logcn+xnXnPn(xn)log(1/Qn(xn))=logcn+H(Pn)+I(Pn,Qn)=n(H(P)+I(P,Q)+n1logcn), (10)

where the last equality holds because H(Pn)=nH(P), and I(Pn,Qn)=nI(P,Q).

Part (d): Dividing (10) throughout by n, and taking limsup on both sides, the result follows. □

 Proposition 4. 

Let ρ>1, α=1/(1+ρ), and Q be a probability distribution on X. For n1, let Qn be the n-fold product distribution of Q on Xn. Suppose ψn:Xn(0,) is such that

ψn(xn)anZQn,αQn(xn)α

for some an>0, then

  • (a)
    for ρ0, we have
    sgn(ρ)·EPn[ψn(Xn)ρ]sgn(ρ)·2nρ(Hα(P)+Iα(P,Q)+n1log an),
  • (b)
    for ρ0, we have
    lim infn1nρlogEPn[ψn(Xn)ρ]Hα(P)+Iα(P,Q)+lim infn1nlogan,
  • (c)
    for ρ=0, we have
    EPn[logψn(Xn)]n(H(P)+I(P,Q)+n1logan),
  • (d)
    for ρ=0, we have
    lim infn1nEPn[logψn(Xn)]H(P)+I(P,Q)+lim infn1nlogan.

 Proof. 

Similar to proof of Proposition 3. □

3. Problem Statements and Known Results

In this section, we discuss Campbell’s source coding problem, Arıkan’s guessing problem, Huleihel et al.’s memoryless guessing problem, and Bunte–Lapidoth’s tasks partitioning problem. Using the general framework presented in the previous section, we re-establish known results, and present a few new results relating to these problems.

3.1. Source Coding Problem

Let X be a random variable that assumes values from a finite alphabet set X={a1,,am} according to a probability distribution P. The tuple (X,P) is usually referred to as a source. A binary codeC is a mapping from X to the set of finite length binary strings. Let L(C(X)) be the length of code C(X). The objective is to find a uniquely decodable code that minimizes the expected code-length, that is,

MinimizeEP[L(C(X))]

over all uniquely decodable codes C. Kraft and McMillan independently proved the following relation between uniquely decodable codes and their code-lengths.

Kraft-McMilan Theorem [26]: If C a uniquely decodable code, then

xX2L(C(x))1. (11)

Conversely, given a length sequence that satisfies the above inequality, there exists a uniquely decodable code C with the given length sequence.

Thus, one can confine the search space for C to codes satisfying the Kraft–McMillan inequality (11).

Theorem 5.3.1 of [26]: If C is a uniquely decodable code, then EP[L(C(X))]H(P).

 Proof. 

Choose ψ(x)=2L(C(x)), where L(C(x)) is the length of code C(x) assigned to alphabet x. Since C is uniquely decodable, from (11), we have xXψ(x)11. Now, an application of Proposition 2 with k=1 yields the desired result. □

 Theorem 1. 

Let Xn:=X1,,Xn be an i.i.d. sequence from Xn following the product distribution Pn(Xn)=i=1nP(Xi). Let Qn(Xn)=i=1nQ(Xi), where Q is another probability distribution. Let Cn be a code such that L(Cn(Xn))=logQn(Xn). Then, Cn satisfies Kraft–McMillan inequality and

limnEPn[L(Cn(Xn))]n=H(P)+I(P,Q).

 Proof. 

Choose ψn(xn)=2L(Cn(xn)), where L(Cn(xn) is the length of code Cn(xn) assigned to sequence xn. Then, we have

ψn(xn)=2L(Cn(xn))=2logQn(xn)2·2logQn(xn)=2/Qn(xn).

An application of Proposition 3 with cn=2 yields lim supnEPn[L(Cn(Xn))]/nH(P)+I(P,Q). Furthermore, we also have

ψn(xn)=2L(Cn(xn))=2logQn(xn)2logQn(xn)=1/Qn(xn).

An application of Proposition 4 with an=1 gives lim infnEPn[L(Cn(Xn))]/nH(P)+I(P,Q). □

3.2. Campbell Coding Problem

Campbell’s coding problem is similar to Shannon’s source coding problem except that, instead of minimizing the expected code-length, one is interested in minimizing the normalized cumulant of code lengths, that is,

Minimize1ρlogEP[2ρL(C(X))],

over all uniquely decodable codes C, and ρ>0. This problem was shown to be equivalent to minimizing buffer overflow probability by Humblet in [27]. A lower bound for the normalized cumulants in terms of Rényi entropy was provided by Campbell [4].

Lemma 1 of [4]: Let C be a uniquely decodable code. Then,

1ρlogEPn[2ρL(C(X))]Hα(P), (12)

where α=1/(1+ρ).

 Proof. 

Apply Proposition 1 with ψ(x)=2L(C(x)) and k=1. □

Notice that, if we ignore the integer constraint of the length function, then

L(C(x))=logZP,αP(x)α, (13)

with ZP,α as in Proposition 1, satisfies (11) and achieves the lower bound in (12). Campbell also showed that the lower bound in (12) can be achieved by encoding long sequences of symbols with code-lengths close to (13).

Theorem 1 of [4]: If Cn is a uniquely decodable code such that

L(Cn(xn))=logZPn,αPn(xn)α, (14)

then

limn1nρlogEPn[2ρL(Cn(Xn))]=Hα(P).

 Proof. 

Choose ψn(xn)=2L(Cn(xn)). Then, from (14), we have

ZPn,αPn(xn)αψn(xn)<2·ZPn,αPn(xn)α·

The result follows by applying Propositions 3 and 4 with cn=2, an=1 and Q=P. □

Mismatch Case:

Redundancy in the mismatched case of the Campbell’s problem was studied in [5,6]. Sundaresan showed that the difference in the normalized cumulant from the minimum when encoding according to an arbitrary uniquely decodable code is measured by Iα-divergence up-to a factor of 1 [5]. We provide a more general version of this result in the following.

 Proposition 5. 

Let X be a random variable that assumes values from set X according to a probability distribution P. Let ρ(1,0)(0,) and L:XZ+ be an arbitrary length function that satisfies (11). Define

Rc(P,L,ρ):=1ρlogEP[2ρL(X)]minK1ρlogEP[2ρK(X)], (15)

where the minimum is over all length functions K satisfying (11). Then, there exists a probability distribution QL such that

Iα(P,QL)logη1Rc(P,L,ρ)Iα(P,QL)logη, (16)

where η=x2L(x).

 Proof. 

Since K satisfies (11), an application of Proposition 1 with ψ(x)=2K(x) gives us 1ρlogEP[2ρK(X)]Hα(P). Since K(x)=log(ZP,α/P(x)α) satisfies (11) and ψ(x)=2K(x)<2·ZP,α/P(x)α, applying Proposition 3 with n=1,c1=2, and Q=P, we have

1ρlogEP[2ρK(X)]Hα(P)+1,

that is, the minimum in (15) is between Hα(P) and Hα(P)+1. Hence,

1ρlogEP[2ρL(X)]Hα(P)1Rc(P,L,ρ)1ρlogEP[2ρL(X)]Hα(P). (17)

Let us now define a probability distribution QL as

QL(x)=2L(x)/αx2L(x)/α·

Then,

2L(x)=QL(x)αZQL,α·1x2L(x)=QL(x)αZQL,α·1η,

where η=x2L(x). Applying Propositions 3 and 4 with n=1, ψ1(x)=2L(x), a1=c1=1/η, we obtain

1ρlogEP[2ρL(X)]=Hα(P)+Iα(P,QL)logη. (18)

Combining (17) and (18), we obtain the desired result. □

We remark that the bound in (16) can be loose when η is small. For example, for a source with two symbols, say x and y, with code lengths L(x)=L(y)=100, we have Rc(P,L,ρ)Iα(P,QL)+98. However, if one imposes the constraint 1/2η1, then (16) simplifies to

|Rc(P,L,ρ)Iα(P,QL)|1,

which is (Theorem 8 of [5]). Iα(P,QL) is, in a sense, the penalty when QL does not match the true distribution P. In view of this, a result analogous to Proposition 5 also holds for the Shannon source coding problem.

3.3. Arıkan’s Guessing Problem

Let X be a set of objects with |X|=m. Bob thinks of an object X (a random variable) from X according to a probability distribution P. Alice guesses it by asking questions of the form “Is X=x?”. The objective is to minimize average number of guesses required for Alice to guess X correctly. By a guessing strategy (or guessing function), we mean a one-one map G:X{1,,m}, where G(x) is to be interpreted as the number of questions required to guess x correctly. Arıkan studied the ρth moment of number of guesses and found upper and lower bounds in terms of Rényi entropy.

Theorem 1 of [13]: Let G be any guessing function. Then, for ρ(1,0)(0,),

1ρlogEPG(X)ρHα(P)log(1+lnm).

 Proof. 

Let G be any guessing function. Let ψ(x)=G(x). Then, we have xXψ(x)1=xX1/G(x)=i=1m1/i1+lnm. An application of Proposition 1 with k=1+lnm yields the desired result. □

Arıkan showed that an optimal guessing function guesses according to the decreasing order of P-probabilities with ties broken using an arbitrary but fixed rule [13]. He also showed that normalized cumulant of an optimal guessing function is bounded above by the Rényi entropy. Next, we present a proof of this using our general framework.

Proposition 4 of [13]: If G* is an optimal guessing function, then, for ρ(1,0)(0,),

1ρlogEPG*(X)ρHα(P).

 Proof. 

Let us rearrange the probabilities {P(x),xX} in non-increasing order, say

p1p2pm.

Then, the optimal guessing function G* is given by G*(x)=i if P(x)=pi. Let us index the elements in set X as {x1,x2,,xm}, according to the decreasing order of their probabilities. Then, for i{1,,m}, we have

ZP,αP(xi)α=j=1mpjαpiαi=G*(xi). (19)

That is, G*(x)ZP,αP(x)α for xX. Now, an application of Proposition 3 with n=1, Q=P, ψ1(x)=G*(x) and c1=1, gives us

1ρlogEP[G*(X)ρ]=1ρlogEP[ψ(X)ρ]Hα(P)+Iα(P,P)+log1=Hα(P).

Arıkan also proved that the upper bound of Rényi entropy can be achieved by guessing long sequences of letters in an i.i.d. fashion.

Proposition 5 of [13] Let X1,X2,,Xn be a sequence of i.i.d. guesses. Let Gn*(X1,,Xn) be an optimal guessing function. Then, for ρ(1,0)(0,),

limn1nρlogEPn[Gn*(X1,X2,,Xn)ρ]=Hα(P).

 Proof. 

Let Gn* be the optimal guessing function from Xn to {1,2,,mn}. An application of Corollary 1 with ψn(xn)=Gn*(xn) and kn=1+nlnm yields

lim infn1nρlogEPn[Gn*(Xn)ρ]Hα(P)lim supnlog(1+nlnm)n=Hα(P). (20)

As in the proof of the previous result, we know that G*(xn)ZPn,αPn(xn)α for xnXn. Hence, an application of Proposition 3 with Qn=Pn, ψn(xn)=Gn*(xn), and cn=1 yields

lim supn1nρlogEPn[Gn*(Xn)ρ]Hα(P). (21)

Combining (20) and (21), we obtain the desired result. □

Henceforth, we shall denote the optimal guessing function corresponding to a probability distribution P by GP.

Mismatch Case:

Suppose Alice does not know the true underlying probability distribution P, and guesses according to some guessing function G. The following proposition tells us that the penalty for deviating from the optimal guessing function can be measured by Iα-divergence.

 Proposition 6. 

Let G be an arbitrary guessing function. Then, for ρ(1,0)(0,), there exists a probability distribution QG on X such that

1ρlogEP[G(X)ρ]Hα(P)+Iα(P,QG)log(1+lnm).

 Proof. 

Let G be a guessing function. Define a probability distribution QG on X as

QG(x)=G(x)1/αxXG(x)1/α. (22)

Then, we have

ZQG,αQG(x)α=G(x)xX1G(x)G(x)·(1+lnm)·

Now, an application of Proposition 4 with n=1, ψ1(x)=G(x), and a1=1/(1+lnm) yields the desired result. □

A converse result is the following.

Proposition 1 of [5]: Let GQ be an optimal guessing function associated with Q. Then, for ρ(1,0)(0,),

1ρlogEPGQ(X)ρHα(P)+Iα(P,Q),

where the expectation is with respect to P.

 Proof. 

Let us rearrange the probabilities (Q(x),xX) in non-increasing order, say

q1q2qm.

By definition, GQ(x)=i if Q(x)=qi. Then, as in (19), we have GQ(x)ZQ,α/Q(x)αforxX. Hence, an application of Proposition 3 with n=1, ψ1(x)=GQ(x), and c1=1 proves the result. □

Observe that, given a guessing function G, if we apply the above proposition for Q=QG, where QG is as in (22), then we obtain

1ρlogEPnGQG(X)ρHα(P)+Iα(P,QG).

Thus, the above two propositions can be combined to state the following, which is analogous to Proposition 5 (refer Section 3.2).

Theorem 6 of [5]: Let G be an arbitrary guessing function and GP be the optimal guessing function for P. For ρ(1,0)(0,), let

Rg(P,G,ρ):=1ρlogEPG(X)ρ1ρlogEPGP(X)ρ.

Then, there exists a probability distribution QG such that

|Rg(P,G,ρ)Iα(P,QG)|log(1+lnm).

3.4. Memoryless Guessing

In memoryless guessing, the setup is similar to that of Arıkan’s guessing problem except that this time the guesser Alice comes up with guesses independent of her previous guesses. Let X^1,X^2, be Alice’s sequence of independent guesses according to a distribution P^. The guessing function in this problem is defined as

GP^(X):=inf{i1:X^i=X},

that is, the number of guesses until a successful guess. Sundaresan [28], inspired by Arıkan’s result, showed that the minimum expected number of guesses required is exp{H12(P)}, and the distribution that achieves this is surprisingly not the underlying distribution P, but the “tilted distribution” P^*(x):=P(x)/yP(y).

Unlike in Arıkan’s guessing problem, Huleihel et al. [19] minimized what are called factorial moments, defined for ρZ+ as

VP^,ρ(X)=1ρ!l=0ρ1GP^(X)+l.

Huleihel et al. [19] (c.f. [20]) studied the following problem.

MinimizeEPVP^,ρ(X),

over all P^P, where P is the probability simplex that is, P={(P(x))xX:P(x)0,xP(x)=1}. Let P^* be the optimal solution of the above problem.

Theorem 1 of [19]: For any integer ρ>0, we have

1ρlogEPVP^*,ρ(X)=Hα(P)

and P^*(x)=P(x)α/ZP,α.

 Proof. 

From [19], we know that

EPVP^,ρ(X)=EP[P^(X)ρ]. (23)

Now, the result follows from Proposition 1 with ψ(x)=P^(x)1 and k=1. Indeed, since P^ is a probability distribution, we have xXψ(x)1=xXP^(x)=1. Hence, 1ρlogEPVP^,ρ(X)Hα(P), and the lower bound is attained by P^*(x)=P(x)α/ZP,α. □

For a sequence of guesses, the above theorem can be stated in the following way. Let X^n=(X^1,,X^n), where X^i’s are i.i.d. guesses, drawn from Xn with distribution P^n—the n-fold product distribution of P^ on Xn. If the true underlying distribution is Pn, then

limn1nρlogEPnVP^n*ρ(Xn)=Hα(P),

where P^n*(x)=Pn(x)α/ZPn,α. For the mismatched case, we have the following result.

 Proposition 7. 

If the true underlying probability distribution is P, but Alice assumes it as Q and guesses according to its optimal one, namely Q^*(x)=Q(x)α/ZQ,α, then

1ρlogEPVQ^*,ρ(X)=Hα(P)+Iα(P,Q).

 Proof. 

Due to (23), the result follows easily by taking n=1, ψ1(x)=Q^*(x)1, c1=1, a1=1 in Propositions 3 and 4. □

3.5. Tasks Partitioning Problem

Encoding of Tasks problem studied by Bunte and Lapidoth [8] can be phrased in the following way. Let X be a finite set of tasks. A task X is randomly drawn from X according to a probability distribution P, which may correspond to the frequency of occurrences of tasks. Suppose these tasks are associated with M keys. Typically, M<|X|. Due to a limited availability of keys, more than one task may be associated with a single key. When a task needs to be performed, the key associated with it is pressed. Consequently, all tasks associated with this key will be performed. The objective in this problem is to minimize the number of redundant tasks performed. Usual coding techniques suggest assigning tasks with high probability to individual keys and leaving the low probability tasks unassigned. It may just be the case that some tasks may have a higher frequency of occurrence than others. However, for an individual, all tasks can be equally important. If M|X|, then one can perform tasks without any redundancy. However, Bunte and Lapidoth [8] showed that, even when M<|X|, one can accomplish the tasks with much less redundancy on average, provided the underlying probability distribution is different from the uniform distribution.

Let A={A1,A2,,AM} be a partition of X that corresponds to the assignment of tasks to M keys. Let A(x) be the cardinality of the subset containing x in the partition. We shall call A the partition function associated with partition A. We shall assume that ρ>0 throughout this section, though some of the results hold even when ρ(1,0).

Theorem I.1 of [8]: The following results hold.

  • (a)
    For any partition of X of size M with partition function A, we have
    1ρlogEP[A(X)ρ]Hα(P)logM.
  • (b)
    If M>log|X|+2, then there exists a partition of X of size at most M with partition function A such that
    1EP[A^(X)ρ]1+2ρ(Hα(P)logM˜),
    where
    M˜:=(Mlog|X|2)/4. (24)

 Proof. 

Part (a): Let ψ(x)=A(x). Then, we have xXψ(x)1=xXA(x)1=M (Prop. III-1 of [8]). Now, an application of Proposition 1 with k=M gives us the desired result.

Part (b): For the proof of this part, we refer to [8]. □

Bunte and Lapidoth also proved the following limit results.

Theorem I.2 of [8]: Then, for every n1, there exists a partition An of Xn of size at most Mn with an associated partition function An such that

limnEPn[An(Xn)ρ]=1iflogM>Hα(P)iflogM<Hα(P),

where Xn:=(X1,,Xn).

It should be noted that, in a general set-up of the tasks partitioning problem, it is not necessary that the partition size is of the form Mn; it can be some Mn (a function of n). Consequently, we have the following result.

 Proposition 8. 

Let {Mn} be a sequence of positive integers such that Mnnlog|X|+3, and

γ:=limnlogMnn

exists. Then, there exists a sequence of partitions of Xn of size at most Mn with partition functions An such that

  • (a) 
    limnEPn[An(Xn)ρ]=1ifγ>Hα(P),
  • (b) 
    limn1nρlogEPn[An(Xn)ρ]=Hα(P)γifγ<Hα(P).

 Proof. 

Let

M˜n:=(Mnnlog|X|2)/4. (25)

We first claim that limnlogM˜nn=γ. Indeed, since log(1/4)nlogM˜nn<logMnn, when γ=0, we have limnlogM˜nn=0. On the other hand, when γ>0, we can find an nγ such that Mn2γn/2nnγ. Thus, we have limnnMn=0. Consequently,

limnlogM˜nn=limnlogMnn+limn1nlog1(nlog|X|+2)Mnlimn2n=γ.

This proves the claim. From Theorem I.1 of [11], for any n1 and Mn>nlog|X|+2, there exists a partition An of Xn of size at most Mn such that the associated partition function An satisfies

EPn[An(Xn)ρ]1+2ρ(Hα(Pn)logM˜n)=1+2nρHα(P)logM˜nn.

Part (a): When γ>Hα(p), let us choose ϵ=(γHα(P))/2>0. Then, there exists an nϵ such that logM˜nnγϵnnϵ. Thus, we have

EPn[An(Xn)ρ]1+2nρ(Hα(P)γ+ϵ)=1+2nρ(γHα(P))/2nnϵ

Consequently, lim supnEPn[An(Xn)ρ]1. We also note that An(xn)1 for all xnXn.

Thus, lim infnEPn[An(Xn)ρ]1.

Part (b): For any ϵ>0, there exists an nϵ such that logM˜nnγϵnnϵ. Thus, we have

EPn[An(Xn)ρ]1+2nρ(Hα(P)γ+ϵ)21+nρ(Hα(P)γ+ϵ)nnϵ

Hence, we have

lim supn1nρlogEPn[An(Xn)ρ]Hα(P)γ+ϵϵ>0

Furthermore, an invocation of Corollary 1 with ψn(xn)=An(xn) and kn=xnXn1/An(xn)=Mn gives us

lim infn1nρlogEPn[An(Xn)ρ]Hα(P)lim supnlogMnn=Hα(P)γ.

 Remark 1. 

It is interesting to note that, when γ<Hα(P), in addition to the fact that limnEPn[An(Xn)ρ]=, we also have EPn[An(Xn)ρ]2nρ(Hα(P)γ) for large values of n.

Mismatch Case:

Let us now suppose that one does not know the true underlying probability distribution P, but arbitrarily partitions X. Then, the penalty due to such a partition can be measured by the Iα-divergence as stated in the following theorem.

 Proposition 9. 

Let A be a partition of X of size M with partition function A. Then, there exists a probability distribution QA on X such that

1ρlogEP[A(X)ρ]=Hα(P)+Iα(P,QA)logM.

 Proof. 

Define a probability distribution QA={QA(x),xX} as

QA(x):=A(x)1/αxXA(x)1/α·

Then,

ZQA,αQA(x)α=A(x)xX1A(x)=A(x)·M,

where the last equality follows due to Proposition III.1 of [8]. Rearranging terms, we have A(x)=ZQA,αM·QA(x)α. Hence, an application of Propositions 3 and 4 with n=1, ψ1(x)=A(x), c1=1/M, a1=1/M, and Q=QA yields the desired result. □

A converse result is the following.

 Proposition 10. 

Let X be a random task from X following distribution P and ρ(0,). Let Q be another distribution on X. If M>log|X|+2, then there exists a partition AQ (with an associated partition function AQ) of X of size at most M such that

EP[AQ(X)ρ]1+2ρ(Hα(P)+Iα(P,Q)logM˜),

where M˜ is as in (24).

 Proof. 

Similar to proof of Theorem I.1 of [8]. □

4. Ordered Tasks Partitioning Problem

In Bunte–Lapidoth’s tasks partitioning problem [8], one is interested in the average number of tasks associated with a key. However, in some scenarios, it might be more important to minimize the average number of redundant tasks performed, before the intended task. To achieve this, tasks associated with a key should be performed in a decreasing order of their probabilities. With such a strategy in place, this problem draws parallel with Arıkan’s guessing problem [13].

Let A={A1,A2,,AM} be a partition of X that corresponds to the assignment of tasks to M keys. Let N(x) be the number of redundant tasks performed until and including the intended task x. We refer to N(·) as the count function associated with partition A. We suppress the dependence of N on A for the sake of notational convenience. If X denotes the intended task, then we are interested in the ρth moment of number of tasks performed, that is, EP[N(X)ρ], where ρ>0.

 Lemma 1. 

For any count function associated with a partition of size M, we have

xX1N(x)M1+ln|X|M. (26)

 Proof. 

For a partition A={A1,A2,,AM} of X, observe that

xX1N(x)=1+12++1|A1|++1+12++1|AM|. (27)

Since 1+12++1|Ak|1+ln|Ak|, for any k{1,,M}, we have

xX1N(x)M+ln(|A1||AM|)=M[1+ln(|A1||AM|)1/M](a)M1+ln|A1|++|AM|M=M1+ln|X|M, (28)

where (a) follows due to the AM–GM inequality. □

 Proposition 11. 

Let X be a random task from X following distribution P. Then, the following hold:

  • (a)
    For any partition of X of size M, we have
    1ρlogEP[N(X)ρ]Hα(P)log{M[1+ln|X|/M]} (29)
  • (b)
    Let M>log|X|+2. Then, there exists a partition of X of size at most M with count function N such that
    1EP[N(X)ρ]1+2ρ(Hα(P)logM˜),
    where M˜ is as in (24).

 Proof. 

Part (a): Applying Proposition 1 with k=M1+ln|X|/M and ψ(x)=N(x), we obtain the desired result.

Part (b): If A and N are, respectively, the partition and count functions of a partition A, then we have 1N(x)A(x) for xX. Once we observe this, the proof is same as Theorem I.1 (b) of [8]. □

 Proposition 12. 

Let {Mn} be a sequence of positive integers such that Mnnlog|X|+3, and γ:=limnlogMn/n exists. Then, there exists a sequence of partitions of Xn of size at most Mn with count functions Nn such that

  • (a) 
    limnEPn[Nn(Xn)ρ]=1ifγ>Hα(P),
  • (b) 
    limn1nρlogEPn[Nn(Xn)ρ]=Hα(P)γifγ<Hα(P).

 Proof. 

Similar to proof of Proposition 8. □

Remark 2. 

  • (a) 
    If we choose the trivial partition, namely An={Xn}, then the ordered tasks partitioning problem simplifies to Arıkan’s guessing problem, that is, we have Mn=1, Nn(xn)=Gn(xn) and (26) simplifies to
    xnXn1Gn(xn)1+nln|X|.

    Hence, all results pertaining to the Arıken’s guessing problem can be derived from the ordered tasks partitioning problem.

  • (b) 

    Structurally, ordered tasks partitioning problem differs from the Bunte–Lapidoth’s problem only due the factor 1+ln(|X|/M) in (28). While this factor matters for one-shot results, for a sequence of i.i.d. tasks, this factor vanishes asymptotically.

Mismatch Case:

Let us now suppose that one does not know the true underlying probability distribution P, but arbitrarily partitions X and executes tasks within each subset of this partition in an arbitrary order. Then, the penalty due to such a partition and ordering can be measured by the Iα-divergence as stated in the following propositions.

 Proposition 13. 

Let A be a partition of X of size M with count function N. Then, there exists a probability distribution QN on X such that

1ρlogEP[N(X)ρ]Hα(P)+Iα(P,QN)log{M1+ln|X|/M}.

 Proof. 

Define a probability distribution QN={QN(x),xX} as

QN(x):=N(x)1/αxXN(x)1/α·

Then, by Lemma 1, we have

ZQN,αQN(x)α=N(x)xX1N(x)N(x)·M[1+ln(|X|/M)].

Now, an application of Proposition 4 with n=1, ψ1(x)=N(x), Q=QN, and a1=1/M[1+ln(|X|/M)] yields the desired result. □

A converse result is the following.

 Proposition 14. 

Let X be a random task from X following distribution P. Let Q be another distribution on X. If M>log|X|+2, then there exists partition AQ (with an associated count function NQ) of X of size at most M such that

EP[NQ(X)ρ]1+2ρ(Hα(P)+Iα(P,Q)logM˜)ifρ(0,),

where M˜ is as in (24).

 Proof. 

Identical to the proof of Proposition 11(b). □

5. Operational Connection among the Problems

In this section, we establish an operational relationship among the five problems (refer Figure 1) that we studied in the previous section. The relationship we are interested in is “Does knowing an optimal or asymptotically optimal solution in one problem helps us find the same in another?” In fact, we end up showing that, under suitable conditions, all the five problems form an equivalence class with respect to the above-mentioned relation.

Figure 1.

Figure 1

Relationships established among the five problems. A directed arrow from problem A to problem B means knowing optimal or asymptotically optimal solution of A helps us find the same in B.

In this section, we assume ρ>0. First, we make the following observations:

  • Among the five problems discussed in the previous section, only Arıkan’s guessing and Huleihel et al.’s memoryless guessing have a unique optimal solution; others only have asymptotically optimal solutions.

  • Optimal solution of Huleihel et al.’s memoryless guessing problem is the α-scaled measure of the underlying probability distribution P. Hence, knowledge about the optimal solution of this problem implies knowledge about an optimal (or asymptotically optimal) solution of all other problems.

  • Among the Bunte–Lapidoth’s and ordered tasks problems, an asymptotically optimal solution of one yields that of the other. The partitioning lemma (Prop. III-2 of [8]) is the key result in these two problems, as it guarantees the existence of the asymptotically optimal partitions in both these problems.

5.1. Campbell’s Coding and Arıkan’s Guessing

An attempt to find a close relationship between these two problems was made, for example, by Hanawal and Sundaresan (Section II of [17]). Here, we show the equivalence between asymptotically optimal solutions of these two problems.

 Proposition 15. 

An asymptotically optimal solution exists for Campbell’s source coding problem if and only if an asymptotically optimal solution exists for Arıkan’s guessing problem.

 Proof. 

Let {Gn*} be an asymptotically optimal sequence of guessing functions, that is,

limn1nρlogEPn[Gn*(Xn)ρ]=Hα(P).

Define

QGn*(xn):=cn1·Gn*(xn)1, (30)

where cn is the normalization constant. Notice that

cn=xnGn*(xn)11+nln|X|. (31)

Let us now define

LGn*(xn):=logQGn*(xn).

Then, by (Proposition 1 of [17]),

LGn*(xn)logGn*(xn)+1+logcn.

Hence,

2ρLGn*(xn)2ρ·cnρ·Gn*(xn)ρ2ρ·(1+nln|X|)ρ·Gn*(xn)ρ.

Thus, we have

lim supn1nρlogEPn[2ρLGn*(Xn)]lim supn1nρlogEPn[Gn*(Xn)ρ]=Hα(P).

We observe that

xnXn2LGn*(xn)=xnXn2logQGn*(xn)xnXn2logQGn*(xn)=1.

Consequently, from (12), we have

lim infn1nρlogEPn[2ρLGn*(Xn)]Hα(P).

Thus, {LGn*} is an asymptotically optimal sequence of length functions for Campbell’s coding problem.

Conversely, given an asymptotically optimal sequence of length functions {Ln*} for Campbell’s coding problem, define

QLn*(xn):=2Ln*(xn)yn2Ln*(xn).

Let GLn* be the guessing function on Xn that guesses according to the decreasing order of QLn*-probabilities. Then, by (Proposition 2 of [17]),

logGLn*(xn)Ln*(xn).

Thus,

lim supn1nρlogEPn[GLn*(Xn)ρ]limn1nρlogEPn[2ρLn*(Xn)]=Hα(P).

Furthermore, from Theorem 1 of [13], we have

lim infn1nρlogEPn[GLn*(Xn)ρ]Hα(P)lim supn1nlog(1+nln|X|)=Hα(P).

This completes the proof. □

5.2. Arıkan’s Guessing and Bunte–Lapidoth’s Tasks Partitioning Problem

Bracher et al. found a close connection between Arıkan’s guessing problem and Bunte–Lapidoth’s tasks partitioning problem in the context of distributed storage [29]. In this section, we establish a different relation between these problems.

 Proposition 16. 

An asymptotically optimal solution of Arıkan’s guessing problem gives rise to an asymptotically optimal solution of tasks partitioning problem.

 Proof. 

Let {Gn*} be an asymptotically optimal sequence of guessing functions. Define

QGn*(xn)=dn1Gn*(xn)1/α,

where dn is the normalization constant. Let AGn* be the partition function satisfying AGn*(xn)βn·ZQGn*,α/QGn*(xn)α guaranteed by (Proposition III-2 of [8]), where

βn=2Mnnlog|X|2·

Thus, we have

AGn*(xn)ρβn·ZQGn*,α/QGn*(xn)αρ(a)βn·(1+nln|X|)·Gn*(xn)ρ(b)1+2ρβnρ·(1+nln|X|)ρ·Gn*(xn)ρ,

where (a) holds because ZQGn*,α=dnαxnXnGn*(xn)1dnα(1+nln|X|); and (b) hold because xρ1+2ρxρ for x>0. Hence,

EPn[AGn*(Xn)ρ]1+2ρβnρ·(1+nln|X|)ρ·EPn[Gn*(Xn)ρ](c)1+22ρ1+nln|X|Mnnlog|X|2ρ·2nρHα(P)=1+2nρHα(P)+log(1+nln|X|)nlogM˜nn,

where M˜n is as in (25); and inequality (c) follows from Proposition 4 of [13] proved in Section 3. Thus, if Mn is such that Mnnlog|X|+3 and if γ:=limn(logMn)/n exists and γ>Hα(P), then we have

lim supnEPn[AGn*(Xn)ρ]1.

Since EPn[AGn*(Xn)ρ]1, we have lim infnEPn[AGn*(Xn)ρ]1. When γ<Hα(P), arguing along the lines of proof of Proposition 8(b), it can be shown than

limn1nρlogEPn[AGn*(Xn)ρ]=Hα(P)γ.

Reverse implication of the above result does not hold always due to the additional parameter Mn in the tasks partitioning problem. For example, if Mn=|X|n and An(xn)=1 for every xn, the partition does not provide any information about the underlying distribution. As a consequence, we will not be able to conclude anything about the optimal (or asymptotically optimal) solutions of other problems. However, if Mn is such that logMn increases sub-linearly, then it does help us find asymptotically optimal solutions of other problems.

 Proposition 17. 

An asymptotically optimal sequence of partition functions {An} with partition sizes {Mn} for the tasks partitioning problem gives rise to an asymptotically optimal solution for the guessing problem if Mnnlog|X|+3 and limn(logMn)/n=0.

 Proof. 

By hypothesis,

limn1nρlogEPn[An(Xn)ρ]=Hα(P).

For every An, define the probability distribution

QAn(xn):=cn1An(xn)1,

where cn:=xnAn(xn)1=Mn. Let GAn* be the guessing function that guesses according to the decreasing order of QAn-probabilities. Then, by (Proposition 2 of [17]), we have

GAn*(xn)QAn(xn)1=cnAn(xn)=MnAn(xn).

Hence,

lim supn1nρlogEPn[GAn*(Xn)ρ]lim supn1nlogMn+lim supn1nρlogEPn[An(Xn)ρ]=Hα(P).

Furthermore, an application of Theorem 1 of [13] gives us

lim infn1nρlogEPn[GAn*(Xn)ρ]Hα(P)lim supn1nlog(1+nln|X|)=Hα(P).

This completes the proof. □

5.3. Huleihel et al.’s Memoryless Guessing and Campbell’s Coding

We already know that, if one knows the optimal solution of Huleihel et al.’s memoryless guessing problem, that is, the α-scaled measure of the underlying probability distribution P, one has knowledge about the optimal (or asymptotically optimal) solution of Campbell’s coding problem. In this section, we prove a converse statement. We first prove the following lemma.

 Lemma 2. 

Let Ln* denote the length function corresponding to an optimal solution for Campbell’s coding problem on the alphabet set Xn endowed with the product distribution Pn. Then, xnXn2Ln*(xn)1/2.

 Proof. 

Suppose xnXn2Ln*(xn)<1/2. Then, we must have Ln*(xn)2 for every xnXn. Define L^n(xn):=Ln*(xn)1. We observe that xnXn2L^n(xn)<1, that is, the length function L^n(·) satisfies (11). Hence, there exists a code Cn for Xn such that L(Cn(xn))=L^n(xn). Then, for ρ>0, we have logEPn[2ρLn*(Xn)]>logEPn[2ρL^n(Xn)]—a contradiction. □

 Proposition 18. 

An asymptotically optimal solution for Huleihel et al.’s memoryless guessing problem exists if an asymptotically optimal solution exists for Campbell’s coding problem.

 Proof. 

Let {Ln*,n1} denote a sequence of asymptotically optimal length functions of Campbell’s coding problem, that is,

limn1nρlogEPn[2ρLn*(Xn)]=Hα(P). (32)

Let us define

QLn*(xn):=2Ln*(xn)/αx¯nXn2Ln*(x¯n)/α.

Then, we have

Iα(Pn,QLn*)=α1αlogxnXnPn(xn)x^nXnQLn*(x^n)QLn*(xn)α1ααHα(Pn)=α1αlogxnXnPn(xn)x^nXn2Ln*(x^n)2Ln*(xn)ρnHα(P)=1ρlogEPn[2ρLn*(Xn)]+logζnnHα(P),

where ζn=x^nXn2Ln*(x^n). Hence,

limn1nIα(Pn,QLn*)=limn1nρlogEPn[2ρLn*(Xn)]+limn1nlogζnHα(P)=(a)0,

where (a) holds because ζn[1/2,1] (refer Lemma 2). If we assume the underlying probability distribution to be QLn* instead of Pn, and perform memoryless guessing according to the escort distribution of QLn*, namely Q^n*(xn)=QLn*(xn)α/ZQLn*,α, due to Proposition 7, we have

limn1nρlogEPnVQ^n*,ρ(Xn)=Hα(P)+limn1nIα(Pn,QLn*)=Hα(P).

6. Summary and Conclusions

This paper was motivated by the need to find a unified framework for the problems on source coding, guessing and the tasks partitioning. To that end, we formulated a general moment minimization problem in the IID-lossless case and observed that the optimal value of its objective function is bounded below by Rényi entropy. We then re-established all achievable lower bounds in each of the above-mentioned problems using the generalized framework. It was interesting to note that the optimal solution did not depend on the moment function ψ, but only on the underlying probability distribution P and order of the moment ρ (refer Proposition 1). We also presented a unified framework for the mismatched version of the above-mentioned problems. This framework not only led to refinement of the known theorems, but also helped us identify a few new results. We went on to extend the tasks partitioning problem by asking a more practical question and solved it using the unified theory. Finally, we established an equivalence among these problems, in the sense that an asymptotically optimal solution of one problem yields the asymptotically optimal solution of all other problems. Although the relationship between source coding and guessing is well-known in the literature [30,31,32,33], their connection to the tasks partitioning problem, and the connection in the mismatched version of the problems are new.

Our unified framework also has the potential to act as a general tool-set and provide insights for similar problems in information theory. For example, in Section 4, this framework enabled us to solve a more general tasks partitioning problem, namely, the ordered tasks partitioning problem using this framework. The presented that a unified approach can also be extended and explored further in several ways. This includes (a) Extension to general alphabet set: The guessing problem was originally studied for countably infinite alphabet set by Massey [12]. Courtade and Verdú have studied the source coding problem for a countably infinite alphabet set with a cumulant generating function of code word lengths as a design criterion [31]. It would be interesting to see if memory-less guessing and tasks partitioning problems can also be formulated for countably infinite alphabet sets and relationships among the problems can be extended. (b) More general sources: Relationship between source coding and guessing is very well-known in the literature. Relationship between guessing and source coding in the ‘with distortion’ case for finite alphabet was established by Merhav and Arıkan [14] and for a countably infinite alphabet by Hanawal and Sundaresan [34]. Relationship between Guessing and Campbell’s coding in the universal case was established by Sundaresan [35]. It would be interesting to see if these can be extended to memoryless guessing and tasks partitioning also, possibly in an unified manner. (c) Applications: Arıkan showed an application of the guessing problem in a sequential decoding problem [13]. Humblet showed that the cumulant of code-lengths arises in minimizing the probability of buffer overflow in source coding problems [27]. Rezaee et al. [36], Salamatian et al. [20], and Sundaresan [28] show the application of guessing in the security aspect of password protected systems. Our unified framework has the potential to help solve problems that arise in real life situations and fall in this framework.

Author Contributions

Conceptualization: M.A.K., A.S., A.T., A.K. and G.D.M.; Formal analysis: M.A.K. and A.S.; Investigation: M.A.K., A.S., A.T. and G.D.M.;Writing – original draft: M.A.K., A.S., A.T., A.K. and G.D.M. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research received no external funding.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Shannon C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948;27:379–423. doi: 10.1002/j.1538-7305.1948.tb01338.x. [DOI] [Google Scholar]
  • 2.Rényi A. On measures of entropy and information; Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics; Berkeley, CA, USA. 20 June–30 July 1960; pp. 547–561. [Google Scholar]
  • 3.Aczél J., Daróczy Z. Mathematicas in Science and Engineering. Volume 115 Academic Press; Cambridge, MA, USA: 1975. On Measures of Information and Their Characterizations. [Google Scholar]
  • 4.Campbell L.L. A coding theorem and Rényi’s entropy. Inf. Control. 1965;8:423–429. doi: 10.1016/S0019-9958(65)90332-3. [DOI] [Google Scholar]
  • 5.Sundaresan R. Guessing Under Source Uncertainty. IEEE Trans. Inf. Theory. 2007;53:269–287. doi: 10.1109/TIT.2006.887466. [DOI] [Google Scholar]
  • 6.Blumer A.C., McEliece R.J. The Rényi redundancy of generalized Huffman codes. IEEE Trans. Inf. Theory. 1988;34:1242–1249. doi: 10.1109/18.21251. [DOI] [Google Scholar]
  • 7.Sundaresan R. A measure of discrimination and its geometric properties; Proceedings of the IEEE International Symposium on Information Theory; Lausanne, Switzerland. 30 June–5 July 2002; p. 264. [Google Scholar]
  • 8.Bunte C., Lapidoth A. Encoding Tasks and Rényi Entropy. IEEE Trans. Inf. Theory. 2014;60:5065–5076. doi: 10.1109/TIT.2014.2329490. [DOI] [Google Scholar]
  • 9.Kumar M.A., Sundaresan R. Minimization problems based on relative α-entropy I: Forward Projection. IEEE Trans. Inf. Theory. 2015;61:5063–5080. doi: 10.1109/TIT.2015.2449311. [DOI] [Google Scholar]
  • 10.Lutwak E., Yang D., Zhang G. Cramér-Rao and moment-entropy inequalities for Rényi entropy and generalized Fisher information. IEEE Trans. Inf. Theory. 2005;51:473–478. doi: 10.1109/TIT.2004.840871. [DOI] [Google Scholar]
  • 11.Ashok Kumar M., Sundaresan R. Minimization Problems Based on Relative α-Entropy II: Reverse Projection. IEEE Trans. Inf. Theory. 2015;61:5081–5095. doi: 10.1109/TIT.2015.2449312. [DOI] [Google Scholar]
  • 12.Massey J.L. Guessing and entropy; Proceedings of the 1994 IEEE International Symposium on Information Theory; Trondheim, Norway. 27 June–1 July 1994; p. 204. [Google Scholar]
  • 13.Arikan E. An inequality on guessing and its application to sequential decoding. IEEE Trans. Inf. Theory. 1996;42:99–105. doi: 10.1109/18.481781. [DOI] [Google Scholar]
  • 14.Arikan E., Merhav N. Guessing subject to distortion. IEEE Trans. Inf. Theory. 1998;44:1041–1056. doi: 10.1109/18.669158. [DOI] [Google Scholar]
  • 15.Pfister C., Sullivan W. Renyi entropy, guesswork moments, and large deviations. IEEE Trans. Inf. Theory. 2004;50:2794–2800. doi: 10.1109/TIT.2004.836665. [DOI] [Google Scholar]
  • 16.Malone D., Sullivan W. Guesswork and entropy. IEEE Trans. Inf. Theory. 2004;50:525–526. doi: 10.1109/TIT.2004.824921. [DOI] [Google Scholar]
  • 17.Hanawal M.K., Sundaresan R. Guessing Revisited: A Large Deviations Approach. IEEE Trans. Inf. Theory. 2011;57:70–78. doi: 10.1109/TIT.2010.2090221. [DOI] [Google Scholar]
  • 18.Christiansen M.M., Duffy K.R. Guesswork, Large Deviations, and Shannon Entropy. IEEE Trans. Inf. Theory. 2013;59:796–802. doi: 10.1109/TIT.2012.2219036. [DOI] [Google Scholar]
  • 19.Huleihel W., Salamatian S., Médard M. Guessing with limited memory; In Proceeding of the 2017 IEEE International Symposium on Information Theory (ISIT); Aachen, Germany. 25–30 June 2017; pp. 2253–2257. [Google Scholar]
  • 20.Salamatian S., Huleihel W., Beirami A., Cohen A., Médard M. Why Botnets Work: Distributed Brute-Force Attacks Need No Synchronization. IEEE Trans. Inf. Forensics Secur. 2019;14:2288–2299. doi: 10.1109/TIFS.2019.2895955. [DOI] [Google Scholar]
  • 21.Arikan E., Merhav N. Joint source-channel coding and guessing with application to sequential decoding. IEEE Trans. Inf. Theory. 1998;44:1756–1769. doi: 10.1109/18.705557. [DOI] [Google Scholar]
  • 22.Csiszár I., Shields P. Information Theory and Statistics: A Tutorial, Foundations and Trends in Communications and Information Theory. Now Publishers; Delft, The Netherlands: 2004. [Google Scholar]
  • 23.Tsallis C., Mendes R.S., Plastino A.R. The role of constraints within generalized non-extensive statistics. Phys. A. 1998;261:534–554. doi: 10.1016/S0378-4371(98)00437-3. [DOI] [Google Scholar]
  • 24.Dupuis P., Ellis R.S. A Weak Convergence Approach to the Theory of Large Deviations. John Wiley & Sons; Hoboken, NJ, USA: 1997. [Google Scholar]
  • 25.Shayevitz O. On Rényi measures and hypothesis testing; Proceedings of the 2011 IEEE International Symposium on Information Theory Proceedings; St. Petersburg, Russia. 31 July–5 August 2011; pp. 894–898. [Google Scholar]
  • 26.Cover T.M., Thomas J.A. Elements of Information Theory. 2nd ed. John Wiley & Sons; Hoboken, NJ, USA: 2012. [Google Scholar]
  • 27.Humblet P. Generalization of Huffman coding to minimize the probability of buffer overflow. IEEE Trans. Inf. Theory. 1981;27:230–232. doi: 10.1109/TIT.1981.1056322. [DOI] [Google Scholar]
  • 28.Hanawal M.K., Sundaresan R. Randomised Attacks on Passwords. DRDO-IISc Programme on Advanced Research in Mathematical Engineering. 2010. [(accessed on 14 August 2022)]. Available online: https://ece.iisc.ac.in/rajeshs/reprints/TR-PME-2010-11.pdf.
  • 29.Bracher A., Hof E., Lapidoth A. Guessing Attacks on Distributed-Storage Systems. IEEE Trans. Inf. Theory. 2019;65:6975–6998. doi: 10.1109/TIT.2019.2933000. [DOI] [Google Scholar]
  • 30.Salamatian S., Liu L., Beirami A., Médard M. Mismatched guesswork and one-to-one codes; Proceedings of the 2019 IEEE Information Theory Workshop (ITW); Gotland, Sweden. 25–28 August 2019; pp. 1–5. [Google Scholar]
  • 31.Courtade T.A., Verdú S. Cumulant generating function of codeword lengths in optimal lossless compression; Proceedings of the 2014 IEEE International Symposium on Information Theory; Honolulu, HI, USA. 29 June –4 July 2014; pp. 2494–2498. [DOI] [Google Scholar]
  • 32.Kosut O., Sankar L. Asymptotics and non-asymptotics for universal fixed-to-variable source coding. IEEE Trans. Inf. Theory. 2017;63:3757–3772. doi: 10.1109/TIT.2017.2686881. [DOI] [Google Scholar]
  • 33.Beirami A., Calderbank R., Christiansen M.M., Duffy K.R., Médard M. A characterization of guesswork on swiftly tilting curves. IEEE Trans. Inf. Theory. 2018;60:2850–2871. [Google Scholar]
  • 34.Hanawal M.K., Sundaresan R. Guessing and Compression Subject to Distortion. IndraStra Global; Sheridan, WY, USA: 2010. [Google Scholar]
  • 35.Sundaresan R. Guessing Based On Length Functions; Proceedings of the 2007 IEEE International Symposium on Information Theory; Cambridge, MA, USA. 1–6 July 2007; pp. 716–719. [DOI] [Google Scholar]
  • 36.Rezaee A., Beirami A., Makhdoumi A., Médard M., Duffy K. Guesswork subject to a total entropy budget; Proceedings of the 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton); Monticello, IL, USA. 3–6 October 2017; pp. 1008–1015. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES