Skip to main content
Springer logoLink to Springer
. 2021 Jul 1;83(9):2833–2858. doi: 10.1007/s00453-021-00844-5

Best Fit Bin Packing with Random Order Revisited

Susanne Albers 1, Arindam Khan 2, Leon Ladewig 1,
PMCID: PMC8550633  PMID: 34776569

Abstract

Best Fit is a well known online algorithm for the bin packing problem, where a collection of one-dimensional items has to be packed into a minimum number of unit-sized bins. In a seminal work, Kenyon [SODA 1996] introduced the (asymptotic) random order ratio as an alternative performance measure for online algorithms. Here, an adversary specifies the items, but the order of arrival is drawn uniformly at random. Kenyon’s result establishes lower and upper bounds of 1.08 and 1.5, respectively, for the random order ratio of Best Fit. Although this type of analysis model became increasingly popular in the field of online algorithms, no progress has been made for the Best Fit algorithm after the result of Kenyon. We study the random order ratio of Best Fit and tighten the long-standing gap by establishing an improved lower bound of 1.10. For the case where all items are larger than 1/3, we show that the random order ratio converges quickly to 1.25. It is the existence of such large items that crucially determines the performance of Best Fit in the general case. Moreover, this case is closely related to the classical maximum-cardinality matching problem in the fully online model. As a side product, we show that Best Fit satisfies a monotonicity property on such instances, unlike in the general case. In addition, we initiate the study of the absolute random order ratio for this problem. In contrast to asymptotic ratios, absolute ratios must hold even for instances that can be packed into a small number of bins. We show that the absolute random order ratio of Best Fit is at least 1.3. For the case where all items are larger than 1/3, we derive upper and lower bounds of 21/16 and 1.2, respectively.

Keywords: Online bin packing, Random arrival order, Probabilistic analysis

Introduction

One of the fundamental problems in combinatorial optimization is bin packing. Given a list I=(x1,,xn) of n items with sizes from (0, 1] and an infinite number of unit-sized bins, the goal is to pack all items into the minimum number of bins. Formally, a packing is an assignment of items to bins such that for any bin, the sum of assigned items is at most 1. While an offline algorithm has complete information about the items in advance, in the online variant, items are revealed one by one. An online algorithm must pack xi without knowing the items following xi and without modifying the packing of previous items.

Bin packing was mentioned first by Ullman [38]. As the problem is strongly NP-complete [17], research mainly focuses on efficient approximation algorithms. The offline problem is well understood and admits even approximation schemes [20, 26, 39]. The online variant is still a very active field in the community [7], as the asymptotic approximation ratio of the best online algorithm is still unknown [3, 4]. The first approximation algorithms for the problem, First Fit and Best Fit, have been analyzed in [38] and a subsequent work by Garey et al. [16]. Johnson published the Next Fit algorithm briefly afterwards [24]. All of these algorithms work in the online setting and attract by their simplicity: Suppose that xi is the current item to pack. The algorithms work as follows:

BestFit(BF)

Pack xi into the fullest bin possible, open a new bin if necessary.

FirstFit(FF)

Maintain a list of bins ordered by the time at which they were opened. Pack xi into the first possible bin in this list, open a new bin if necessary.

NextFit(NF)

Pack xi into the bin opened most recently if possible; open a new bin if necessary.

Another important branch of online algorithms is based on the harmonic algorithm [29]. This approach has been massively tuned and generalized in a sequence of papers [3, 35, 36].

To measure the performance of an algorithm, different metrics exist. For an algorithm A, let A(I) and OPT(I) denote the number of bins used by A and an optimal offline algorithm, respectively, to pack the items in I. Let I denote the set of all item lists. The most common metric for bin packing algorithms is the asymptotic (approximation) ratio defined as

RA=lim supksupII{A(I)/OPT(I)OPT(I)=k}.

Note that RA focuses on instances where OPT(I) is large. This avoids anomalies typically occurring on lists that can be packed optimally into few bins. However, many bin packing algorithms are also studied in terms of the stronger absolute (approximation) ratio

RA=supII{A(I)/OPT(I)}.

Here, the approximation ratio RA must hold for each possible input. An online algorithm with (absolute or asymptotic) ratio α is also called α-competitive.

Table 1 shows the asymptotic and absolute approximation ratios of the three heuristics Best Fit, First Fit, and Next Fit. Interestingly, for these algorithms both metrics coincide. While the asymptotic ratios of Best Fit and Next Fit were established already in early work [25], the absolute ratios have been settled rather recently [11, 12].

Table 1.

Approximation ratios in different metrics of common bin packing heuristics. In RNF, the symbol γ refers to the total size of items in the list

Algorithm A Abs. ratio RA Asym. ratio RA Asym. random order ratio RRA
Best Fit 1.7 [12] 1.7 [25] 1.08RRBF1.5 [28]
First Fit 1.7 [11] 1.7 [25]
Next Fit 2-1/γ [5] 2 [24] 2 [9]

Note that the above performance measures are clearly worst-case orientated. An adversary can choose items and present them in an order that forces the algorithm into its worst possible behavior. In the case of Best Fit, hardness examples are typically based on lists where small items occur before large items [16]. In contrast, it is known that Best Fit performs significantly better if items appear in non-increasing order [25]. For real-world instances, it seems overly pessimistic to assume adversarial order of input. Moreover, sometimes worst-case ratios hide interesting properties of algorithms that occur in average cases. This led to the development of alternative measures.

A natural approach that goes beyond worst-case was introduced by Kenyon [28] in 1996. In the model of random order arrivals, the adversary can still specify the items, but the arrival order is permuted randomly. The performance measure described in [28] is based on the asymptotic ratio, but can be applied to absolute ratios likewise. In the resulting performance metrics, an algorithm must satisfy its performance guarantee in expectation over all permutations. We define

RRA=lim supksupIIE[A(Iσ)]/OPT(I)OPT(I)=kandRRA=supIIE[A(Iσ)]/OPT(I)

as the asymptotic random order ratio and the absolute random order ratio of algorithm A, respectively. Here, σ is drawn uniformly at random from Sn, the set of permutations of n elements, and Iσ=(xσ(1),,xσ(n)) is the permuted list.

Related Work

The following literature review only covers results that are most relevant to our work. We refer the reader to the article [8] by Coffman et al. for an extensive survey on (online) bin packing. For further problems studied in the random order model, see [19].

Bin packing Kenyon introduced the notion of asymptotic random order ratio RRA for online bin packing algorithms in [28]. For the Best Fit algorithm, Kenyon proves an upper bound of 1.5 on RRBF, demonstrating that random order significantly improves upon RBF=1.7. However, it is conjectured in [8, 28] that the actual random order ratio is close to 1.15. The proof of the upper bound crucially relies on the following scaling property: With high probability, the first t items of a random permutation can be packed optimally into tnOPT(I)+o(n) bins. On the other side, Kenyon proves that RRBF1.08. This lower bound is obtained from the weaker i.i.d.-model, where item sizes are drawn independently and identically distributed according to a fixed probability distribution.

Coffman et al. [9] analyzed next-fit in the random order model and showed that RRNF=2, matching the asymptotic approximation ratio RRNF=2 (see Table 1). Fischer and Röglin [14] obtained analogous results for Worst Fit [24] and Smart Next Fit [34]. Therefore, all three algorithms fail to perform better in the random order model than in the adversarial model.

A natural property of bin packing algorithms is monotonicity, which holds if an algorithm never uses fewer bins to pack I than for I, where I is obtained from I by increasing item sizes. Murgolo [33] showed that next-fit is monotone, while Best Fit and First Fit are not monotone in general. The concept of monotonicity also arises in related optimization problems, such as scheduling [18] and bin covering [14].

Bin covering The dual problem of bin packing is bin covering, where the goal is to cover as many bins as possible. A bin is covered if it receives items of total size at least 1. Here, a well-studied and natural algorithm is Dual Next Fit (DNF). In the adversarial setting, DNF has asymptotic ratio RDNF=1/2 which is best possible for any online algorithm [6]. Under random arrival order, Christ et al. [6] showed that RRDNF4/5. This upper bound was improved later by Fischer and Röglin [13] to RRDNF2/3. The same group of authors further showed that RRDNF0.501, i.e., DNF performs strictly better under random order than in the adversarial setting [14].

Matching Online matching can be seen as the key problem in the field of online algorithms [32]. Inspired by the seminal work of Karp et al. [27], who introduced the online bipartite matching problem with one-sided arrivals, the problem has been studied in many generalizations. Extensions include fully online models [15, 21, 22], vertex-weighted versions [1, 23] and, most relevant to our work, random arrival order [23, 31].

Our Results

While several natural algorithms fail to perform better in the random order model, Best Fit emerges as a strong candidate in this model. The existing gap between 1.08 and 1.5 clearly leaves room for improvement; closing (or even narrowing) this gap has been reported as challenging and interesting open problem in several papers [6, 9, 19]. To the best of our knowledge, our work provides the first new results on the problem since the seminal work by Kenyon. Below we describe our results in detail. In the following theorems, the expectation is over the permutation σ drawn uniformly at random.

If all items are strictly larger than 1/3, the objective is to maximize the number of bins containing two items. This problem is closely related to finding a maximum-cardinality matching in a vertex-weighted graph; our setting corresponds with the fully online model studied in [1] under random order arrival. Also in the analysis from [28], this special case arises. There, it is sufficient to argue that BF(I)32OPT(I)+1 under adversarial order. We show that Best Fit performs significantly better under random arrival order:

Theorem 1

For any list I of items larger than 1/3, we have E[BF(Iσ)]54OPT(I)+14.

The proof of Theorem 1 is developed in Sect. 3 and based on several pillars. First, we show that Best Fit is monotone in this case (Proposition 3), unlike in the general case [33]. This property can be used to restrict the analysis to instances with well-structured optimal packing. The main technical ingredient is introduced in Sect. 3.3 with Lemma 2 as the key lemma. Here, we show that Best Fit maintains some parts of the optimal packing, depending on certain structures of the input sequence. We identify these structures and show that they occur with constant probability for a random permutation. It seems likely that this property can be used in a similar form to improve the bound RRBF1.5 for the general case: Under adversarial order, much hardness comes from relatively large items of size more than 1/3; in fact, if all items have size at most 1/3, an easy argument shows 43-competitiveness even for adversarial arrival order [25].

Moreover, it is natural to ask for the performance in terms of absolute random order ratio. It is a surprising and rather recent result that for Best Fit, absolute and asymptotic ratios coincide. The result of [28] has vast additive terms and it seems that new techniques are required for insights into the absolute random order ratio. In Sect. 3.4, we investigate the absolute random order ratio for items larger than 1/3 and obtain the following result.

Proposition 1

For any list I of items larger than 1/3, we have E[BF(Iσ)]2116OPT(I).

The upper bound of 21/16 is complemented by the following lower bound.

Proposition 2

There is a list I of items larger than 1/3 with E[BF(Iσ)]>65OPT(I).

The proof of Proposition 2 is given in Appendix B.

We also make progress on the hardness side in the general case, which is presented in Sect. 4. First, we show that the asymptotic random order ratio is larger than 1.10, improving the previous lower bound of 1.08 from [28].

Theorem 2

The asymptotic random order ratio of Best Fit is RRBF>1.10.

As it is typically challenging to obtain lower bounds in the random order model, we exploit the connection to the i.i.d.-model. Here, items are drawn independently and identically distributed according to a fixed probability distribution. By defining an appropriate distribution, the problem can be analyzed using Markov chain techniques. Moreover, we present the first lower bound on the absolute random order ratio:

Theorem 3

The absolute random order ratio of Best Fit is RRBF1.30.

Interestingly, our lower bound on the absolute random order ratio is notably larger than in the asymptotic case (see [28] and Theorem 2). This suggests either

  • a significant discrepancy between RRBF and RRBF, which is in contrast to the adversarial setting (RBF=RBF, see Table 1), or

  • a disproof of the conjecture RRBF1.15 mentioned in [8, 28].

Notation

We consider a list I=(x1,,xn) of n items throughout the paper. Due to the online setting, I is revealed in rounds 1,,n. In round t, item xt arrives and in total, the prefix list I(t):=(x1,,xt) is revealed to the algorithm. The items in I(t) are called the visible items of round t. We use the symbol xt for the item itself and its size xt(0,1] interchangeably. An item xt is called large (L) if xt>1/2, medium (M) if xt1/3,1/2, and small (S) if xt1/3. We also say that xt is α-large if xt>α.

Bins contain items and therefore can be represented as sets. As a bin usually can receive further items in later rounds, the following terms refer always to a fixed round. We define the load of a bin B as xiBxi. Sometimes, we classify bins by their internal structure. We say B is of configuration LM (or B is an LM-bin) if it contains one large and one medium item. The configurations L, MM, etc. are defined analogously. Moreover, we call B a k-bin if it contains exactly k items. If a bin cannot receive further items in the future, it is called closed; otherwise, it is called open.

The number of bins which Best Fit uses to pack a list I is denoted by BF(I). We slightly abuse the notation and refer to the corresponding packing by BF(I) as well whenever the exact meaning is clear from the context. Similarly, we denote by OPT(I) the number of bins and the corresponding packing of an optimal offline solution.

Finally, for any natural number n we define [n]:={1,,n}. Let Sn be the set of permutations in [n]. If not stated otherwise, σ refers to a permutation drawn uniformly at random from Sn.

Upper Bound for 1/3-Large Items

In this section, we consider the case where I contains no small items, i.e., where all items are 13-large. We develop the technical foundations in Sects. 3.1 to 3.3. The final proofs of Theorem 1 and Proposition 1 are presented in Sect. 3.4.

Monotonicity

We first define the notion of monotone algorithms.

Definition 1

We call an algorithm monotone if increasing the size of one or more items cannot decrease the number of bins used by the algorithm.

One might suspect that any reasonable algorithm is monotone. While this property holds for an optimal offline algorithm and some online algorithms as Next Fit [10], Best Fit is not monotone in general [33]. As a counterexample, consider the lists

I=(0.36,0.65,0.34,0.38,0.28,0.35,0.62)andI=(0.36,0.65,0.36,0.38,0.28,0.35,0.62).

Before arrival of the fifth item, BF(I(4)) uses the two bins {0.36,0.38} and {0.65,0.34}, while BF(I(4)) uses three bins {0.36,0.36}, {0.65}, and {0.38}. Now, the last three items fill up the existing bins in BF(I(4)) exactly. In contrast, these items open two further bins in the packing of BF(I(4)). Therefore, BF(I)=4>3=BF(I).

However, we can show that Best Fit is monotone for the case of 13-large items. Interestingly, 1/3 seems to be the threshold for the monotonicity of Best Fit: As shown in the counterexample from the beginning of this section, it is sufficient to have one item x1/4,1/3 to force Best Fit into anomalous behavior. Anyway, we have the following proposition.

Proposition 3

Given a list I of items larger than 1/3 and a list I obtained from I by increasing the sizes of one or more items, we have BF(I)BF(I).

We provide the proof of Proposition 3 in Appendix A. Enabled by the monotonicity, we can reduce an instance of 13-large items to an instance of easier structure. This construction is described in the following.

Simplifying the Instance

Let I be a list of items larger than 1/3. Note that both the optimal and the Best Fit packing use only bins of configurations L, LM, MM, and possibly one M-bin. However, we can assume a simpler structure without substantial implications on the competitiveness of Best Fit.

Lemma 1

Let I be any list that can be packed optimally into OPT(I) LM-bins. If Best Fit has (asymptotic or absolute) approximation ratio α for I, then it has (asymptotic or absolute) approximation ratio α for any list of items larger than 1/3 as well.

Proof

Let I0 be a list of items larger than 1/3 and let a, b, c, and d1 be the number of bins in OPT(I0) with configurations L, LM, MM, and M, respectively (see Fig. 1a). In several steps, we eliminate L-, MM-, and M-bins from OPT(I0) while making the instance only harder for Best Fit.

Fig. 1.

Fig. 1

Construction from Lemma 1 to eliminate L-, MM-, and M-bins in the optimal packing

First, we obtain I1 from I0 by replacing items of size 1/2 by items of size 1/2-ε. By choosing ε>0 small enough, i.e., ε<min{δ+-1/2,1/2-δ-}, where δ+=min{xixi>1/2} and δ-=max{xixi<1/2}, it is ensured that Best Fit packs all items in the same bins as before the modification. Further, the modification does not decrease the number of bins in an optimal packing, so we have BF(I0)=BF(I1) and OPT(I0)=OPT(I1).

Now, we obtain I2 from I1 by increasing item sizes: We replace each of the a+d items packed in 1-bins in OPT(I1) by large items of size 1. Moreover, any 2-bin (MM or LM) in OPT(I1) contains at least one item smaller than 1/2. These items are enlarged such that they fill their respective bin completely. Therefore, OPT(I2) has a+d L-bins and b+c LM-bins (see Fig. 1b). We have OPT(I2)=OPT(I1) and, by Proposition 3, BF(I2)BF(I1).

Finally, we obtain I3 from I2 by deleting the a+d items of size 1. As size-1 items are packed separately in any feasible packing, OPT(I3)=OPT(I2)-(a+d) and BF(I3)=BF(I2)-(a+d).

Note that OPT(I3) contains only LM-bins (see Fig. 1c) and, by assumption, Best Fit has (asymptotic or absolute) approximation ratio α for such lists. Therefore, in general we have a factor α1 and an additive term β such that BF(I3)αOPT(I3)+β. It follows that

BF(I0)BF(I2)=BF(I3)+(a+d)αOPT(I3)+(a+d)+βαOPT(I0)+β,

which concludes the proof.

By Lemma 1, we can impose the following constraints on I without loss of generality.

Assumption. For the remainder of the section, we assume that the optimal packing of I has k=OPT(I) LM-bins. For i[k], let li and mi denote the large item and the medium item in the i-th bin, respectively. We call {li,mi} an LM-pair.

Good Order Pairs

If the adversary could control the order of items, he would send all medium items first, followed by all large items. This way, Best Fit opens k/2 MM-bins and k L-bins and therefore is 1.5-competitive. In a random permutation, we can identify structures with a positive impact on the Best Fit packing. This is formalized in the following random event.

Definition 2

Consider a fixed permutation πSn. We say that the LM-pair {li,mi} arrives in good order (or is a good order pair) if li arrives before mi in π.

Note that in the adversarial setting, no LM-pair arrives in good order, while in a random permutation, this holds for any LM-pair independently with probability 1/2. The next lemma is central for the proof of Theorem 1. It shows that the number of LM-pairs in good order bounds the number of LM-bins in the final Best Fit packing from below.

Lemma 2

Let πSn be any permutation and let X be the number of LM-pairs arriving in good order in Iπ. The packing BF(Iπ) has at least X LM-bins.

To prove Lemma 2, we model the Best Fit packing by the following bipartite graph: Let Gt=(MtLt,EtBFEtOPT), where Mt and Lt are the sets of medium and large items in Iπ(t), respectively. The sets of edges represent the LM-matchings in the Best Fit packing and in the optimal packing at time t, i.e.,

EtBF={{m,l}(Mt×Lt)m\,and\,l\,are\,packed\,into\,the\,same\,bin\,inBF(Iπ(t))}EtOPT={{mi,li}(Mt×Lt)i[k]}.

We distinguish OPT-edges in good and bad order, according to the corresponding LM-pair. Note that Gt is not necessarily connected and may contain parallel edges. We illustrate the graph representation by a small example.

Example 1

Let ε>0 be sufficiently small and define for i[4] large items li=1/2+iε and medium items mi=1/2-iε. Consider the list Iπ=(l2,l1,m3,m4,l4,m1,m2,l3). Figure 2a, b show the Best Fit packing and the corresponding graph G7 before arrival of the last item. Note that Iπ has two good order pairs ({l1,m1} and {l2,m2}) and, according to Lemma 2, the packing has two LM-bins.

Fig. 2.

Fig. 2

Visualization of Example 1. In Fig. 2, BF-edges are solid, while OPT-edges are thin. An asterisk indicates an OPT-edge in good order

The proof of Lemma 2 essentially boils down to the following claim:

Claim 1

In each round t and in each connected component C of Gt, the number of BF-edges in C is at least the number of OPT-edges in good order in C.

We first show how Lemma 2 follows from Claim 1. Then, we work towards the proof of Claim 1.

Proof of Lemma 2

Claim 1 implies that in Gn, the total number of BF-edges (summed over all connected components) is at least X. Therefore, the packing has at least X LM-bins and thus not less than the number of good order pairs X.

Before proving Claim 1, we show the following property of Gt.

Claim 2

Consider the graph Gt for some t[n]. Let Q=(bw,aw-1,bw-1,,,a1,b1) with w1 be a maximal alternating path such that {aj,bj} is an OPT-edge in good order and {aj,bj+1} is a BF-edge for any j[w-1] (i.e., a-items and b-items represent medium and large items, respectively). It holds that bwb1.

Proof

We show the claim by induction on w. Note that the items’ indices only reflect the position along the path, not the arrival order. For w=1, we have Q=(bw)=(b1) and thus, the claim holds trivially.

Now, fix w2 and suppose that the claim holds for all paths Q with ww-1. We next prove bwb1. Let tt be the arrival time of the a-item ad that arrived latest among all a-items in Q. We consider the graph Gt-1, i.e., the graph immediately before arrival of ad and its incident edges. Note that in Gt-1, all items ai with i[w-1]\{d} and bi with i[w-1] are visible. Let Q=(bw,,ad+1,bd+1) and Q=(bd,,a1,b1) be the connected components of bw and b1 in Gt-1. As Q and Q are maximal alternating paths shorter than Q, we obtain from the induction hypothesis bwbd+1 and bdb1.

Note that bd+1 and b1 were visible and packed into L-bins on arrival of ad. Further, ad and b1 would fit together, as ad+b1ad+bd1. However, Best Fit packed ad with bd+1, implying bd+1b1. Combining the inequalities yields bwbd+1b1, concluding the proof.

Now, we are able to prove the remaining technical claim.

Proof of Claim 1

Note that the number of OPT-edges in good order can only increase on arrival of a medium item mi where {mi,li} is an LM-pair in good order. Therefore, it is sufficient to verify Claim 1 in rounds t1<<tj such that in round ti, item mi arrives and li arrived previously.

Induction base. In round t1, there is one OPT-edge {m1,l1} in good order. We need to show that there exists at least one BF-edge in Gt1, or, alternatively, at least one LM-bin in the packing. If the bin of l1 contains a medium item different from m1, we identified one LM-bin. Otherwise, Best Fit packs m1 together with l1 or some other large item, again creating an LM-bin.

Induction hypothesis. Fix i2 and assume that Claim 1 holds up to round ti-1.

Induction step. We only consider the connected component of mi, as by the induction hypothesis, the claim holds for all remaining connected components. If mi is packed into an LM-bin, the number of BF-edges increases by one and the claim holds for round ti. Therefore, assume that mi is packed by Best Fit in an M- or MM-bin. This means that in Gti, vertex mi is incident to an OPT-edge in good order, but not incident to any BF-edge. Let P=(mi,li,,v) be the maximal path starting from mi alternating between OPT-edges and BF-edges.

Case 1: v is a medium item. For illustration, consider Fig. 2 with mi=m2 and v=m3. Since P begins with an OPT-edge and ends with a BF-edge, the number of BF-edges in P equals the number of OPT-edges in P. The latter number is clearly at least the number of OPT-edges in good order in P.

Case 2: v is a large item. For illustration, consider Fig. 2 with mi=m1 and v=l4. We consider two cases. If P contains at least one OPT-edge which is not in good order, the claim follows by the same argument as in Case 1.

Now, suppose that all OPT-edges in P are in good order. Let P be the path obtained from P by removing the item mi. As P satisfies the premises of Claim 2, we obtain liv. This implies that mi and v would fit together, as mi+vmi+li1. However, mi is packed in an M- or MM-bin by assumption, although v is a feasible option on arrival of mi. As this contradicts the Best Fit rule, we conclude that case 2 cannot happen.

Final Proofs

Finally, we prove the main result of this section.

Proof of Theorem 1

Let X be the number of good order pairs in Iσ and let Y be the number of LM-bins in the packing BF(Iσ). We have YX by Lemma 2. For the remaining large and medium items, Best Fit uses (k-Y) L-bins and (k-Y)/2 MM-bins (including possibly one M-bin), respectively. Therefore,

BF(Iσ)=Y+(k-Y)+k-Y2k+k-X2=3k2-X2+ξ(X)2, 1

where ξ(X)=(k-X)mod2. Using linearity and monotonicity of expectation, we obtain

E[BF(Iσ)]3k2-E[X]2+Pr[ξ(X)=1]2. 2

Since σ is uniformly distributed on Sn, each LM-pair arrives in good order with probability 1/2. Therefore, E[X]=k/2 and Pr[ξ(X)=1]=1/2. Hence,

E[BF(Iσ)]3k2-k/22+1/22=5k4+14=54OPT(I)+14, 3

where we used k=OPT(I). This concludes the proof.

To obtain the upper bound of 21/16 on the absolute random order ratio (Proposition 1), we analyze a few special cases more carefully.

Proof of Proposition 1

For k4, the claim follows immediately from Equation (3):

E[BF(Iπ)]OPT(I)=(5k)/4+1/4k=54+14k2116.

Since Best Fit is clearly optimal for k=1, it remains to verify the cases k{2,3}.

k=2
It is easily verified that there are 16 out of 4!=24 permutations where Best Fit is optimal and that it opens at most 3 bins otherwise. Therefore,
E[BF(Iσ)]=14!·16OPT(I)+8·32OPT(I)=76OPT(I)<2116OPT(I).
k=3
When k is odd, there must be at least one LM-bin in the Best Fit packing: Suppose for contradiction that all M-items are packed in MM- or M-bins. As k is odd, there must be an item mi packed in an M-bin. If li arrives before mi, item li is packed in an L-bin, as there is no LM-bin. Therefore, Best Fit packs mi with li or some other L-item instead of opening a new bin. If li arrives after mi, Best Fit packs li with mi or some other M-item. We have a contradiction in both cases. Therefore, for k=3 we have at least one LM-bin, even if no LM-pair arrives in good order. Consider the proof of Theorem 1. Instead of YX, we can use the stronger bound YX with X:=max{1,X} on the number of LM-bins. The new random variable satisfies E[X]=k/2+1/2k and Pr[ξ(X)=1]=1/2-1/2k. Adapting Equations (1) and (2) appropriately, we obtain
E[BF(Iσ)]OPT(I)=1k·3k2-k/2+1/2k2+1/2-1/2k2=54+14k-1k2k=3124<2116.

Lower Bounds

In this section, we present the improved lower bound on RRBF (Theorem 2) and the first lower bound on the absolute random order ratio RRBF (Theorem 3).

Asymptotic Random Order Ratio

Another model of probabilistic analysis is the i.i.d.-model, where the input of the algorithm is a sequence of independent and identically distributed (i.i.d.) random variables. Here, the performance measure of algorithm A is E[A(In(F))]/E[OPT(In(F))], where In(F):=(X1,,Xn) is a list of n random variables drawn i.i.d. according to F. This model is in general weaker than the random order model, which is why lower bounds in the random order model can be obtained from the i.i.d.-model. This is formalized in the following lemma.

Lemma 3

Consider any online bin packing algorithm A. Let F be a discrete distribution and In(F)=(X1,,Xn) be a list of i.i.d. samples. For n, there exists a list I of n items such that

E[A(Iσ)]OPT(I)E[A(In(F))]E[OPT(In(F))].

Moreover, if there exists a constant c>0 such that Xic for all i[n], we have OPT(I)cn.

This technique has already been used in [28] to establish the lower bound of 1.08, however, without a formal proof. Apparently, the only published proofs of this connection address bin covering [6, 13]. We provide a constructive proof of Lemma 3 in Appendix C for completeness. The improved lower bound from Theorem 2 now follows by combining Lemma 3 with the next lemma.

Lemma 4

There exists a discrete distribution F such that for n, we have E[A(In(F))]>1110E[OPT(In(F))] and each sample Xi satisfies Xi1/4.

Proof

Let F be the discrete distribution which gives an item of size 1/4 with probability p and an item of size 1/3 with probability q:=1-p. First, we analyze the optimal packing. Let N4 and N3 be the number of items with size 1/4 and 1/3 in In(F), respectively. We have

E[OPT(In(F))]EN44+N33+2=np4+nq3+2=n13-p12+2n.

Now, we analyze the expected behavior of Best Fit for In(F). As the only possible item sizes are 1/4 and 1/3, we can consider each bin of load more than 3/4 as closed. Moreover, the number of possible loads for open bins is small and Best Fit maintains at most two open bins at any time. Therefore, we can model the Best Fit packing by a Markov chain as follows. Let the nine states A,B,,I be defined as in Fig. 3b. The corresponding transition diagram is depicted in Fig. 3. This Markov chain converges to the stationary distribution

ω=(ωA,,ωI)=1λ1,p,q+pqϑ,p2,2pq+p2qϑ,q2+2pq2ϑ,ϑ,qϑ,q2ϑ,

where we defined ϑ:=p31-q3 and λ:=ϑq3-q2+ϑ+3. A formal proof of this fact can be found in Appendix C.2.

Fig. 3.

Fig. 3

Markov chain from Lemma 4. Bold arcs in Figure 3a indicate transitions where Best Fit opens a new bin

Let VS(t) denote the number of visits to state S{A,,I} up to time t. By a basic result from the theory of ergodic Markov chains (see [30, Sect. 4.7]), it holds that limt1t·VS(t)=ωS. In other words, the proportion of time spent in state S approaches its probability ωS in the stationary distribution.

This fact can be used to bound the total number of opened bins over time. Note that Best Fit opens a new bin on the transitions AB, AC, and GH (see Fig. 3a).

Hence, E[BF(In(F))]=VA(n)+qVG(n). Setting p=0.60, we obtain finally

limnE[BF(In(F))]E[OPT(In(F))]limnnωA+nqωGn13-p12+2n=1+qϑλ·13-p12>1110. 4

Absolute Random Order Ratio

Theorem 3 follows from the following lemma.

Lemma 5

There exists a list I such that E[BF(Iσ)]=1310OPT(I).

Proof

Let ε>0 be sufficiently small and let I:=(a1,a2,b1,b2,c) where

a1=a2=13+4ε,b1=b2=13+16ε,c=13-8ε.

An optimal packing of I has two bins {a1,a2,c} and {b1,b2}, thus OPT(I)=2. Subsequently, we argue that Best Fit needs two or three bins depending on the order of arrival.

Let E be the event that exactly one b-item arrives within the first two rounds. After the second item, the first bin is closed, as its load is at least 13+16ε+13-8ε=23+8ε. Among the remaining three items, there is a b-item of size 13+16ε and at least one a-item of size 13+4ε. This implies that a third bin needs to be opened for the last item. As there are exactly 2·3·2!·3!=72 permutations where E happens, we have Pr[E]=725!=35.

On the other side, Best Fit needs only two bins if one of the events F and G, defined in the following, happen. Let F be the event that both b-items arrive in the first two rounds. Then, the remaining three items fit into one additional bin. Moreover, let G be the event that the set of the first two items is a subset of {a1,a2,c}. Then, the first bin has load at least 23-4ε, thus no b-item can be packed there. Again, this ensures a packing into two bins.

By counting permutations, we obtain Pr[F]=2!·3!5!=110 and Pr[G]=3·2!·3!5!=310.

As the events E, F, and G partition the probability space, we obtain

E[BF(Iσ)]OPT(I)=Pr[E]·3+Pr[F]+Pr[G]·22=35·3+110+310·22=1310.

The construction from the above proof is used in [25] to prove that Best Fit is 1.5-competitive under adversarial arrival order if all item sizes are close to 1/3. Interestingly, it gives a strong lower bound on the absolute random order ratio as well.

Appendix A: Monotonicity

Proposition 3 follows by applying the following lemma iteratively. A technically similar proof appeared in [37], where Shor showed that the MBF algorithm from [37] is monotone under removal of items.

Lemma 6

Let I=(x1,,xn) be any list of items larger than 1/3. Let I=(x1,,xn) with xi>xi for a single i[n] and xj=xj for all ji. We have BF(I)BF(I).

Proof

All bins in any packing of I or I contain at most two items. We call two 1-bins of BF(I) and BF(I) pairwise-identical if they contain items of the same size. Moreover, we call any two 2-bins of BF(I) and BF(I) pairwise-closed, as neither of the two bins can receive a further item.

For ease of notation, let It=I(t) and It=I(t). We show that at any time t, the packings BF(It) and BF(It) are related in one of three ways (see Fig. 4).

(1)

All bins are pairwise-identical or pairwise-closed.

(2)

All bins are pairwise-identical or pairwise-closed, except for two 1-bins B={b} and B={b} in BF(It) and BF(It), respectively, where b<b.

(3)

All bins are pairwise-identical or pairwise-closed, except for a 2-bin C={c1,c2} in BF(It) which does not exist in BF(It), and two 1-bins B1={b1}, B2={b2} in BF(It) which do not exist in BF(It).

Fig. 4.

Fig. 4

Relating the packings BF(It) and BF(It) in the proof of Lemma 6

Note in all three cases, BF(It)BF(It). As this property is maintained until t=n, it implies the lemma. Subsequently, we show the claim inductively for each round.

Before round i, both lists contain items of identical size and thus, the packings are clearly related by (1). In round i, both packings deviate possibly, since xi<xi. Here, three cases can occur.

  1. Both items go into new bins. Then, (2) holds.

  2. Both items go into identical 1-bins. Then, (1) holds.

  3. Best-fit packs xi into an existing 1-bin B1 in BF(Ii), while it opens a new bin B2 for xi in BF(Ii). Before packing xi into B1, there was a 1-bin B1 in BF(Ii) such that B1 and B1 were pairwise identical. By packing xi, bin B1 becomes a 2-bin. Therefore, (3) holds with C=B1 as 2-bin and B1, B2 as 1-bins.

Note that these three cases are exhaustive: The algorithm packs either both items into new bins (a), both items into existing 1-bins (b), or xi into an existing 1-bin and xi into a new bin (c). In case (b), the 1-bins must be pairwise-identical as this holds for all bins up to time i. The inverse situation of (c) cannot occur since xi<xi.

For the induction step, we consider round ti+1 and suppose that BF(It-1) and BF(It-1) are related either by (1), (2), or (3). Note that in round t, in both lists the current item has equal size xt=xt again.

Case 1 If (1) holds at time t-1, then (1) is maintained in round t.

Case 2 Suppose (2) holds at time t-1. If xt is packed in BF(It) into an existing 1-bin, we know that Best Fit does not open a new bin for xt in BF(It) as well. Depending on the chosen bins, either (2) is maintained, or (1) holds.

Now, assume that xt is packed into a new bin in BF(It). Either, a new bin is opened in BF(It) as well, or BF(It) packs xt into bin B. In the first case, (2) is maintained, in the second case (3) holds. Note that BF(It) cannot pack xt into any existing 1-bin other than B, as otherwise BF(It) would pack xt into the corresponding identical bin instead of opening a new bin.

Case 3 Suppose (3) holds at time t-1. If Best Fit packs xt into an existing 1-bin, we know that it packs xt into an existing 1-bin as well. Hence, in both packings a 1-bin becomes a 2-bin and (3) is maintained.

If Best Fit packs xt into a new bin, we know that none of the existing 1-bins in BF(It), except for possibly B1 or B2, are suitable for xt. Therefore, Best Fit either opens a new bin to pack xt as well (then, (3) is maintained), or it packs xt into B{B1,B2}. Suppose that B=B1. Then, all bins in the packings are pairwise-identical or closed, except for B2={b2} in BF(It) and the new bin {xt} in BF(It). Hence, (2) holds if b2>xt. To see this, observe that b1+b2>1 since b1 and b2 have been packed into separate bins previously. Moreover, since Best Fit packed xt with b1, we have xt+b11. Therefore, b2>1-b11-(1-xt)=xt. The case B=B2 is analogous.

Appendix B: Lower Bound for 13-Large Items

Here, we prove the existence of a list I with 13-large items and E[BF(Iσ)]>65OPT(I).

Proof of Proposition 2

We construct a list of k=3 LM-pairs. For sufficiently small ε>0 and i[k] define li=12+iε and mi=12-iε. This way, l1<l2<l3 and m1>m2>m3. Clearly, OPT(I)=3. We can show that Best Fit uses 4 instead of 3 bins in at least 440 permutations. Therefore,

E[BF(Iσ)]OPT(I)16!·440·4+(6!-440)·33=6554>65.

We call the event that Best Fit packs two items different from {li,mi} into the same bin a non-optimal match. This event occurs if Best Fit packs either li and mj for 1i<j3, or two medium items mi and mj together into the same bin. It is easy to see that any packing with a non-optimal match needs at least 4 bins. However, Best Fit never uses more than 4 bins.

In the following, we partition the set of all permutations with a non-optimal match according to the first time where this happens. Note that the first possibility is after arrival of the second item. Moreover, if no non-optimal match happened before arrival of the fifth item, the fifth and sixth item build an LM-pair and therefore, no non-optimal match occurs at all.

Let a1,,a4 denote the first four items in a fixed permutation.

  • Case1:First non-optimal matchafter thesecond item

We have either a1,a2{li,mj} with i<j, or a1,a2{mi,mj} with ij. In both cases, we can choose the set {a1,a2} in three ways, can arrange a1 and a2 in 2! ways and the remaining items in 4! ways. The total number of such permutations is

2·3·2!·4!=288.
  • Case2:First non-optimal matchafter thethird item

Before a3 arrives and is matched non-optimally, a1 and a2 need to be packed into separate bins and a3 is matched non-optimally. This happens in the following cases.

  • Case 2a

a1,a2{l1,l2} and a3=m3.

  • Case 2b

a1,a2{l1,l3} and a3=m2.

  • Case 2c

a1,a2{l2,m1} and a3=m3.

  • Case 2d

a1,a2{l3,m1} and a3=m2.

  • Case 2e

a1,a2{m2,l3} and a3{m1,l1}. Note that for each case 2a-2d, we have 2!·3! permutations, while in case 2e, in addition, we can choose a3 in one of two ways. Therefore, we have a total number of permutations for case 2 of

4·(2!·3!)+2!·2·3!=48+24=72.
  • Case3:First non-optimal matchafter thefourth item

Here, we must have an optimal match either after the second item (Case 3a) or after the third item (Case 3b).

  • Case 3a

We have a1,a2{li,mi} for some i[3]. Now, {a3,a4} can contain either the remaining two medium items, or the larger of the remaining two large items and the smaller of the remaining two medium items. The fifth and sixth item can be arranged in 2! ways. The total number of such permutations is

3·2!·2·2!·2!=48.
  • Case 3b

We have an optimal match after the third item followed by a non-optimal match after the fourth item. This happens in the following cases.

  • a1,a2{l3,m1}, a3=m3, a4=m2.

  • a1,a2{l1,l3}, a3=m3, a4=m2.

  • a1,a2{l2,m1}, a3{m2,l1}, a4=m3.

  • a1,a2{l1,l2}, a3{m1,m2}, a4=m3.

  • a1,a2{l3,m2}, a3=m3, a4={m1,l1}.

In each of the first two cases, we get 2!·2! permutations. In each of the remaining three cases, we have 2!·2·2! permutations, since we can choose one additional item among two elements. The total number of permutations in case 3b is thus

2·(2!·2!)+3·(2!·2·2!)=8+24=32.

Appendix C: Lower Bound on the Asymptotic Random Order Ratio

Appendix C.1: Proof of Lemma 3

Let I={IPr[In(F)=I]>0} be the set of possible outcomes of In(F). We say that two lists I1,I2I are similar (I1I2) if there is a permutation πSn such that I1=I2π. Note that defines an equivalence relation on I. Let H be a complete set of representatives of . This way, I=HH{HππSn}.

We will use the following two technical claims which we will prove later.

Claim 3

Let ALG{A,OPT}. For each HH, there exists λH>0 such that

E[ALG(In(F))]=HHλH·πSnALG(Hπ).
Claim 4

For any two nonnegative sequences (ai)i[m] and (bi)i[m], we have

j=1majj=1mbjmaxj[m]ajbj.
Proof of Lemma 3

Using Claims 3 and 4, Lemma 3 follows from the following reasoning. Let

I=argmaxHHλH·πSnA(Hπ)λH·πSnOPT(Hπ).

Now, it holds that

E[A(In(F))]E[OPT(In(F))]=HHλH·πSnA(Hπ)HHλH·πSnOPT(Hπ)Claim3λI·πSnA(Iπ)λI·πSnOPT(Iπ)Claim 4=1n!·πSnA(Iπ)1n!·πSnOPT(Iπ)=E[A(Iσ)]OPT(I).

The second claim follows immediately from OPT(I)i=1nXicn.

It remains to show Claims 3 and 4.

Proof of Claim 3

We first show that all lists of the same equivalence class have the same probability to be an outcome of In(F)=(X1,,Xn).

Let I1=(a1,,an) and I2=(b1,,bn) with I1I2. Note that both lists contain the same multiset of items. Let m be the number of different item sizes and let zi be the multiplicity of size yi for i[m] in I1 (and thus in I2). As the random variables Xi are drawn i.i.d., we have

Pr[In(F)=I1]=i=1nPr[Xi=ai]=j=1mPr[X1=yj]zj=i=1nPr[Xi=bi]=Pr[In(F)=I2].

Now, consider any equivalence class [H] for HH. All lists I[H] have the same probability Pr[In(F)=I]=:λH to be an outcome of In(F). Further, we have [H]={HππSn} and thus

E[ALG(In(F))]=IIALG(I)·Pr[In(F)=I]=HHI[H]ALG(I)·Pr[In(F)=I]=HHλH·I[H]ALG(I)=HHλH·πSnALG(Hπ).

Claim 4 is easy to see:

Proof of Claim 4

Define M:=maxj[m](aj/bj). We have

j=1maj=j=1majbj·bjj=1mMbj=M·j=1mbj.

Appendix C.2: Markov Chain from Fig. 3

Lemma 7

The Markov chain from Fig3converges to the stationary distribution ω=(ωA,,ωI) with

ω=1λ1,p,q+pqϑ,p2,2pq+p2qϑ,q2+2pq2ϑ,ϑ,qϑ,q2ϑ,

where ϑ=p31-q3 and λ=ϑq3-q2+ϑ+3.

Proof

As the Markov chain from Fig. 3 is irreducible and aperiodic, it converges to a unique stationary distribution. Let ω=(ωA,,ωI) be this distribution, then we have the following system of equations.

ωA=ωE+ωF+pωG+qωD Q1
ωB=pωA Q2
ωC=qωA+pωH Q3
ωD=pωB Q4
ωE=qωB+pωC Q5
ωF=qωC+pωI Q6
ωG=pωD+qωI Q7
ωH=qωG Q8
ωI=qωH Q9
1=ωA+ωB+ωC+ωD+ωE+ωF+ωG+ωH+ωI. Q10

We define ϑ:=p31-q3 and λ:=ϑq3-q2+ϑ+3 and claim that

ω=1λ1,p,q+pqϑ,p2,2pq+p2qϑ,q2+2pq2ϑ,ϑ,qϑ,q2ϑ

satisfies (Q1) to (Q10). In fact, the validity of equations (Q2) to (Q9) can be seen immediately. To verify (Q1) and (Q10), we first observe that for any i0 and 1j3, it holds that

piqj=pi-pi+1j=1pi-2pi+1+pi+2j=2pi-3pi+1+3pi+2-pi+3j=3. 5

Now, we prove (Q1). We have

λ·ωE+ωF+pωG+qωD=(2pq+p2qϑ)+(q2+2pq2ϑ)+pϑ+qp2=ϑ·p2q+2pq2+p+2pq+q2+p2q=eq.5ϑ·p2-p3+2·(p-2p2+p3)+p+2·(p-p2)+(1-2p+p2)+(p2-p3)=ϑ(3p-3p2+p3)+1-p3=1=λ·ωA,

which implies (Q1). Equation (Q10) follows from the following calculation.

λ·ωA+ωB+ωC+ωD+ωE+ωF+ωG+ωH+ωI=1+p+(q+pqϑ)+p2+(2pq+p2qϑ)+(q2+2pq2ϑ)+ϑ+qϑ+q2ϑ=qϑ·p+p2+2pq+1+q+(1+p+q+p2+2pq+q2)+ϑ=eq.5qϑ·p+p2+2·(p-p2)+1+1-p+(1+p+1-p+p2+2·(p-p2)+1-2p+p2)+ϑ=qϑ·3-q2+3+ϑ=λ.

Funding

Open Access funding enabled and organized by Projekt DEAL.

Declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Footnotes

Work supported by the European Research Council, Grant Agreement No. 691672. A preliminary version of this paper received the best paper award at 45th International Symposium on Mathematical Foundations of Computer Science (MFCS 2020) and was published in the conference proceedings [2].

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Susanne Albers, Email: albers@in.tum.de.

Arindam Khan, Email: arindamkhan@iisc.ac.in.

Leon Ladewig, Email: leon.ladewig@tum.de.

References

  • 1.Al-Herz, A., Pothen, A.: A 2/3-approximation algorithm for vertex-weighted matching. In: Discrete Applied Mathematics (2019). ((in press))
  • 2.Albers, S., Khan, A., Ladewig, L.: Best fit bin packing with random order revisited. In: 45th International Symposium on Mathematical Foundations of Computer Science (MFCS), pp. 7:1–7:15 (2020)
  • 3.Balogh, J., Békési, J., Dósa, G., Epstein, L., Levin, A.: A new and improved algorithm for online bin packing. In: Proceedings of the 26th Annual European Symposium on Algorithms (ESA), LIPIcs, vol. 112, pp. 5:1–5:14 (2018)
  • 4.Balogh, J., Békési, J., Dósa, G., Epstein, L., Levin, A.: A new lower bound for classic online bin packing. In: Approximation and Online Algorithms—17th International Workshop, (WAOA), Lecture Notes in Computer Science, vol. 11926, pp. 18–28. Springer (2019)
  • 5.Boyar J, Dósa G, Epstein L. On the absolute approximation ratio for first fit and related results. Discret. Appl. Math. 2012;160(13–14):1914–1923. doi: 10.1016/j.dam.2012.04.012. [DOI] [Google Scholar]
  • 6.Christ MG, Favrholdt LM, Larsen KS. Online bin covering: expectations vs. guarantees. Theor. Comput. Sci. 2014;556:71–84. doi: 10.1016/j.tcs.2014.06.029. [DOI] [Google Scholar]
  • 7.Christensen HI, Khan A, Pokutta S, Tetali P. Approximation and online algorithms for multidimensional bin packing: a survey. Comput. Sci. Rev. 2017;24:63–79. doi: 10.1016/j.cosrev.2016.12.001. [DOI] [Google Scholar]
  • 8.Coffman Jr., E.G., Csirik, J., Galambos, G., Martello, S., Vigo, D.: Bin packing approximation algorithms: survey and classification. In: Handbook of Combinatorial Optimization, pp. 455–531. Springer, New York (2013)
  • 9.Coffman EG, Jr, Csirik J, Rónyai L, Zsbán A. Random-order bin packing. Discret. Appl. Math. 2008;156(14):2810–2816. doi: 10.1016/j.dam.2007.11.004. [DOI] [Google Scholar]
  • 10.Coffman EG, Jr, Lueker GS. Probabilistic Analysis of Packing and Partitioning Algorithms. Wiley-Interscience Series in Discrete Mathematics and Optimization. Hoboken: Wiley; 1991. [Google Scholar]
  • 11.Dósa, G., Sgall, J.: First fit bin packing: a tight analysis. In: 30th International Symposium on Theoretical Aspects of Computer Science (STACS), LIPIcs, vol. 20, pp. 538–549 (2013)
  • 12.Dósa, G., Sgall, J.: Optimal analysis of best fit bin packing. In: Proceedings of the 41st International Colloquium on Automata, Languages, and Programming (ICALP), pp. 429–441 (2014)
  • 13.Fischer, C., Röglin, H.: Probabilistic analysis of the dual next-fit algorithm for bin covering. In: LATIN 2016: Theoretical Informatics—12th Latin American Symposium, pp. 469–482 (2016)
  • 14.Fischer, C., Röglin, H.: Probabilistic analysis of online (class-constrained) bin packing and bin covering. In: LATIN 2018: Theoretical Informatics—13th Latin American Symposium, Lecture Notes in Computer Science, vol. 10807, pp. 461–474. Springer (2018)
  • 15.Gamlath, B., Kapralov, M., Maggiori, A., Svensson, O., Wajc, D.: Online matching with general arrivals. In: 60th IEEE Annual Symposium on Foundations of Computer Science (FOCS), pp. 26–37 (2019)
  • 16.Garey, M.R., Graham, R.L., Ullman, J.D.: Worst-case analysis of memory allocation algorithms. In: Proceedings of the 4th Annual ACM Symposium on Theory of Computing (STOC), pp. 143–150 (1972)
  • 17.Garey MR, Johnson DS. Strong NP-completeness results: motivation, examples, and implications. J. ACM. 1978;25(3):499–508. doi: 10.1145/322077.322090. [DOI] [Google Scholar]
  • 18.Graham RL. Bounds on multiprocessing timing anomalies. SIAM J. Appl. Math. 1969;17(2):416–429. doi: 10.1137/0117039. [DOI] [Google Scholar]
  • 19.Gupta, A., Singla, S.: Random-order models. CoRR, arXiv:abs/2002.12159 (2020)
  • 20.Hoberg, R., Rothvoss, T.: A logarithmic additive integrality gap for bin packing. In: Proceedings of the 28th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pp. 2616–2625 (2017)
  • 21.Huang, Z., Kang, N., Tang, Z.G., Wu, X., Zhang, Y., Zhu, X.: How to match when all vertices arrive online. In: Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing (STOC), pp. 17–29 (2018)
  • 22.Huang, Z., Peng, B., Tang, Z.G., Tao, R., Wu, X., Zhang, Y.: Tight competitive ratios of classic matching algorithms in the fully online model. In: Proceedings of the 30th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pp. 2875–2886 (2019)
  • 23.Huang Z, Tang ZG, Wu X, Zhang Y. Online vertex-weighted bipartite matching: beating 1-1/e with random arrivals. ACM Trans. Algorithms. 2019;15(3):38:1–38:15. doi: 10.1145/3326169. [DOI] [Google Scholar]
  • 24.Johnson DS. Fast algorithms for bin packing. J. Comput. Syst. Sci. 1974;8(3):272–314. doi: 10.1016/S0022-0000(74)80026-7. [DOI] [Google Scholar]
  • 25.Johnson DS, Demers AJ, Ullman JD, Garey MR, Graham RL. Worst-case performance bounds for simple one-dimensional packing algorithms. SIAM J. Comput. 1974;3(4):299–325. doi: 10.1137/0203025. [DOI] [Google Scholar]
  • 26.Karmarkar, N., Karp, R.M.: An efficient approximation scheme for the one-dimensional bin-packing problem. In: Proceedings of the 23rd Annual Symposium on Foundations of Computer Science (FOCS), pp. 312–320 (1982)
  • 27.Karp, R.M., Vazirani, U.V., Vazirani, V.V.: An optimal algorithm for on-line bipartite matching. In: Ortiz, H. (ed.) Proceedings of the 22nd Annual ACM Symposium on Theory of Computing (STOC), pp. 352–358 (1990)
  • 28.Kenyon, C.: Best-fit bin-packing with random order. In: Proceedings of the 7th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pp. 359–364 (1996)
  • 29.Lee CC, Lee DT. A simple on-line bin-packing algorithm. J. ACM. 1985;32(3):562–572. doi: 10.1145/3828.3833. [DOI] [Google Scholar]
  • 30.Levin DA, Peres Y. Markov Chains and Mixing Times. Providence: American Mathematical Society; 2017. [Google Scholar]
  • 31.Mahdian, M., Yan, Q.: Online bipartite matching with random arrivals: an approach based on strongly factor-revealing LPs. In: Proceedings of the 43rd ACM Symposium on Theory of Computing (STOC), pp. 597–606 (2011)
  • 32.Mehta A. Online matching and ad allocation. Found. Trends Theor. Comput. Sci. 2013;8(4):265–368. doi: 10.1561/0400000057. [DOI] [Google Scholar]
  • 33.Murgolo FD. Anomalous behavior in bin packing algorithms. Discret. Appl. Math. 1988;21(3):229–243. doi: 10.1016/0166-218X(88)90069-8. [DOI] [Google Scholar]
  • 34.Ramanan PV. Average-case analysis of the smart next fit algorithm. Inf. Process. Lett. 1989;31(5):221–225. doi: 10.1016/0020-0190(89)90077-X. [DOI] [Google Scholar]
  • 35.Ramanan PV, Brown DJ, Lee CC, Lee DT. On-line bin packing in linear time. J. Algorithms. 1989;10(3):305–326. doi: 10.1016/0196-6774(89)90031-X. [DOI] [Google Scholar]
  • 36.Seiden SS. On the online bin packing problem. J. ACM. 2002;49(5):640–671. doi: 10.1145/585265.585269. [DOI] [Google Scholar]
  • 37.Shor PW. The average-case analysis of some on-line algorithms for bin packing. Combinatorica. 1986;6(2):179–200. doi: 10.1007/BF02579171. [DOI] [Google Scholar]
  • 38.Ullman J. The Performance of a Memory Allocation Algorithm. Princeton: Department of Electrical Engineering, Computer Science Laboratory, Princeton University; 1971. [Google Scholar]
  • 39.de la Vega WF, Lueker GS. Bin packing can be solved within 1+epsilon in linear time. Combinatorica. 1981;1(4):349–355. doi: 10.1007/BF02579456. [DOI] [Google Scholar]

Articles from Algorithmica are provided here courtesy of Springer

RESOURCES