Abstract
The problem of optimizing over random structures emerges in many areas of science and engineering, ranging from statistical physics to machine learning and artificial intelligence. For many such structures, finding optimal solutions by means of fast algorithms is not known and often is believed not to be possible. At the same time, the formal hardness of these problems in the form of the complexity-theoretic NP-hardness is lacking. A new approach for algorithmic intractability in random structures is described in this article, which is based on the topological disconnectivity property of the set of pairwise distances of near-optimal solutions, called the Overlap Gap Property. The article demonstrates how this property 1) emerges in most models known to exhibit an apparent algorithmic hardness; 2) is consistent with the hardness/tractability phase transition for many models analyzed to the day; and, importantly, 3) allows to mathematically rigorously rule out a large class of algorithms as potential contenders, specifically the algorithms that exhibit the input stability (insensitivity).
Keywords: algorithms and computation, phase transition, random structures, spin glasses
Optimization problems involving uncertainty emerge in many areas of science and engineering, including statistics, machine learning and artificial intelligence, computer science, physics, biology, management science, economics, and social sciences. The exact nature and the sources of uncertainty vary from field to field. The modern paradigm of Big Data brought forward optimization problems involving, in particular, many dimensions, thus creating a new metafield of ”high-dimensional probability” and ”high-dimensional statistics” (1–3). Two recent special semester programs at the Simons Institute for the Theory of Computing were devoted to this burgeoning topic,* among a multitude of other conferences and workshops. While many of the optimization problems involving randomness can be solved to optimality by fast, and in fact, sometimes borderline trivial, algorithms, other problems have resisted decades of attempts, and slowly, it has been accepted that these problems are likely nonamenable to fast algorithms, with polynomial time algorithms broadly considered to be the gold standards for what constitutes fast algorithms. The debate surrounding the actual algorithmic hardness of these problems, though, is by no means settled, unlike its ”worst-case” algorithmic complexity-theoretic counterparts, expressed in the form of the widely believed conjecture.
What is the ”right” algorithmic complexity theory explaining the persistent failure to find tractable algorithms for these problems? We discuss in this paper some existing theories and explain their shortcomings in light of what we know now about the state-of-the-art algorithms. We then propose an approach, largely inspired by the field of statistical physics, which offers a new topological/geometric theory of algorithmic hardness that is based on the disconnectivity of the overlaps of near-optimal solutions, dubbed the ”Overlap Gap Property” (OGP). The property can and has been rigorously verified for many concrete models and, importantly, can be used to mathematically rule out large classes of algorithms as potential contenders, specifically algorithms exhibiting a form of input stability. This includes both classical and quantum (Quantum Approximate Optimization Algorithm [QAOA]) algorithms (4, 5), which recently gained a lot of attention as one of the most realistic algorithms to be implementable on quantum computers (6). A widely studied random Number Partitioning problem will be used to illustrate both the OGP and the mechanics by which this property manifests itself as a barrier for tractable algorithms.
The Largest Clique of a Random Graph: The Most ”Embarrassing” Open Problem in Random Structures
Imagine a club with members, in which about 50% of the member pairs know each other personally, and the remaining 50% of the members do not. You want to find a largest clique in this club, namely, the largest group of members out of the members who all know each other. What is the typical size of such a clique? How easy is it to find one? This question can be modeled as the problem of finding a largest fully connected subgraph (formally called a ”clique”) in a random Erdös–Rényi graph , which is a graph on nodes, where every pair of nodes is connected with probability , independently for all pairs. The first question regarding the largest clique size is a textbook example of the so-called probabilistic method, a simple application of which tells us that is likely to be near with high degree of certainty as gets large (7). A totally different matter is the problem of actually finding such a clique, and this is where the embarrassment begins. Richard Karp, one of the founding fathers of the algorithmic complexity theory, observed in his 1976 paper (8) that a very simple algorithm, both in terms of the analysis and the implementation, finds a clique of roughly half-optimum size, namely, with about members, and challenged to do better. The problem is still open, and this is embarrassing for two reasons: 1) The best-known algorithm, namely, the one above, is also an extremely naive one. So it appears that the significant progress that the algorithmic community has achieved over the past decades in constructing ever more clever and powerful algorithms is totally helpless in improving upon this extremely simple and naive algorithm. 2) We don’t have a workable theory of algorithmic hardness, which rigorously explains why finding cliques of size half-optimum is easy, and improving on it does not seem possible within the realm of polynomial time algorithms. The classical paradigm and its variants are of no help here. A variant of this problem, called the Hidden Clique Problem, was introduced in a seminal paper by Jerrum (9), and it exhibits a similar algorithmic gap.
The largest clique problem turns out to be one of very many other problems exhibiting a similar phenomenon: Using nonconstructive analysis method, one shows that the optimal value of some optimization problem involving randomness is some value , the best known polynomial time algorithm achieves value , and there is a significant gap between the two: . A partial list (but one growing seemingly at the speed of the Moore’s Law) is the following: the randomly generated constraint satisfaction problem (also known as [aka] the random K-satisfiability [K-SAT] problem), the largest independent set in a random graph, proper coloring of a random graph, finding a ground state of a spin-glass model in statistical mechanics, discovering communities in a network (the so-called community detection problem), group testing, statistical learning of a mixture of Gaussian distributions, the sparse linear regression problem, the sparse covariance estimation, the graph matching problem, the spiked tensor problem, the Number Partitioning problem, and many other problems.
The Number Partitioning problem is motivated in particular by the statistical problem of designing a randomized controlled study with two groups possessing roughly ”equal” attributes. It is also a special case of the bin-packing problem and also has been widely studied in the statistical physics literature. The Number Partitioning problem will serve as one of the running examples in this article, so we introduce it now. Given items with weights , the goal is to split it into two groups so that the difference of total weights in two groups is the smallest possible. Namely, the goal is finding a subset {} such that is as small as possible. An NP-hard problem in the worst case (10), it is more tractable in the presence of randomness. Suppose the weights are generated independently, according to the standard Gaussian distribution . A rather straightforward application of the same probabilistic method shows that the optimum value is typically for large . Namely, the value of in our context is . An algorithm proposed by Karmarkar and Karp in 1982 (11) achieves the value of order , with value predicted (though not proven) to be (12). The multiplicative gap between the two is thus very significant (exponential): . No improvement of this result is known to date, and no algorithmic hardness result is known either.
In Search of the Right Algorithmic Complexity Theory
We now describe some existing approaches for understanding the algorithmic complexity and discuss their shortcomings in explaining the existential/algorithmic gap exhibited by the multitude of problems described above. Starting with the classical algorithmic complexity theory based on the algorithmic complexity classes such as , , etc., this theory is of no help, since these complexity classes are based on the worst-case assumptions on the problem description. For example, assuming the widely believed conjecture that , finding a largest clique in a graph with nodes within a multiplicative factor is not possible by polynomial time algorithms (13) for any constant if . This is a statement, however, about the algorithmic problem of finding large cliques in all graphs, and it is in sharp contrast with the factor achievable by polynomial time algorithms for random graphs , according to the discussion above.
A significant part of the algorithmic complexity theory does in fact deal with problems with random input, and here, impressive random-case to worst-case reductions are achievable for some of the problems. These problems enjoy a wide range applications in cryptography, where the average case-hardness property is paramount for designing secure cryptographic protocols. For example, if there exists a polynomial time algorithm for computing the permanent of a matrix with independent and identically distributed (i.i.d.) random inputs, then, rather surprisingly, one can use it to design a polynomial time algorithm for all matrices (14). The latter problem is known to be in the complexity class, which subsumes . A similar reduction exists (15) for the problem of computing the partition function of a Sherrington–Kirkpatrick model described below, thus implying that computing partition functions for spin-glass models is not possible by polynomial time algorithms unless . Another problem admitting average-case to worst-case reduction is the problem of finding a shortest vector in a lattice (16). The random to worst-case types of reduction described above would be ideal for our setting, as they would provide the most compelling evidence of hardness of these problems. For example, it would be ideal to show that finding a clique of size at least, say, in implies a polynomial time algorithm with the same approximation guarantee for all graphs. Unfortunately, such a result appears out of reach for the existing proof techniques.
Closer to the approach described in this paper, and the one that in fact has largely inspired the present line of work, is a range of theories based on the solution space geometry of the underlying optimization problem. This approach emerged in the statistical physics literature, specifically, the study of spin glasses, and more recently found its way to questions above and beyond statistical physics, in particular, questions arising in the context of studying neural networks, as in ref. 17. The general philosophy of this take on the algorithmic complexity is that when the problem appears to be algorithmically hard, this somehow should be reflected in the nontrivial geometry of optimal or near-optimal solutions. One of the earliest such approaches was formulated for the decision problems, such as random constraint satisfaction problems (aka the random K-SAT problem). It links the algorithmic hardness with the proximity to the satisfiability phase-transition threshold (18, 19) and the order (first vs. second) of the associated phase transition (20). To elaborate, we need to introduce the random K-SAT problem first. It is a Boolean constraint satisfaction problem involving variables defined as a conjunction of clauses , where each clause is a disjunction of exactly variables from or their negation. Thus, each is of the form . An example of such a formula with , and is, say, . A formula is called satisfiable if there exists an assignment of Boolean variables to values zero and one such that the value of the formula is one.
A random instance of a K-SAT problem is obtained by selecting variables into each clause uniformly at random, independently for all , and applying the negation operation with probability independently for all variables. The random K-SAT problem is viewed by statistical physicists as a problem exhibiting the so-called frustration property, which is of great interest to physics of spin glasses, and, hence, it has enjoyed a great deal of attention in the statistical physics literature (21, 22).
As it turns out, for each , there is a conjectured critical threshold for the satisfiability property, which was rigorously proven (23) for large enough . In particular, for every , the formula admits a satisfying assignment when , and for every , the formula does not admit a satisfying assignment when , both with overwhelming probability as . The sharpness of this transition was established rigorously earlier by general methods in ref. 24. The algorithmic counterparts, however, came short of achieving the value for every (more on this below). Even before the results in refs. 23 and 24, it was conjectured in the physics literature that perhaps the existence of the sharp phase-transition property itself is the culprit for the algorithmic hardness (19, 20). This was argued by studying the heuristic running times of finding the satisfying solutions or demonstrating nonexistence of ones for small values of and observing that the running time explodes near and subsides when is far from on either side. The values of are known thanks to the powerful machinery of the replica symmetry methods developed in seminal works of Parisi (21, 25). These methods give very precise predictions for the values of for every . For example, is ∼4.26 when . An in-depth analysis of the exponent of the running times was reported in ref. 19.
The theory postulating that the algorithmic hardness of the K-SAT problem is linked with the satisfiability phase transition, however, appears inconsistent with the later rigorous discoveries obtained specifically for large values of . In particular, while is known to be approximately for large values of , all of the known algorithms stall long before that, and specifically at (26). Furthermore, there is evidence that breaking this barrier might be extremely challenging. This was argued by proving that the model undergoes the so-called clustering phase transition near (27, 28) (more on this below), and also by ruling out various families of algorithms. These algorithms include sequential local algorithms and algorithms based on low-degree polynomials (using the OGP discussed in the next section) (29, 30), the Survey Propagation algorithm (31), and a variant of a random-search algorithm called WalkSAT (32). The Survey Propagation algorithm is a highly effective heuristics for finding satisfying assignments in random K-SAT and many other similar problems. It is particularly effective for low values of and beats many other existing approaches in terms of the running times and sizes of instances it is capable to handle (33) (see also ref. 34 for the perspective on the approach). It is entirely possible that this algorithm and even its more elementary version, the Belief Propagation guided decimation algorithm, solves the satisfiability problem for small values of . An evidence of this can be found in ref. 35, and rigorous analysis of this algorithm was conducted by Coja-Oghlan in ref. 36, which shows it to be effective when . However, the analysis by Hetterich in ref. 31 rules out the Survey Propagation algorithm for sufficiently large values of , beyond the threshold, which we recall is . Additionally, the theory of algorithmic hardness based on the existence of the satisfiability-type phase transition does not appear to explain the algorithmic hardness associated with optimization problems, such as the largest clique problem in .
The clustering property was rigorously established (27, 28) for random K-SAT for values of above , which is known to be close to for large . We specifically refer to this as weak clustering property in order to distinguish it from the strong clustering property, both of which we now define. (While the distinction between weak and strong clustering was discussed in many prior works, the choice of the terminology is entirely by the author.) The model exhibits the weak clustering property if there exists a subset of satisfying assignments, which contains all but exponentially small in fraction of the satisfying assignments and which can be partitioned into subsets (clusters) separated by order distances, such that within each cluster, one can move between any two assignments by changing, at most, constantly many bits. In other words, by declaring two assignments and connected if can be obtained from by flipping the values of, at most, variables, the set consists of several disconnected components separated by linear in distances. This is illustrated in SI Appendix, Fig. S1, where blue regions represent clusters in . The gray ”tunnel”-like sets depicted in the figure are part of the complement (exception) subset of the set of satisfying assignments, which may potentially connect (tunnel between) the clusters. Furthermore, the clusters are separated by exponentially large in cost barriers, meaning that any path connecting two solutions from different clusters (which then necessarily contains assignments violating at least some of the constraints), at some point, contains in it assignments, which, in fact, violate order constraint. For this reason, the clustering property is sometimes referred to as an “energetic” barrier.
As mentioned earlier, for large values of , the onset of weak clustering occurs near , and, in fact, the suggestion that this value indeed corresponds to the algorithmic threshold was partially motivated by this discovery of weak clustering property around this value. See also ref. 37 for the issues connecting the clustering property with the algorithmic hardness. Below , the bulk of the set of satisfying assignments constitutes one connected subset.
On the other hand, the strong clustering property is the property that all satisfying assignments can be partitioned into clusters like the ones above, however, with no exceptions. Namely, is the set of all satisfying assignments. This can be visualized from SI Appendix, Fig. S1, where the gray tunnels would be removed. It turns out that for large values of , the model does exhibit the strong clustering property as well, but the known lower bounds for it are of the order , as opposed to , for large . While not proven formally, it is expected though that the onset of the strong clustering property indeed occurs at values order , as opposed to , for large . Namely, the weak and strong clustering properties appear at different in scales.
The existence of the strong clustering property is established as an implication of the OGP (which is the subject of this paper), and this was the approach used in refs. 27 and 28. In contrast, the weak clustering property is established by using the so-called planting solution, which amounts to consider the solution-space geometry from the perspective of a uniformly at random selected satisfying assignment.
The necessity of distinguishing between the two modes of clustering described above is driven by algorithmic implications. Solutions generated by most algorithms typically are not generated uniformly at random and, thus, in principle, can easily fall into the exception set . Thus, the obstruction arguments based somehow on linear separations between the clusters and exponentially large energetic barriers might simply be of no algorithmic relevance; see the later OGP, the Clustering Property, and the Curious Case of the Perceptron Model section devoted to the relationship between clustering and the OGP.
As the notion of ”exception” size in the definition of weak clustering property hinges on the counting measure associated with the space of satisfying assignments (and the uniform measure was assumed in the discussion above), a potential improvement might stem from changing this counting measures and introducing solution-specific weights. Interestingly, for small values of , this can be done effectively delaying the onset of weak clustering, as was shown by Budzynski et al. in ref. 38. Unfortunately, though, for large , the gain is only in second-order terms, and up to those orders, the threshold for the onset of weak clustering based on biased counting measures is still , as shown by Budzynski and Semerjian in ref. 39. This, arguably, provides an even more robust evidence that this value marks a phase-transition point of fundamental nature.
More recently, algorithmic barriers were suggested to be linked with refined properties of the clustering phase transition, specifically, the so-called condensation-phase transition (40) and the emergence of so-called frozen variables. Algorithmic implications of both are discussed in ref. 41. Using again a powerful machinery of the replica symmetry breaking, providing nonrigorous, though highly believable, predictions, one identifies another phase transition satisfying for large . As the value passes through , the number of solution clusters covering the majority of satisfying assignment in the random K-SAT problem drops dramatically from the exponentially many to only constantly many ones, with the largest cluster containing a nontrivial fraction of all of the assignments (40) (SI Appendix, Fig. S1). At this stage, a long-range dependence between the spin magnetization (appropriately defined) emerges, and, furthermore, at this stage, the random overlaps (inner products) of pairs of assignments generated according to the associated Gibbs measure have a nontrivial limiting distribution, described by the so-called Parisi measure.
Even before the value exceeds the value for large , each cluster contains a nontrivial fraction of the frozen variables, which are variables always taking the same values within a given cluster of satisfying assignment. The authors in ref. 41 conjecture that the emergence of frozen variables is the primary culprit of the algorithmic hardness and construct a variant of the Survey Propagation algorithm called the Backtracking Survey Propagation algorithm, which is conjectured to be effective all the way up to . Numerical simulations demonstrate an excellent performance of this algorithm on randomly generated instances of K-SAT for small values of .
The conjecture linking the emergence of frozen variables with algorithmic hardness stumbles, however, upon similar challenges as the theory based on the existence of the phase transition. A rigorous evidence has been established that when is large, is of the order for some (explicitly computable) value , which does not depend on . In particular, is substantially above the value at which all known algorithms fail, and, furthermore, classes of algorithms, including the standard Survey Propagation algorithm, are ruled out, as discussed above. Of course, the nonexistence of evidence is not necessarily the evidence of nonexistence, and it is very much possible that in the future, successful polynomial time algorithms will be constructed in the regime . But there is another limitation of the theory of algorithmic hardness based on the emergence of frozen variables: Just like the notion of the satisfiability phase transition, the notion of frozen variables does not apply to problems of optimization type, such as the largest clique problem, or an optimization problem appearing in the context of spin-glass models—for example, the problem of finding a ground state of a spherical -spin-glass model. In this model, spins take values in a continuous range, and, thus, the notion of frozenness, which is exclusively discrete, is lost.
Topological Complexity Barriers: The OGP
We now introduce a new approach for predicting and proving algorithmic complexity for solving random constraint satisfaction problems and optimization problems on random structures, such as the ones introduced in sections The Largest Clique of a Random Graph: the Most “Embarrassing” Open Problem in Random Structures and In Search of the Right Algorithmic Complexity Theory. The approach bears a lot of similarity with the predictions based on the clustering property described in the latter of the two aforementioned sections, but has important and nuanced differences, which allow us to rule out large classes of algorithms, something that did not seem possible before. An important special case of such algorithms is stable algorithms, namely, algorithms exhibiting low sensitivity to the data input. A special, yet very broad and powerful, class of such algorithms is algorithms based on low-degree polynomials. A fairly recent stream of research (42–44) puts forward an evidence that, in fact, algorithms based on low-degree polynomials might be the most powerful class of polynomial time algorithms for optimization problems on random structures, such as the problems discussed in this paper.
A Generic Formulation of the Optimization Problem.
To set the stage, we consider a generic optimization problem . Here, encodes the objective function (cost) to be minimized. The solutions lie in some ambient solution space , which is typically very high-dimensional, often discrete, with encoding its dimension. The space is equipped with some metric (distance) defined for each pair of solutions . For the max-clique problem, we can take . For the Number Partitioning problem, we can take . is intended to denote the associated randomness of the problem. So, for example, for the max-clique problem, encodes the random graph , and encodes a set of nodes constituting a clique, with if node is in the clique and otherwise. We denote by an instance generated according to the probability law of . In the present context, is any graph, for which the likelihood of generating is . Finally, is the number of ones in the vector . Now, if does not actually encode a clique (that is, for at least one of the edges of the graph, we have ), we can easily augment the setup by declaring for such ”nonclique” encoding vectors. For the Number Partitioning problem, is just the probability distribution associated with an -vector of independent standard Gaussians. For each and each instance of , the value of the associated partition is . Here, denotes an inner product between vectors and . Thus, for the largest clique problem, we have that the random variable is approximately with high degree of likelihood. For the Number Partitioning problem, we have that is approximately with high degree of likelihood as well.
The OGP and its Variants.
We now introduce the OGP and its several variants. The term was introduced in ref. 45, but the property itself was discovered by Achlioptas and Ricci-Tersenghi (27) and Mezard, Mora, and Zecchina (28). The definition of the OGP pertains to a particular instance of the randomness . We say that the optimization problem exhibits the OGP with values if for every two solutions , which are -optimal in the additive sense, namely, satisfy , it is the case that either or . Intuitively, the definition says that every two solutions that are ”close” (within an additive error ) to optimality are either close (at most distance ) to each other or ”far” (at least distance ) from each other, thus exhibiting a fundamental topological discontinuity of the set of distances of near-optimal solutions. In the case of random instances , we say that the problem exhibits the OGP if the problem exhibits the OGP with high likelihood when is generated according to the law . An illustration of the OGP is depicted in SI Appendix, Fig. S2. The notion of the ”overlap” refers to the fact that distances in normed space are directly relatable to inner products, commonly called overlaps in the spin-glass theory, via and the fact that solutions themselves often have identical or close to identical norms. The OGP is of interest only for certain choices of parameters , which are always problem-dependent. Furthermore, all of these parameters along with the optimal value are usually dependent on the problem size, such as the number of Boolean variables in the K-SAT model or the number of nodes in a graph. In particular, the property is of interest only if there are pairs of -optimal solutions satisfying both and property. The first case is trivial, as we can take , but the existence of pairs with distance at least needs to be verified. Establishing the OGP rigorously can be either straightforward or technically very involved, again depending on the problem. It is should not be surprising that the presence of the OGP presents a potential difficulty in finding an optimal solution, due to the presence of multiple local minima, similarly to the lack of convexity property. An important distinction should be noted, however, between the OGP and the lack of convexity property. The function can be nonconvex, but not exhibiting the OGP, as depicted in SI Appendix, Fig. S3. Thus, the ”common” intractability obstruction presented by nonconvexity is not identical with the OGP. Also, the solution space is often discrete (such as the binary cube), rendering the notions of convexity nonapplicable.
We now extend the OGP definition to the so-called ensemble OGP, abbreviated as the e-OGP, and the multi-OGP, abbreviated as the m-OGP. We say that a set of problem instances satisfies the e-OGP with parameters if for every pair of instances , for every -optimal solution of the instance (namely, ), and every -optimal solution of the instance (namely, ), it is the case that either or , and in the case when instances and are probabilistically independent, the former case (i.e., ) is not possible. The set represents a collection of problem instances over which the optimization problem is considered. For example, might represent a collection of correlated random graphs or random matrices or random tensors. Indeed, below, we will provide an example of a family of correlated random graphs as a main example of the case when the e-OGP holds. We see that the OGP is the special case of the e-OGP when consists of a single instance .
Finally, we say that a family of instances satisfies the m-OGP with parameters and if for every -collection of instances and every collection of solutions that are -optimal with respect to , at least one pair satisfies or . Informally, the m-OGP means that one cannot find near-optimal solutions for instances such that all pairs of distances lie in the interval . Clearly, the case boils down to the e-OGP discussed earlier. In many applications, the difference between and is often significantly smaller than the value of itself; that is, . Thus, roughly speaking, the m-OGP means that one cannot find near-optimal solutions so that all pairwise distances are . The first usage of OGP as an obstruction to algorithms was adopted by Gamarnik and Sudan in ref. 46. The first application of m-OGP as an obstruction to algorithms was adopted by Rahman and Virag in ref. 47. While their variant of m-OGP proved useful for some applications (29), since then, the definition of m-OGP has been extended to ”less symmetric” variants, as one by Wein in ref. 48 and by Bresler and Huang in ref. 30. We will not discuss the nature of these extensions and, instead, refer to the aforementioned references for further details. e-OGP was introduced by Chen et al. (49).
OGP Is an Obstruction to Stability.
Next, we discuss how the presence of variants of the OGP constitutes an obstruction to algorithms. The broadest class of algorithms for which such an implication can be established is algorithms that are stable with respect to a small perturbation of instances . An algorithm is viewed here simply as a mapping from instances to solutions , which we abbreviate as . Suppose we have a parametrized collection of instances with discrete parameter taking values in some interval . We say that the algorithm is stable, or specifically -stable, with respect to the family if for every , . Informally, if we think of as an instance obtained from by a small perturbation, the output of the algorithm does not change much: The algorithm is not very sensitive to the input. Continuous versions of such stability have been considered as well, specifically in the context of models with Gaussian distribution. Here, is a continuous parameter, and the algorithm is stable with sensitivity value if for every . Often, these bounds are established only with probabilistic guarantees, both in the case of discrete and continuously valued .
Now, assume that the e-OGP holds for a family of instances . Assume two additional conditions: 1) The regime for -optimal solutions of , and -optimal solutions of is nonexistent (namely, every -optimal solution of is far from every -optimal solution of ); and 2) . The property (1) is typically verified since the endpoints and of the interpolated sequence are often independent, and (1) can be checked by straightforward moment arguments. The verification of condition (2) typically involves technical and problem-dependent arguments. The conditions (1) and (2) above plus the OGP allow us to conclude that any -stable algorithm fails to produce a -optimal solution. This is seen by a simple contradiction argument based on continuity: Consider the sequence of solutions , produced by the algorithm . We have for every and by the OGP and assumption (1) above. Suppose, for the purposes of contradiction, that every solution produced by the algorithm is -optimal for every instance . Then, by the OGP, it is the case that, in particular, is either at most or at least for every . Since , there exists , possibly , such that and . Then, , which is a contradiction with the assumption (2).
Simply put, the crux of the argument is that the algorithm cannot ”jump” over the gap , since the incremental distances, bounded by , are too small to allow for it. This is the main method by which stable algorithms are ruled out in the presence of the OGP and is illustrated in SI Appendix, Fig. S4. In this figure, the smaller circle represents all of the -optimal solutions that are at most distance from , across all instances . The complement to the larger circle represents all of the -optimal solutions that are at least distance from , again across all instances . All -optimal solutions across all instances should fall into one of these two regions, according to the e-OGP. As the sequence of solutions has to cross between these two regions, at some point , the distance between and will have to exceed , contradicting the stability property.
Suppose the model does not exhibit the OGP, but does exhibit the m-OGP (ensemble version). We now describe how this can also be used as an obstruction to stable algorithms. As above, the argument is based on first verifying that for two independent instances and , the regime is not possible with overwhelming probability. Consider independent instances and consider an interpolated sequence with the property that and for all . In other words, the sequence starts with identical copies of instance and slowly interpolates toward instance . The details of such a construction are usually guided by concrete problems. Typically, such constructions are also symmetric in distribution, so that, in particular, all pairwise expected distances between the algorithmic outputs are the same, say, denoted by . Furthermore, in some cases, a concentration around the expectation can be also established. As an implication, this set of identical pairwise distances spans values from zero to value larger than by the assumption (1). But the stability of the algorithm also implies that at some point , it will be the case that , contradicting the m-OGP. This is illustrated in SI Appendix, Fig. S5. The high-concentration property discussed above was established by using standard methods in refs. 29 and 48, but the approach in ref. 50 was based on different arguments employing methods from the Ramsey extremal combinatorics field. Here, the idea is to generate many more independent instances than the value of arising in m-OGP and showing using Ramsey theory results that there should exist a subset of -instances such that all pairwise distances fall into interval .
OGP for Concrete Models.
We now illustrate the OGP and its variants for the Number Partitioning problem and the maximum clique problem as examples. The discussion will be informal, and references containing detailed derivations will be provided. The instance of the Number Partitioning problem is a sequence of independent standard normal random variables. For every binary vector , the inner product is a mean zero normal random variable with variance . The likelihood that takes value in an interval around zero with length is roughly times the length of the interval, namely, times . Thus, the expected number of such vectors is of constant order since there are choices for . This gives an intuition for why there do exist vectors achieving value . Details can be found in ref. 11. Any choice of order magnitude smaller length interval leads to expectation converging to zero and, thus, is nonachievable with overwhelming probability. Now, to discuss the OGP, fix constants and consider pairs of vectors , which achieve objective value and have the scaled inner product . The expected number of such pairs is roughly times the area of the square with side length approximately , namely, . This is because the likelihood that a pair of correlated Gaussians falls into a very small square around zero is dominated by the area of this square up to a constant. Approximating as , where is the standard binary entropy, the expectation of the number of such pairs evaluates to roughly . As the largest value of is , we obtain an important discovery: For every , there is a region of of the form for which the expectation is exponentially small in , and, thus, the value is not achievable by pairs of partitions and with scaled inner products in , with overwhelming probability. Specifically, for every pair of -optimal solutions with , it is the case that is either one or at most . With this value of and , we conclude that the model exhibits the OGP.
Showing the ensemble version of OGP (namely, e-OGP) can be done using very similar calculations, but what about values of smaller than ? After all, the state-of-the-art algorithms achieve only order , which corresponds to at scale. It turns out (50) that, in fact, the m-OGP holds for this model for all strictly positive with a judicious choice of constant value . Furthermore, the m-OGP extends all the way to by taking growing with , but not beyond that value, at least using the available proof techniques. Thus, at this stage, we may conjecture that the problem is algorithmically hard for objective values smaller than order , but beyond this value, we don’t have plausible predictions either for hardness or tractability.
The largest clique problem is another example where the OGP-based predictions are consistent with the state-of-the-art algorithms. Recall from the earlier section that while the largest clique in is roughly , the largest size clique achievable by known algorithm is roughly half-optimal with size approximately . Fixing , consider pairs of cliques with sizes roughly . This range of corresponds to clique sizes known to be present in the graph existentially with overwhelming probability, but not achievable algorithmically. As it turns out (46, 51), the model exhibits the OGP for and overlap parameters , both of which are at the scale . Specifically, every two cliques of size have intersection size either at least or at most , for some values , both at scale . This easily implies OGP with . The calculations in refs. 46 and 51 were carried out for a similar problem of finding a largest independent set (a set with no internal edges) in sparse random graphs with of the order , but the calculation details for the dense graph are almost identical. The end result is that , where and are the two roots of the quadratic equation , namely, , and roots exist if and only if . For recent algorithmic results for the related problem on random regular graphs with small degree, see ref. 52.
To illustrate the ensemble version of the OGP, consider an ensemble of random graphs, constructed as follows. Generate according to the probability law of . Introduce a random uniform order of pairs of nodes , and generate graph from by resampling the pair —the -th pair in the order —and leaving other edges of intact. This way, each graph individually has the law of , but neighboring graphs and differ by at most one edge. The beginning and the end graphs and are clearly independent. Now, fix and A simple calculation shows that the expected number of size cliques in and same-size cliques in , such that the size of the intersection of and is , is approximately given by , where is the (rescaled) number of edges of , which were resampled from . Easy calculation shows that the function has two roots in the interval given by , when the larger root is smaller than , namely, when . On the other hand, when , including the extreme case corresponding to , there is only one root in . The model thus not only does exhibit the e-OGP for the ensemble . Additionally, the value represents a new type of phase transition. When , every two cliques with size have intersection size either at least or at most . On the other hand, when , the second regime disappears, and any two such cliques have intersection size only at most . The situation is illustrated in SI Appendix, Fig. S6 for the special case and .
The value is not tight with respect to the apparent algorithmic hardness, which occurs at , as discussed earlier. This is where the multi-OGP comes to rescue. A variant of an (asymmetric) m-OGP has been established recently in ref. 48, building on an earlier symmetric version of m-OGP, discovered in ref. 47, ruling out the class of stable algorithms as potential contenders for solving the Independent Set problem above . More specifically, in both works, it is shown that for every fixed , there exists a constant such that the model exhibits m-OGP with this value of for independent sets corresponding to .
Stability of Low-Degree Polynomials.
How does one establish the link of the form OGP implies Class of algorithms fails, and what classes of algorithms can be proven to fail in presence of the OGP? As discussed earlier, the failure is established by showing stability (insensitivity) of the algorithms in class C to the changes of the input. The largest class of algorithms to the day that is shown to be stable is the class of algorithms informally dubbed as ”low-degree polynomials.” This class of algorithms has deep connections with the so-called Sum-of-Squares method (42–44). Roughly speaking, this is a class of algorithms where the solution is obtained by constructing an -sequence of multivariate polynomials with variables evaluated at entries of the instance . The largest degree of the polynomial is supposed to be low, though depending on the problem, it can take even significantly high value. What can be said about this method in the context of the problems exhibiting the OGP? It was shown in refs. 30, 48, and 51 that degree- polynomial-based algorithms are stable. This is true even when is as large as . Thus, algorithms based on polynomials even with degree fail to find near-optimum solutions for these problems due to the presence of the OGP. Many algorithms can be modeled as special cases of low-degree polynomials, and therefore ruled out by OGP, including the so-called Approximate Message Passing algorithms (51, 53), local algorithms considered in refs. 46 and 47, and quantum versions of local algorithms known as QAOA (4).
OGP and the Problem of Finding Ground States of p-Spin Models.
There is another large class of optimization problems for which the OGP approach provides a tight classification of solvable vs. not yet solvable in polynomial time problems. It is the problem of finding ground states of p-spin models—the very class of optimization problems that led to studying the overlap distribution and overlap gap-type properties to begin with. The problem is described as follows: Suppose one is given a -order tensor of side length . Namely, is described as a collection of real values , where range between one and . Given the solution space , the goal is finding a solution (called ground state), which minimizes the inner product . A common assumption, capturing most of the structural and algorithmic complexity of the model, is that the entries are independent standard normal entries. Two important special cases of the solution space include the case when is a binary cube and when it is a sphere of a fixed radius (say, one) in the -dimensional real space, the latter case being referred to as the spherical spin-glass model. The special case and is the celebrated Sherrington–Kirkpatrick model (54)—the centerpiece of the spin-glass theory. The problem of computing the value of this optimization problem above was at the core of the spin-glass theory in the past four decades, which led to the replica symmetry-breaking technique first introduced in the physics literature by Parisi (25) and then followed by a series of mathematically rigorous developments, including those by Talagrand (55) and Panchenko (56).
The algorithmic progress on this problem, however, is quite recent. In the important development by Subag (57) and Montanari (58), polynomial time algorithms were constructed for finding near ground states in these models. Remarkably, this was done in the regimes where the model does not exhibit the OGP. A follow-up development in ref. 59 extended this to models that do exhibit OGP, but in this case, near ground states are not reached. Instead, one constructs the best solution within the part of the solution space not exhibiting the OGP. To put it differently, the algorithms produce a -optimal solution for the smallest value of for which the OGP does not hold. Conversely (51, 53, 60), the optimization problem above cannot be solved within the class of stable algorithms.
OGP, the Clustering Property, and the Curious Case of the Perceptron Model.
We now return to the subject of weak and strong clustering property and discuss it from the perspective of the OGP and algorithmic hardness. The relationship is nontrivial and perhaps best exemplified by yet another model, which exhibits statistical to computation gap: the binary perceptron model. The input to the model is matrix of i.i.d. standard normal entries . A fixed parameter is fixed. The goal is to find a binary vector such that for each . This is also known as the binary symmetric perceptron model, to contrast it with the asymmetric model, where the requirement is that , for some fixed parameter (positive or negative), with inequalities interpreted coordinate-wise. In the asymmetric case, a typical choice is . Replica symmetry-breaking-based heuristic methods developed in this context by Krauth and Mézard (61) predict a sharp threshold at for the case , so that when and such a vector exists, and when such a vector does not exist, both with high probability as . This was confirmed rigorously as a lower bound in ref. 62. For the symmetric case, a similar sharp threshold is known at a -dependent constant , as established in refs. 63 and 64.
The known algorithmic results, however, are significantly weaker. The algorithm by Kim and Rouche (65) finds for the asymmetric case when , notably much smaller than . It is quite likely (though not known) that a version of this algorithm should work in the symmetric case as well for some positive -dependent constant . Heuristically, the message-passing algorithm was found to be effective at small positive densities, as reported in ref. 66. Curiously, though, it is known that the symmetric model exhibits the weak clustering property at all positive densities (63, 64)! Furthermore, quite remarkably, each cluster consists of singletons! Namely, the cardinality of each cluster is one, akin to the ”needles in the haystack” metaphor. A similar picture was established in the context of the Number Partitioning problem (50). This interesting entropic (due to the subextensive cardinality of clusters) phenomenon is fairly novel, and its algorithmic implications have been also discussed in ref. 67. Thus, assuming the extendibility of the Kim–Roche algorithm and/or validity of the message-passing algorithm, we have an example of a model exhibiting weak clustering property, but amenable to polynomial time algorithms. As an explanation of this phenomena, the aforementioned references point out the ”nonuniformity” in algorithmic choices of the solutions, allowing it somehow to bypass the overwhelming clustering picture exhibited by the weak clustering property.
What about the strong clustering property and the OGP? For any model, the presence of the pairwise () OGP for a single instance implies the strong clustering property: If every pair of -optimal solutions is at distance at most or at least , and the second case is nonvacuous, clearly, the set of solutions is strongly clustered. It, furthermore, indicates that the diameter of each cluster is strictly smaller (at scale) than the distances between the clusters (SI Appendix, Figs. S2 and S3). In fact, this is precisely how the strong clustering property was discovered to begin with in refs. 27 and 28. However, the converse does not need to be the case, as one could not rule out an example of a strong clustering property where the cluster diameters are larger than the distances between them, and, as a result, the overlaps of pairs of solutions span continuous intervals. At this stage, we are not aware of any such examples.
Back to the symmetric perceptron model, it is known that the model does exhibit the OGP, at a value strictly smaller than the critical threshold (68), but strictly larger than the algorithmic value 0.005 achieved algorithmically in ref. 65. As a result, the strong clustering property holds as well above this . Naturally, we conjecture that the problem of finding a solution is hard in this regime. In conclusion, the binary perceptron model demonstrates that the weak clustering property is provably not an obstruction to polynomial time algorithms, but OGP and its implication, the strong clustering property, likely is, at least for the case of stable algorithms.
Finally, as the strong clustering property appears to be the most closely related to the OGP, it raises a question as to whether it can be used as a ”signature” of hardness in lieu of OGP and/or whether it can be used as a method to rule out classes of algorithm, similarly to the OGP-based method. There are two challenges associated with this proposal. First, we are not aware of any effective method of establishing the strong clustering property other than the one based on the OGP. Second, and more importantly, it is not entirely clear how to link meaningfully the strong clustering property with m-OGP, which appears necessary for many problems in order to bridge the computational gaps to the known algorithmically achievable values. Likewise, it is not entirely clear what is the appropriate definition of either weak or strong clustering property in the ensemble version, when one considers a sequence of correlated instances, which is again another important ingredient in the implementation of the refutation-type arguments. Perhaps some form of dynamically evolving clustering picture is of relevance here. Some recent progress on the former question regarding the geometry of the m-OGP was obtained recently in ref. 69 in the context of the spherical spin-glass model. We leave both questions as interesting open venues for future investigation.
Discussion
In this article, we have discussed a new approach to computational complexity arising in optimality studies of random structures. The approach, which is based on topological disconnectivity of the overlaps of near-optimal solutions, called the OGP, is both a theory and a method. As a theory, it predicts the onset of algorithmic hardness in random structures at the onset of the OGP. The approach has been verified at various levels of precision in many classes of problems, and the summary of the state of the art is presented in SI Appendix, Table S1. In this table, the current list of models known to exhibit an apparent algorithmic hardness is provided, along with references and notes indicating whether the OGP-based method matches the state-of-the-art algorithmic knowledge. The ”Yes” indicates this to be the case, and ”Not known” indicates that the formal analysis has not been completed yet. Remarkably, to the day, we are not aware of a model that does not exhibit the OGP, but that is known to exhibit some form of an apparent algorithmic hardness.
The OGP-based approach also provides a concrete method for rigorously establishing barriers for classes of algorithms. Typically, such barriers are established by verifying certain stability (input insensitivity) of algorithms, making them inherently incapable of overcoming the gaps appearing in the overlap structures. While the general approach for ruling out such classes of algorithm is more or less problem-independent, the exact nature of such stability, as well as the OGP itself, is very much problem-dependent, and the required mathematical analysis varies from borderline trivial to extremely technical, often relying on the state-of-the-art development in the field of mathematical theory of spin glasses.
Supplementary Material
Acknowledgments
This work was supported by NSF Grant DMS-2015 517. I thank the anonymous reviewers for many suggestions, which led to significant improvement of the presentations of the manuscript.
Footnotes
The author declares no competing interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2108492118/-/DCSupplemental.
Data Availability
There are no data underlying this work.
References
- 1.Vershynin R., High-Dimensional Probability: An Introduction with Applications in Data Science (Cambridge Series in Statistical and Probabilistic Mathematics, Cambridge University Press, Cambridge, UK, 2018), vol. 47. [Google Scholar]
- 2.Bühlmann P., Van De Geer S., Statistics for High-Dimensional Data: Methods, Theory and Applications (Springer Series in Statistics, Springer Science & Business Media, Berlin, Germany, 2011). [Google Scholar]
- 3.Foucart S., Rauhut H., A Mathematical Introduction to Compressive Sensing (Applied and Numerical Harmonic Analysis, Birkhäuser, New York, 2013). [Google Scholar]
- 4.Farhi E., Gamarnik D., Gutmann S., The quantum approximate optimization algorithm needs to see the whole graph: A typical case. arXiv [Preprint] (2020). https://arxiv.org/abs/2004.09002 (Accessed 15 September 2021).
- 5.Farhi E., Gamarnik D., Gutmann S., The quantum approximate optimization algorithm needs to see the whole graph: Worst case examples. arXiv [Preprint] (2020). https://arxiv.org/abs/2005.08747 (Accessed 15 September 2021).
- 6.Farhi E., Goldstone J., Gutmann S., A quantum approximate optimization algorithm. arXiv [Preprint] (2014). https://arxiv.org/abs/1411.4028 (Accessed 15 September 2021).
- 7.Alon N., Spencer J. H., The Probabilistic Method (John Wiley & Sons, New York, 2004). [Google Scholar]
- 8.Karp R. M., “The probabilistic analysis of some combinatorial search algorithms” in Algorithms and Complexity: New Directions and Recent Results, J. F. Traub, Ed. (Academic, New York, 1976), vol. 1, pp. 1–19. [Google Scholar]
- 9.Jerrum M., Large cliques elude the metropolis process. Random Structures Algorithms 3, 347–359 (1992). [Google Scholar]
- 10.Garey M. R., Johnson D.S., Computers and Intractability (Series of Books in the Mathematical Sciences, Freeman, San Francisco, CA, 1979), vol. 174. [Google Scholar]
- 11.Karmarkar N., Karp R. M., The Differencing Method of Set Partitioning (Computer Science Division, University of California Berkeley, Berkeley, CA, 1982). [Google Scholar]
- 12.Boettcher S., Mertens S., Analysis of the Karmarkar-Karp differencing algorithm. Eur. Phys. J. B 65, 131–140 (2008). [Google Scholar]
- 13.Hastad J., Clique is hard to approximate within n1–ε. Acta Math. 182, 105–142 (1999). [Google Scholar]
- 14.Lipton R., New directions in testing. Distributed Comput. Cryptogr. 2, 191–202 (1991). [Google Scholar]
- 15.Gamarnik D., Kızıldağ E. C., Computing the partition function of the Sherrington–Kirkpatrick model is hard on average. Ann. Appl. Probab. 31, 1474–1504 (2021). [Google Scholar]
- 16.Ajtai M., “Generating hard instances of lattice problems” in Proceedings of the Twenty-Eighth Annual ACM Symposium on Theory of Computing (Association for Computing Machinery, New York, 996), pp. 99–108. [Google Scholar]
- 17.Choromanska A., Henaff M., Mathieu M., Ben Arous G., LeCun Y., “The loss surfaces of multilayer networks” in Proceedings of the 18th International Conference on Artificial Intelligence and Statistics (AISTATS) (JMLR, 2015), pp. 192–204.
- 18.Fu Y., Anderson P. W., Application of statistical mechanics to NP-complete problems in combinatorial optimisation. J. Phys. Math. Gen. 19, 1605 (1986). [Google Scholar]
- 19.Kirkpatrick S., Selman B., Critical behavior in the satisfiability of random Boolean expressions. Science 264, 1297–1301 (1994). [DOI] [PubMed] [Google Scholar]
- 20.Monasson R., Zecchina R., Kirkpatrick S., Selman B., Troyansky L., Determining computational complexity from characteristic ‘phase transitions’. Nature 400, 133–137 (1999). [Google Scholar]
- 21.Mezard M., Parisi G., Virasoro M. A., Spin-Glass Theory and Beyond (Lecture Notes in Physics, World Scientific, Singapore, 1987), vol. 9. [Google Scholar]
- 22.Mezard M., Montanari A., Information, Physics and Computation (Oxford Graduate Texts, Oxford University Press, Oxford, UK, 2009). [Google Scholar]
- 23.Ding J., Sly A., Sun N., “Proof of the satisfiability conjecture for large k” in STOC’15: Proceedings of the Forty-Seventh Annual ACM Symposium on Theory of Computing (Association for Computing Machinery, New York, 2015), pp. 59–68.
- 24.Friedgut E., Sharp thresholds of graph properties, and the k-SAT problem. J. Am. Math. Soc. 4, 1017–1054 (1999). [Google Scholar]
- 25.Parisi G., A sequence of approximated solutions to the SK model for spin glasses. J. Phys. Math. Gen. 13, L115 (1980). [Google Scholar]
- 26.Coja-Oghlan A., A better algorithm for random k-SAT. SIAM J. Comput. 39, 2823–2864 (2010). [Google Scholar]
- 27.Achlioptas D., Ricci-Tersenghi F., “On the solution-space geometry of random constraint satisfaction problems” in STOC’06: Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing (Association for Computing Machinery, New York, NY, 2006), pp. 130–139.
- 28.Mézard M., Mora T., Zecchina R., Clustering of solutions in the random satisfiability problem. Phys. Rev. Lett. 94, 197205 (2005). [DOI] [PubMed] [Google Scholar]
- 29.Gamarnik D., Sudan M., Performance of sequential local algorithms for the random NAE-k-SAT problem. SIAM J. Comput. 46, 590–619 (2017). [Google Scholar]
- 30.Bresler G., Huang B., The algorithmic phase transition of random k-sat for low degree polynomials. arXiv [Preprint] (2021). https://arxiv.org/abs/2106.02129 (Accessed 15 September 2021).
- 31.Hetterich S., “Analysing survey propagation guided decimationon random formulas” in 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), Chatzigiannakis I., Mitzenmacher M., Rabani Y., Sangiorgi D., Eds. (Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, Dagstuhl, Germany, 2016), pp. 65:1–65:12. [Google Scholar]
- 32.Coja-Oghlan A., Haqshenas A., Hetterich S., Walksat stalls well below satisfiability. SIAM J. Discrete Math. 31, 1160–1173 (2017). [Google Scholar]
- 33.Mézard M., Parisi G., Zecchina R., Analytic and algorithmic solution of random satisfiability problems. Science 297, 812–815 (2002). [DOI] [PubMed] [Google Scholar]
- 34.Gomes C. P., Selman B., Computer science. Satisfied with physics. Science 297, 784–785 (2002). [DOI] [PubMed] [Google Scholar]
- 35.Ricci-Tersenghi F., Semerjian G., On the cavity method for decimated random constraint satisfaction problems and the analysis of belief propagation guided decimation algorithms. J. Stat. Mech. 2009, P09001 (2009). [Google Scholar]
- 36.Coja-Oghlan A., Belief propagation guided decimation fails on random formulas. J. ACM, 63, 1–55 (2017). [Google Scholar]
- 37.Ricci-Tersenghi F., Mathematics. Being glassy without being hard to solve. Science 330, 1639–1640 (2010). [DOI] [PubMed] [Google Scholar]
- 38.Budzynski L., Ricci-Tersenghi F., Semerjian G., Biased landscapes for random constraint satisfaction problems. J. Stat. Mech. 2019, 023302 (2019). [Google Scholar]
- 39.Budzynski L., Semerjian G., Biased measures for random constraint satisfaction problems: larger interaction range and asymptotic expansion. J. Stat. Mech. 2020, 103406 (2020). [Google Scholar]
- 40.Krzakała F., Montanari A., Ricci-Tersenghi F., Semerjian G., Zdeborová L., Gibbs states and the set of solutions of random constraint satisfaction problems. Proc. Natl. Acad. Sci. U.S.A. 104, 10318–10323 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Marino R., Parisi G., Ricci-Tersenghi F., The backtracking survey propagation algorithm for solving random K-SAT problems. Nat. Commun. 7, 12996 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Hopkins S. B., Steurer D., “Efficient Bayesian estimation from few samples: community detection and related problems” in 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS) (IEEE, New York, 2017), pp. 379–390. [Google Scholar]
- 43.Hopkins S. B., et al., “The power of sum-of-squares for detecting hidden structures” in 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS) (IEEE, New York, 2017), pp. 720–731. [Google Scholar]
- 44.Hopkins S., “Statistical inference and the sum of squares method,” PhD thesis, Cornell University, Ithaca, NY (2018).
- 45.Gamarnik D., Li Q., Finding a large submatrix of a Gaussian random matrix. Ann. Stat. 46 (6A), 2511–2561 (2018). [Google Scholar]
- 46.Gamarnik D., Sudan M., Limits of local algorithms over sparse random graphs. Ann. Probab. 45, 2353–2376 (2017). [Google Scholar]
- 47.Rahman M., Virag B., Local algorithms for independent sets are half-optimal. Ann. Probab. 45, 1543–1577 (2017). [Google Scholar]
- 48.Wein A. S., Optimal low-degree hardness of maximum independent set. arXiv [Preprint] (2020). https://arxiv.org/abs/2010.06563 (Accessed 15 September 2021).
- 49.Chen W.-K., Gamarnik D., Panchenko D., Rahman M., Suboptimality of local algorithms for a class of max-cut problems. Ann. Probab. 47, 1587–1618 (2019). [Google Scholar]
- 50.Gamarnik D., Kızıldağ E. C., Algorithmic obstructions in the random number partitioning problem. arXiv [Preprint] (2021). https://arxiv.org/abs/2103.01369 (Accessed 15 September 2021).
- 51.Gamarnik D., Jagannath A., Wein A. S., “Low-degree hardness of random optimization problems” in 61st Annual Symposium on Foundations of Computer Science (2020).
- 52.Marino R., Kirkpatrick S., Large independent sets on random d-regular graphs with d small. arXiv [Preprint] (2020). https://arxiv.org/abs/2003.12293 (Accessed 15 September 2021).
- 53.Gamarnik D., Jagannath A., The overlap gap property and approximate message passing algorithms for p-spin models. Ann. Probab. 49, 180–205 (2021). [Google Scholar]
- 54.Sherrington D., Kirkpatrick S., Solvable model of a spin-glass. Phys. Rev. Lett. 35, 1792 (1975). [Google Scholar]
- 55.Talagrand M., Basic Examples (Mean Field Models for Spin Glasses, Springer, Berlin, Germany, 2010), vol. 1. [Google Scholar]
- 56.Panchenko D., The Sherrington-Kirkpatrick Model (Springer Monographs in Mathematics, Springer Science & Business Media, New York, 2013). [Google Scholar]
- 57.Subag E., Following the ground states of full-RSB spherical spin glasses. Commun. Pure Appl. Math. 74.5, 1021–1044 (2021). [Google Scholar]
- 58.Montanari A., Optimization of the Sherrington–Kirkpatrick Hamiltonian. SIAM J. Sci. Comput. 2021, FOCS19-1 (2021). [Google Scholar]
- 59.El Alaoui A., Montanari A., Sellke M., Optimization of mean-field spin glasses. arXiv [Preprint] (2020). https://arxiv.org/abs/2001.00904 (Accessed 15 September 2021).
- 60.Sellke M., Approximate ground states of hypercube spin glasses are near corners. arXiv [Preprint] (2020). https://arxiv.org/abs/2009.09316 (Accessed 15 September 2021).
- 61.Krauth W., Mézard M., Storage capacity of memory networks with binary couplings. J. Phys. (Paris) 50, 3057–3066 (1989). [Google Scholar]
- 62.Ding J., Sun N., “Capacity lower bound for the Ising perceptron” in Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing (Association for Computing Machinery, New York, 2019), pp. 816–827.
- 63.Perkins W., Xu C., “Frozen 1-RSB structure of the symmetric Ising perceptron” in Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing (Association for Computer Machinery, New York, 2021), pp. 1579–1588.
- 64.Abbe E., Li S., Sly A., Proof of the contiguity conjecture and lognormal limit for the symmetric perceptron. arXiv [Preprint] (2021). https://arxiv.org/abs/2102.13069 (Accessed 15 September 2021).
- 65.Kim J. H., Roche J.R., Covering cubes by random half cubes, with applications to binary neural networks. J. Comput. Syst. Sci. 56, 223–252 (1998). [Google Scholar]
- 66.Braunstein A., Zecchina R., Learning by message passing in networks of discrete synapses. Phys. Rev. Lett. 96, 030201 (2006). [DOI] [PubMed] [Google Scholar]
- 67.Bellitti M., Ricci-Tersenghi F., Scardicchio A., Entropic barriers as a reason for hardness in both classical and quantum algorithms. arXiv [Preprint] (2021). https://arxiv.org/abs/2102.00182 (Accessed 15 September 2021).
- 68.Baldassi C., Della Vecchia R., Lucibello C., Zecchina R., Clustering of solutions in the symmetric binary perceptron. J. Stat. Mech. 2020, 073303 (2020). [Google Scholar]
- 69.Ben Arous G., Jagannath A., Shattering versus metastability in spin glasses. arXiv [Preprint] (2021). https://arxiv.org/abs/2104.08299 (Accessed 15 September 2021).
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
There are no data underlying this work.
