Skip to main content
Entropy logoLink to Entropy
. 2019 Jan 24;21(2):108. doi: 10.3390/e21020108

Doubly Nonnegative and Semidefinite Relaxations for the Densest k-Subgraph Problem

Chuan-Hao Guo 1,*, Yuan Guo 1, Bei-Bei Liu 1
PMCID: PMC7514591  PMID: 33266824

Abstract

The densest k-subgraph (DkS) maximization problem is to find a set of k vertices with maximum total weight of edges in the subgraph induced by this set. This problem is in general NP-hard. In this paper, two relaxation methods for solving the DkS problem are presented. One is doubly nonnegative relaxation, and the other is semidefinite relaxation with tighter relaxation compare with the relaxation of standard semidefinite. The two relaxation problems are equivalent under the suitable conditions. Moreover, the corresponding approximation ratios’ results are given for these relaxation problems. Finally, some numerical examples are tested to show the comparison of these relaxation problems, and the numerical results show that the doubly nonnegative relaxation is more promising than the semidefinite relaxation for solving some DkS problems.

Keywords: densest k-subgraph, doubly nonnegative relaxation, semidefinite relaxation, approximation ratio, NP-hard

MSC: 90C10, 90C20, 49M20, 65K05

1. Introduction

In this paper, the densest k-subgraph (DkS) problem [1,2] is considered. For a given graph G and a parameter k, the DkS problem consists in finding a maximal average degree in the subgraph induced by the set of k vertices. This problem was first introduced by Corneil and Perl as a natural generalization of the maximum clique problem [3]. It is NP-hard on restricted graph classes such as chordal graphs [3], bipartite graphs [3] and planar graphs [4]. The DkS problem is a classical problem of combinatorial optimization and arises in several applications, such as facility location [5], community detection in social networks, identifying protein families and molecular complexes in protein–protein interaction networks [6], etc. Since the DkS problem is in general NP-hard, there are a few approximation methods [7,8,9] for solving it. It is well-known that semidefinite relaxation is a powerful and computationally efficient approximation technique for solving a host of very difficult optimization problems, for instance, the max-cut problem [10] and the boolean quadratic programming problem [11]. It also has been at the center of some of the very exciting developments in the area of signal processing [12,13].

Optimization problems over the doubly nonnegative cone arise, for example, as a strengthening of the Lovasz-ϑ-number for approximating the largest clique in a graph [14]. The recent work by Burer [15] stimulated the interest in optimization problems over the completely positive cone. A tractable approximation to such problem being defined as an optimization problem over the doubly nonnegative cone. By using the technique of doubly nonnegative relaxation, Bai and Guo proposed an effective and promising method for solving multiple objective quadratic programming problems in [16]. For more details and developments of this technique, one may refer to [17,18,19] and the references therein. It is worth pointing out that the cone of doubly nonnegative matrices is a subset of a positive semidefinite matrices cone. Thus, the doubly nonnegative relaxation is more promising than the basic semidefinite relaxation. Moreover, such relaxation problems can be efficiently solved by some popular package software.

In this paper, motivated by the idea of doubly nonnegative relaxation and semidefinite relaxation, the two relaxation methods for solving the DkS problem are presented. One is doubly nonnegative relaxation, and the other is semidefinite relaxation with tighter relaxation. Furthermore, we prove that the two relaxation problems are equivalent under the suitable conditions. Some approximation accuracy results about these relaxation problems are also given. Finally, we report some numerical examples to show the comparison of the two relaxation problems. The numerical results show that the doubly nonnegative relaxation is more promising than the semidefinite relaxation for solving some DkS problems.

The paper is organized as follows: we present doubly nonnegative relaxation and a new semidefinite relaxation with tighter relaxation for the DkS problem in Section 2.1 and Section 2.2, respectively. In Section 3, we prove that the two new relaxations proposed in Section 2 are equivalent. In Section 4, some approximation accuracy results for the proposed relaxation problems are given. Some comparative numerical results are reported in Section 5 to show the efficiency of the proposed new relaxations. Moreover, some concluding remarks are given in Section 6.

2. Two Relaxations for the Densest k-Subgraph Problem

First of all, the definition of the densest k-subgraph (DkS) problem is given as follows.

Definition 1 (Densest k-subgraph).

For a given graph G(V,E), where V is the vertex set and E is the edge set. The DkS problem on G(V,E) is the problem of finding a vertex subset of V of size k with the maximum induced average degree.

Given a symmetric n×n matrix A=(aij), the weighted graph with vertex set {1,2,,n} associates with A in such a way: the edge [i,j] with the weight aij is introduced in the graph. Then, A is interpreted as the weighted adjacency matrix of the graph with the vertex set V={1,2,,n}. Based on Definition 1, the DkS problem consists of determining a subset V1V consisting of k vertices such that the total weight of edges in the subgraph spanned by V1 is maximized. To select subgraphs, assign a decision variable yi{0,1} for each node (yi=1 if the node is taken, and yi=0 if the node is not). The weight of the subgraph given by y is yTAy. Thus, the DkS problem can be phrased as the 01 quadratic problem

(DkS)max12yTAy,s.t.yTe=k,y{0,1}n.

It is known that the (DkS) problem is NP-hard [20], even though A is assumed to be positive semidefinite, since the feasible space of the (DkS) problem is nonconvex. For solving this problem efficiently, we present the two new relaxations for the (DkS) problem in the following subsections, based on the idea of approximation methods.

2.1. Doubly Nonnegative Relaxation

Note that the quadratic term yTAy in the (DkS) problem can also be expressed as AyyT. By introducing a new variable Y=yyT and taking lifting techniques, we could reformulate the (DkS) problem into the following completely positive programming problem:

(CPPDkS)max12AY,s.t.yTe=k,eTYe=k2,diag(Y)=y,1yTyYC1+n,

where C1+n is defined as follows:

C1+n:=XS1+n:X=hHzh(zh)T{0},

and for some finite vectors {zh}hHR1+n+\{0}.

The following theorem shows the relationship between the (DkS) problem and the (CPPDkS) problem. Its proof is similar to the one of Theorem 2.6 in [15] and is omitted here.

Theorem 1.

(i) The (DkS) problem and the (CPPDkS) problem have the same optimal values of objective functions, i.e., Opt(DkS)=Opt(CPPDkS); (ii) if (y*,Y*) is an optimal solution for the (CPPDkS) problem; then, y* is in the convex hull of optimal solutions for the (DkS) problem.

On one hand, according to Definition 2, it is obviously that the (CPPDkS) problem is equivalent to the (DkS) problem. On the other hand, in view of the definition of convex cone, C1+n is a closed convex cone, and is called completely positive matrices cone. Thus, the (CPPDkS) problem is convex. However, since checking whether or not a given matrix belongs to C1+n is NP-hard, which has been shown by Dickinson and Gijen in [21], the (CPPDkS) problem is still NP-hard. Thus, C1+n has to be replaced or approximated by some computable cones. For example, Rn+ and Sn+ are both computable cones; furthermore, Nn+ is also a computable cone.

It is worth mentioning that Diananda’s decomposition theorem [22] can be reformulated as follows, and its proof can be found in it.

Theorem 2.

CnSn+Nn+ holds for all n. If n4, then Cn=Sn+Nn+.

The matrices cone Sn+Nn+ is sometimes called “doubly nonnegative matrices cone”. Of course, in dimension n5, there are matrices which are doubly nonnegative but not completely positive, the counterexample can be seen in [23].

By using Theorem 2, the (CPPDkS) problem can be relaxed to the problem

(DNNPDkS)max12AY,s.t.yTe=k,eTYe=k2,diag(Y)=y,1yTyYS1+n+N1+n+,

which is called the doubly nonnegative relaxation for the (DkS) problem. Some explanations are given below for this relaxation problem.

Remark 1.

Obviously, the (DNNPDkS) problem has a linear objective function and the linear constraints as well as a convex conic constraint, so it is a linear conic programming problem. Meanwhile, it is notable that S1+n+N1+n+S1+n+ and the types of variables in both the sets are the same, which further implies that the (DNNPDkS) problem could be solved by some popular package softwares for solving semidefinite programs.

2.2. New Semidefinite Relaxation

It is well-known that semidefinite relaxation is a powerful approximation technique for solving a host of combinatorial optimization problems. In this subsection, we present a new semidefinite relaxation with tighter bound for the (DkS) problem.

The idea of the standard lifting is to introduce the symmetric matrix of rank one Y=yyT. With the help of Y, we could express the integer constraints yi{0,1} as Yii=yi, and the quadratic objective function yTAy as AY. Thus, we can get the following equivalent formulation of the (DkS) problem

max12AY,s.t.yTe=k,eTYe=k2,Y=yyT,diag(Y)=y,rank(Y)=1.

Notice then that the hard constraint in the above problem is the constraint rank(Y)=1, which is moreover difficult to handle. Thus, we can relax the above problem to the following standard semidefinite relaxation problem by dropping the rank-one constraint

(ISDRDkS)max12AY,s.t.yTe=k,eTYe=k2,diag(Y)=y,1yTyYS1+n+.

For the (ISDRDkS) problem, some remarks are given below.

Remark 2.

(i) Obviously, the (ISDRDkS) problem is also a linear conic programming problem, it has the same objective function and the equality constraints with the (DNNPDkS) problem. The only difference between the (ISDRDkS) problem and the (DNNPDkS) problem is that the (DNNPDkS) problem has n(n+1)2+n nonnegative constraints more than the (ISDRDkS) problem.

(ii) Since S1+n+N1+n+S1+n+, it holds that Opt(DNNPDkS)Opt(ISDRDkS) from Remark 2 (i). Thus, the bound of the (DNNPDkS) problem is not larger than the one of the (ISDRDkS) problem. In Section 5, we implement some numerical experiments to show the comparison between the (ISDRDkS) problem and the (DNNPDkS) problem from the computational point of view.

Note that the (DkS) problem is inhomogeneous, but we can homogenize it as follows. First, let z=2ye in the (DkS) problem, it follows that z{1,1}n. Thus, the change of variable yz gives the following equivalent formulation of the (DkS) problem:

(DkS)^max18zTAz+14zTAe+18eTAe,s.t.zTe=2kn,z{1,1}n.

Then, with the introduction of the extra variable t, the (DkS) problem can be expressed as a homogeneous problem

(DkS)¯maxtzT18eTAe18eTA18Ae18Atz,s.t.tzT012eT12e0tz=2kn,tz{1,1}1+n,

where 0 is a zero matrix with appropriate dimension.

Remark 3.

The (DkS)¯ problem is equivalent to the (DkS)^ problem in the following sense: if t*z* is an optimal solution to the (DkS)¯ problem, then z* (resp. z*) is an optimal solution to the (DkS)^ problem with t*=1 (resp. t*=1).

By using the standard semidefinite relaxation technique, and letting let S=tztzT, the (DkS)¯ problem can be relaxed to the following problem:

(SDRDkS¯)max18eTAe18eTA18Ae18AS,s.t.012eT12e0S=2kn,diag(S)=e,SS1+n+.

Moreover, again by using the standard semidefinite relaxation technique directly to the (DkS)^ problem, we have from Z=zzT,

(SDRDkS^)max18AZ+14zTAe+18eTAe,s.t.zTe=2kn,eTZe=(2kn)2,diag(Z)=e,1zTzZS1+n+.

The (SDRDkS¯) problem and the (SDRDkS^) problem are both standard semidefinite relaxation problems for the (DkS) problem. The upshot of the formulations of these two relaxation problems is that they can be solved very conveniently and efficiently, to some arbitrary accuracy, by some readily available software packages, such as CVX. Note that there is only one difference between these two relaxation problems, i.e., the (SDRDkS^) problem has one equality constraint more than the (SDRDkS¯) problem. In Section 5, some comparative numerical results are reported to show the effectiveness of these two relaxations problems for solving some random (DkS) problems, respectively.

It is worth noting that the constraint z{1,+1}n in the (DkS)^ problem further implies

(1+zi)(1+zj)01+zi+zj+Zij0,1ijn (1)

always holds. Thus, adding Formula (1) to the (SDRDkS^) problem, we come up with the following new semidefinite relaxation problem

(IISDRDkS)max18AZ+14zTAe+18eTAe,s.t.zTe=2kn,eTZe=(2kn)2,diag(Z)=e,1+zi+zj+Zjj0,1ijn,1zTzZS1+n+.

Obviously, the relationship Opt(IISDRDkS)Opt(SDRDkS^) holds since the feasible set of the (IISDRDkS) problem is the subset of the feasible set of the (SDRDkS^) problem and the two problems have the same objective function.

Up to now, three new semidefinite relaxation problems for the (DkS) problem are established, i.e., the (SDRDkS¯) problem, the (SDRDkS^) problem and the (IISDRDkS) problem, in which the upper bound of the (IISDRDkS) problem is more promising than the one of the (SDRDkS^) problem. In the following sections, we will further investigate the relationship between these three problems with the (DNNPDkS) problem.

3. The Equivalence between the Relaxation Problems

The previous section establishes the doubly nonnegative relaxation (i.e., the (DNNPDkS) problem) and the semidefinite relaxation with tighter bound (i.e., the (IISDRDkS) problem) for the (DkS) problem. Note that the (DNNPDkS) problem has n inequality constraints more than the (IISDRDkS) problem. In this section, we will prove the equivalence between the two relaxations. First of all, the definition of the equivalence of two optimization problems is given as follows.

Definition 2.

Two problems P1 and P2 are called equivalent if they satisfy the following two conditions:

(i) Opt(P1)=Opt(P2).

(ii) If from a solution of P1, then a solution of P2 is readily found, and vice versa.

In order to establish the equivalence for the (DNNPDkS) problem and the (IISDRDkS) problem, a crucial theorem is given below and the details of its proof can be seen in [24] (Appendix A.5.5).

Theorem 3 (Schur Complement).

Let matrix MSn is partitioned as

M=DBBTC.

If detD0, the matrix H=CBTD1B is called the Schur complement of D in M. Then, the following relations hold:

(i) M0 if and only if D0 and H0.

(ii) If D0, then M0 if and only if H0.

To the end, by using Definition 2 and Theorem 3, we have the following main equivalence theorem.

Theorem 4.

Suppose that the feasible sets Fea(DNNPDkS) and Fea(IISDRDkS) are both nonempty. Then, the (DNNPDkS) problem and the (IISDRDkS) problem are equivalent.

Proof. 

First of all, we prove that Opt(DNNPDkS)Opt((IISDRDkS)).

Suppose that (z*,Z*) is an optimal solution of the (IISDRDkS) problem, and let

Y=14(eeT+e(z*)T+z*eT+Z*),y=12(e+z*). (2)

Directly from eTz*=2kn and Equation (2), we have

yTe=12(e+z*)Te=12(n+eTz*)=k. (3)

By Equation (2) and eTZ*e=(2kn)2, it holds that

eTYe=14eT(eeT+e(z*)T+(z*)eT+Z*)e=14(n2+2n(z*)Te+(2kn)2)=k2. (4)

Since diag(Z*)=e, Equation (2) further implies that

diag(Y)=14(e+2z*+diag(Z*)=12(e+z*)=y. (5)

From 1+zi*+zj*+Zij*0 for all 1ijn, it holds that

Yij0,1ijn. (6)

Combining with Equation (5), it is true that from Formula (6)

yi0,1in. (7)

By Theorem 3 (ii) and Equation (2), it follows that

YyyT=14(eeT+e(z*)T+z*eT+Z*)14(e+z*)(e+z*)T=14(Z*z*(z*)T)0,

i.e.,

1yTyYS1+n+. (8)

From Equations (3), (4) and (6)–(8), we have (y,Y) defined by Equation (2) is a feasible solution of the (DNNPDkS) problem. Moreover, again from Equation (2), we obtain

12AY=18A(eeT+e(z*)T+z*eT+Z*)=18AZ*+14(z*)TAe+18eTAe=Opt((IISDRDkS)),

i.e., Opt(DNNPDkS)Opt((IISDRDkS)).

Conversely, given an optimal solution (y*,Y*) to the (DNNPDkS) problem, and let

Zij=12yi*2yj*+4Yij*,zi=2yi*1,1ijn. (9)

Since diag(Y*)=y*, Equation (9) further implies that

diag(Z)=e4y*+4diag(Y*)=e. (10)

Moreover,

1+zi+zj+Zij=1+(2yi*1)+(2yj*1)+12yi*2yj*+4Yij*=4Yij*0,1ijn. (11)

From Equation (9) and eTy*=k, it follows that

zTe=(2y*e)Te=2(y*)Ten=2kn. (12)

Again from Equation (9) and eTY*e=k2, it is true that

eTZe=eT(eeT2e(y*)T2y*eT+4Y*)e=n24kn+4k2=(2kn)2. (13)

From Equation (9) and Theorem 3 (ii), it holds that

ZzzT=eeT2e(y*)T2y*eT+4Y*(2y*e)(2y*e)T=4(Y*y*(y*)T)0. (14)

By Equations (11)–(14), we can conclude that (z,Z) defined by Equation (9) is a feasible solution of the (IISDRDkS) problem. Furthermore, we have

18AZ+14zTAe+18eTAe=18A(eeT2e(y*)T2y*eT+4Y*)+14(2y*e)TAe+18eTAe=12AY*=Opt(DNNPDkS),

i.e., Opt(DNNPDkS)Opt(IISDRDkS).

Summarizing the analysis above, we obtain Opt(DNNPDkS)=Opt(IISDRDkS). From Equations (2) and (9), we observe that (y,Y) defined by Equation (2) is an optimal solution for the (DNNPDkS) problem and (z,Z) defined by Equation (9) is also an optimal solution for the (IISDRDkS) problem, respectively. According to Definition 2, we conclude that the (DNNPDkS) problem and the (IISDRDkS) problem are equivalent. ☐

The above Theorem 4 shows that Opt(DNNPDkS)=Opt(IISDRDkS). Note that the (DNNPDkS) problem has n inequality constraints more than the (IISDRDkS) problem, thus the computational cost of solving (DNNPDkS) problem may be greater than that of the (IISDRDkS) problem.

4. The Approximation Accuracy

The above section shows that the (DNNPDkS) problem is equivalent to the (IISDRDkS) problem which has the tighter upper bound compared to the (SDRDkS^) problem (see Theorem 4). In this section, we further investigate the approximation accuracy of the (DNNPDkS) problem for solving the (DkS) problem, comparing with the standard semidefinite relaxation problems which was proposed in the above sections, under some conditions.

To simplify the expression, we denote 18eTAe18eTA18Ae18A by N. First of all, note that, if k=n2, then the (SDRDkS^) problem is simplified to the following problem:

(SDRDkS^)˜maxN1zTzZ,s.t.diag(Z)=e,1zTzZS1+n+,

and the (IISDRDkS) problem can be simplified as follows:

(IISDRDkS)˜maxN1zTzZ,s.t.diag(Z)=e,1+zi+zj+Zij0,1ijn,1zTzZS1+n+.

Combining Theorem 3 in [25] with the corresponding known approximation accuracy of semidefinite relaxation for some quadratic programming problems [26], we immediately have that the following theorem holds.

Theorem 5.

If k=n2, then we have

2πOpt(SDRDkS^)=2πOpt(SDRDkS^)˜Opt(DkS)=Opt(DkS^)Opt(SDRDkS^)˜=Opt(SDRDkS^)

and

2πOpt(SDRDkS^)Opt(DNNPDkS)=Opt(IISDRDkS)=Opt(IISDRDkS)˜Opt(SDRDkS^)

for N0. Moreover, if N0 and Nij0,ij, we further have

0.87856Opt(SDRDkS^)Opt(DkS)=Opt(DkS^)Opt(SDRDkS^)

and

0.87856Opt(SDRDkS^)Opt(DNNPDkS)=Opt(IISDRDkS)Opt(SDRDkS^)

hold.

In the following analysis, we assume that kn2. We first observe that

Opt(DkS)=Opt(DkS¯),

since the (DkS¯) problem is the homogeneous problem for the (DkS) problem. In addition, note that the constraint tz{1,1}1+n in the (DkS¯) problem can be relaxed to the quadratic form tzTtz=n+1. Thus, the (DkS¯) problem can be relaxed to

(DkS)˜maxtzT18eTAe18eTA18Ae18Atz,s.t.tzT012eT12e0tz=2kn,tzTtz=n+1,

which can further be relaxed to the standard semidefinite programming problem

(SDRDkS˜)max18eTAe18eTA18Ae18AS,s.t.012eT12e0S=2kn,IS=n+1,SS1+n+.

Obviously, diag(S)=e implies that iSii=n+1, i.e., IS=n+1, but we could not obtain diag(S)=e from IS=n+1. These results further imply that

Opt(SDRDkS¯)Opt(SDRDkS˜).

Similar to the Theorem 4.2 in [27], we have that the following approximation accuracy theorem holds.

Theorem 6.

Opt(DkS)Opt(DkS)˜Opt(SDRDkS˜)2log(67)Opt(DkS)˜.

Up to now, we not only establish the equivalence between the (DNNPDkS) problem and the (IISDRDkS) problem, but also some approximation accuracy results about the (DNNPDkS) problem and some standard semidefinite relaxation problems are given. In the following Section 5, we will implement some numerical experiments to give a flavour of the actual behaviour of the (DNNPDkS) problem and some semidefinite relaxation problems.

5. Numerical Experiments

In this section, some random (DkS) examples are tested to show the efficiency of the proposed relaxation problems. These relaxation problems are all solved by CVX [28], which is implemented by using MATLAB R2010a on the Windows XP platform, and on a PC with 2.53 GHz CPU. The corresponding comparative numerical results are reported in the following parts.

To give a flavour of the behaviour of the above relaxation problems, we consider results for the following test examples. The data of the test examples are given in Table 1.

Table 1.

Data of the test examples.

P25 n = 25; k = 5; seed = 1, 2, …, 50; rand(‘seed’,seed);
A = round(10 × rand(n)); A = tril(A,−1) + triu(A’,0);
P30 n= 30; k= 10; seed = 2012; randn(‘seed’,seed); A = randn(n,n);
A = tril(A,−1) + triu(A’,0);
P40 n= 40; k= 15; seed = 2017; rand(‘seed’,seed); A = rand(n,n);
A = tril(A,−1) + triu(A’,0);
P50 n= 50; k= 10; seed = 2001, 2002, …, 2050; rand(‘seed’,seed);
G = (rand(n)<0.5); A = round(20 × rand(n)−10); A = triu(A,1);
A = A.*G; A100 = A + A’;
P60 n= 60; k= 20; seed = 2020; rand(‘seed’,seed); A = round(rand(n,n));
A = tril(A,−1) + triu(A’,0);
P80 n= 80; k= 30; seed = 2023; rand(‘seed’,seed); A = round(rand(n,n));
A = tril(A,−1) + triu(A’,0);

The first column of Table 1 denotes the name of the test examples, n and k stand for the number of vertices of the given graph and the finding subgraph, respectively. The last column denotes the procedures for generating the coefficient matrices A in the (DkS) problem. The more detailed explanations for the procedures are given as follows:

• P25. 50 random examples are generated from the ‘seed = 1,2,…,50’. The corresponding coefficient matrices A of order n=25 with integer weights are drawn from {0,1,,10}.

• P30. This example is generated by the MATLAB function randn from the ‘seed = 2012’. The elements of A satisfy the standard normal distribution.

• P40. This example is generated by MATLAB function rand from the ‘seed = 2017’. The elements of A satisfy the standard uniform distribution on the interval (0,1).

• P50. We generate 50 examples of order n=50 from the ‘seed = 2001,2002,…,2050’. For each coefficient matrix of these examples, half of the entries are randomly set to zero, and the remaining ones are chosen from {10,9,,9,10}.

• P60. This example is generated by MATLAB function rand from the ‘seed = 2020’. The elements of A are drawn from {0,1}.

• P80. This example is generated by MATLAB function rand from the ‘seed = 2023’. The coefficient matrix A of order n=80 with integer weights drawn from {0,1,,10}.

First of all, the performances of the (DNNPDkS) problem and the (IISDRDkS) problem as well as the (ISDRDkS) problem, for solving P25 and P50, are compared. We use the performance profiles described in Dolan and Moré’s paper [29]. Our profiles are based on optimal values (i.e., average degree) and the number of iterations of these relaxation problems. The Cumulative Probability denotes the cumulative distribution function for the performance ratio within a factor τR, i.e., is the probability that the solver will win over the rest of the solvers. The corresponding comparative results of performance are shown in Figure 1 and Figure 2.

Figure 1.

Figure 1

The performance of the (DNNPDkS) problem and the (IISDRDkS) problem as well as the (ISDRDkS) problem for solving P25. The performance profile of (left) figure is based on optimal values, and (right) figure is based on the number of iterations.

Figure 2.

Figure 2

The performance of the (DNNPDkS) problem and the (IISDRDkS) problem as well as (ISDRDkS) problem for solving P50. The performance profile in the (left) figure is according to optimal values, and the (right) figure is based on the number of iterations.

The comparative results for P25 are shown in Figure 1. It is obvious that the (DNNPDkS) problem and the (IISDRDkS) problem have the same performance, which is a bit better than that of the (ISDRDkS) problem from the viewpoint of optimal values. In view of the number of iterations, the performance of the (ISDRDkS) problem is the best, and the performance of the (IISDRDkS) problem is better than that of the (DNNPDkS) problem.

The performance of the three relaxation problems for solving P50 is shown in Figure 2. The results show that the performance of the (DNNPDkS) problem is the same as that of the (IISDRDkS) problem; they are both much better than that of the (ISDRDkS) problem in view of optimal values—although the performance of the (ISDRDkS) problem is better than the one of the (DNNPDkS) problem and the (IISDRDkS) problem from the viewpoint of the number of iterations.

All of results show in Figure 1 and Figure 2 further imply that the (DNNPDkS) problem and the (IISDRDkS) problem can generate more promising bounds for solving P25 and P50, compared with the (ISDRDkS) problem, while the number of iterations is a bit more. Moreover, the (DNNPDkS) problem and the (IISDRDkS) problem have the same performance based on optimal values, although the performance of the (IISDRDkS) problem is better than that of the (DNNPDkS) problem from the viewpoint of the number of iterations, for solving P25 and P50.

In order to further show the computational efficiency of the (DNNPDkS) problem, which is compared with the (IISDRDkS) problem and some other types of semidefinite relaxation problems proposed in [30], for solving some (DkS) problems. The test examples A50 and A100 are chosen from [30]. (R-20), (R-24) and (R-MET) denote the three semidefinite relaxation problems proposed in [30], respectively. The corresponding numerical results are shown in Table 2, where “−” means that the corresponding information about the number of iterations is not given in [30]. The results show that the computational efficiency of the (DNNPDkS) problem is better than the one of the (IISDRDkS) problem from the viewpoints of optimal values and number of iterations, respectively. Note that the performance of the (DNNPDkS) problem and the (IISDRDkS) problem are both much better than that of (R20) and (R24). Moreover, the performance of the (DNNPDkS) problem is more competitive with (RMET) for solving these two problems.

Table 2.

Numerical results for some DkS examples in [30].

Example Solving Model Optimal Value Number of Iterations
A50 (DNNPDkS) 337.4 51
(IISDRDkS) 337.416 57
(R20) 365.508
(R24) 365.831
(RMET) 333.65
A100 (DNNPDkS) 505.989 54
(IISDRDkS) 506.126 68
(R20) 607.691
(R24) 611.070
(RMET) 490.50

Finally, we further compare the efficiency of the (DNNPDkS), (IISDRDkS), (ISDRDkS) and (SDRDkS^) problems, for solving examples P30, P40, P60 and P80. The comparative results are shown in Table 3. The results signify that the efficiency of the (DNNPDkS) problem is always better than that of the (IISDRDkS) problem from the viewpoint of optimal values and the number of iterations as well as CPU time, respectively, for solving these examples. The performance of the (ISDRDkS) problem and the (SDRDkS^) problem are almost the same for solving these examples. Moreover, note that the optimal value of the (DNNPDkS) problem for solving P80 is larger than that of the (IISDRDkS) problem. Thus, we can conclude that it may be more promising to use the (DNNPDkS) problem than to use the (IISDRDkS) problem for solving some specific (DkS) problems in practice.

Table 3.

Numerical results for some (DkS) examples.

Example Solving Model Optimal Value Number of Iterations CPU Time (seconds)
P30 (DNNPDkS) 30.5894 52 3.4
(IISDRDkS) 30.5903 70 5.0
(ISDRDkS) 33.3382 33 0.7
(SDRDkS^) 33.3383 34 0.8
P40 (DNNPDkS) 74.4657 50 14.7
(IISDRDkS) 74.4658 66 20.6
(ISDRDkS) 75.5198 33 1.5
(SDRDkS^) 75.5197 34 1.6
P60 (DNNPDkS) 150.385 52 105.5
(IISDRDkS) 150.387 60 125.3
(ISDRDkS) 154.272 40 3.2
(SDRDkS^) 154.273 35 2.2
P80 (DNNPDkS) 230.912 84 739.9
(IISDRDkS) 312.644 85 770.4
(ISDRDkS) 318.945 41 5.7
(SDRDkS^) 318.945 34 3.2

6. Conclusions

In this paper, the DkS problem is studied, whose goal is to find a k-vertex subgraph such that the total weight of edges in this subgraph is maximized. This problem is NP-hard on bipartite graphs, chordal graphs, and planar graphs. By using the advantages of the structure of the DkS problem, the doubly nonnegative relaxation and the new semidefinite relaxation with tighter relaxation for solving the DkS problem are established, respectively. Moreover, we prove that the two relaxation problems are equivalent under the suitable conditions, and give some approximation accuracy results for these relaxation problems. Finally, the comparative numerical results show that the efficiency of the doubly nonnegative relaxation is better than the one of semidefinite relaxation for solving some DkS problems.

Acknowledgments

The authors thank the reviewers for their very helpful suggestions, which led to substantial improvements of the paper.

Abbreviations

The following abbreviations are used in this manuscript:

Rn+ the cone of nonnegative vectors with dimension n
Sn the cone of n×n symmetric matrices
Sn+ the cone of n×n positive semidefinite matrices
Sn++ the cone of n×n positive definite matrices
Nn+ the cone of n×n nonnegative symmetric matrices
e the vector of ones
I the identity matrix with appropriate dimension
X()0 stands for X is positive semidefinite (positive definite), i.e, XSn+(Sn++)
AB trace(ATB) for two conformal matrices A and B
diag(M) the column vector whose elements are the diagonal elements of the given matrix M
Opt()/Fea() the optimal objective value/the set of feasible solutions of the given problem ()

Author Contributions

Methodology, C.-H.G.; software, C.-H.G., Y.G. and B.-B.L.; validation, C.-H.G.; formal analysis, C.-H.G. and B.-B.L.; investigation, C.-H.G. and Y.G.; data curation, C.-H.G. and Y.G.; writing—original draft preparation, C.-H.G.; writing—review and editing, C.-H.G., Y.G. and B.-B.L.; visualization, C.-H.G. and Y.G.; supervision, C.-H.G.; project administration, C.-H.G.; funding acquisition, C.-H.G.

Funding

This work is supported by National Natural Science Foundation of China under Grant No. 11501350 and Science Foundation of Zhejiang Sci-Tech University (ZSTU) under Grant No. 16092098-Y.

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Feige U., Kortsarz G., Peleg D. The Dense k-Subgraph Problem1. Algorithmica. 2001;29:410–421. doi: 10.1007/s004530010050. [DOI] [Google Scholar]
  • 2.Roupin F., Billionnet A. A deterministic approximation algorithm for the densest k-subgraph problem. Int. J. Oper. Res. 2008;3:301–314. doi: 10.1504/IJOR.2008.017534. [DOI] [Google Scholar]
  • 3.Corneil D., Perl Y. Clustering and domination in perfect graphs. Discret. Appl. Math. 1984;9:27–39. doi: 10.1016/0166-218X(84)90088-X. [DOI] [Google Scholar]
  • 4.Keil J., Brecht T. The complexity of clustering in planar graphs. J. Combin. Math. Combin. Comput. 1991;9:155–159. [Google Scholar]
  • 5.Pisinger P. Upper bounds and exact algorithms for dispersion problems. Comput. Oper. Res. 2006;33:1380–1398. doi: 10.1016/j.cor.2004.09.033. [DOI] [Google Scholar]
  • 6.Goldstein D., Langberg M. The dense k-subgraph problem. arXiv. 2009. 0912.5327
  • 7.Backer J., Keil J. Constant factor approximation algorithms for the densest k-subgraph problem on proper interval graphs and bipartite permutation graphs. Inf. Process. Lett. 2010;110:635–638. doi: 10.1016/j.ipl.2010.05.011. [DOI] [Google Scholar]
  • 8.Bhaskara A., Charikar M., Chlamtac E., Feige U., Vijayaraghavan A. Detecting high log-densities: An O (n 1/4) approximation for densest k-subgraph; Proceedings of the 42nd ACM Symposium on Theory of Computing; Cambridge, MA, USA. 6–8 June 2010; pp. 201–210. [Google Scholar]
  • 9.Liazi M., Milis I., Zissimopoulos V. A constant approximation algorithm for the densest k-subgraph problem on chordal graphs. Inf. Process. Lett. 2008;108:29–32. doi: 10.1016/j.ipl.2008.03.016. [DOI] [Google Scholar]
  • 10.Goemans M., Williamson D. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. J. ACM. 1995;42:1115–1145. doi: 10.1145/227683.227684. [DOI] [Google Scholar]
  • 11.Poljak S., Rendl F., Wolkowicz H. A recipe for semidefinite relaxation for (0, 1)-quadratic programming. J. Glob. Optim. 1995;7:51–73. doi: 10.1007/BF01100205. [DOI] [Google Scholar]
  • 12.Ma W., Davidson T., Wong K., Luo Z., Ching P. Quasi-maximum-likelihood multiuser detection using semi-definite relaxation with application to synchronous CDMA. IEEE Trans. Signal Process. 2002;50:912–922. [Google Scholar]
  • 13.Ma W., Su C., Jalden J., Chang T., Chi C. The equivalence of semidefinite relaxation MIMO detectors for higher-order QAM. IEEE J. Sel. Top. Signal Process. 2009;3:1038–1052. doi: 10.1109/JSTSP.2009.2035798. [DOI] [Google Scholar]
  • 14.Lovasz L. Stable sets and polynomials. Discret. Math. 1994;124:137–153. doi: 10.1016/0012-365X(92)00057-X. [DOI] [Google Scholar]
  • 15.Burer S. On the copositive representation of binary and continuous nonconvex quadratic programs. Math. Program. 2009;120:479–495. doi: 10.1007/s10107-008-0223-z. [DOI] [Google Scholar]
  • 16.Bai Y., Guo C. Doubly nonnegative relaxation method for solving multiple objective quadratic programming problems. J. Ind. Manag. Optim. 2014;10:543–556. doi: 10.3934/jimo.2014.10.543. [DOI] [Google Scholar]
  • 17.Burer S. Optimizing a polyhedral-semidefinite relaxation of completely positive programs. Math. Program Comput. 2010;2:1–19. doi: 10.1007/s12532-010-0010-8. [DOI] [Google Scholar]
  • 18.Arima N., Kim S., Kojima M.T.K. Lagrangian-Conic relaxations, part I: A unified framework and its applications to quadratic optimization problems. Pac. J. Optim. 2018;14:161–192. [Google Scholar]
  • 19.Bomze I. Copositive optimization-Recent developments and applications. Eur. J. Oper. Res. 2012;216:509–520. doi: 10.1016/j.ejor.2011.04.026. [DOI] [Google Scholar]
  • 20.Pardalos P., Vavasis S. Quadratic programming with one negative eigenvalue is NP-hard. J. Glob. Optim. 1991;1:15–22. doi: 10.1007/BF00120662. [DOI] [Google Scholar]
  • 21.Dickinson P., Gijben L. On the Computational Complexity of Membership Problems for the Completely Positive Cone and its Dual. University of Groningen; Groningen, The Netherlands: 2011. [Google Scholar]
  • 22.Diananda P. Mathematical Proceedings of the Cambridge Philosophical Society. Cambridge University Press; Cambridge, UK: 1962. On non-negative forms in real variables some or all of which are non-negative. [Google Scholar]
  • 23.Berman A., Shaked-Monderer N. Completely Positive Matrices. World Scientific; Singapore: 2003. [Google Scholar]
  • 24.Boyd S., Vandenberghe L. Convex Optimization. Cambridge University Press; Cambridge, UK: 2004. [Google Scholar]
  • 25.Ge D., Ye Y. On Doubly Positive Semidefinite Programming Relaxations. [(accessed on 17 August 2010)];2010 Available online: http://www.optimization-online.org/DB_HTML/2010/08/2709.pdf.
  • 26.Luo Z., Ma W., So A.C., Ye Y., Zhang S. Semidefinite relaxation of quadratic optimization problems. Signal Proc. Mag. IEEE. 2010;27:20–34. doi: 10.1109/MSP.2010.936019. [DOI] [Google Scholar]
  • 27.He S., Luo Z., Nie J., Zhang S. Semidefinite relaxation bounds for indefinite homogeneous quadratic optimization. SIAM J. Optim. 2008;19:503–523. doi: 10.1137/070679041. [DOI] [Google Scholar]
  • 28.Grant M., Boyd S. CVX: Matlab Software for Disciplined Convex Programming, Version 1.21. [(accessed on 15 April 2010)]; Available online: http://cvxr.com/cvx.
  • 29.Dolan E., Moré J. Benchmarking optimization software with performance profiles. Math. Program. 2002;91:201–213. doi: 10.1007/s101070100263. [DOI] [Google Scholar]
  • 30.Rendl F. Semidefinite relaxations for partitioning, assignment and ordering problems. Q. J. Oper. Res. 2012;10:321–346. doi: 10.1007/s10288-012-0210-3. [DOI] [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES