Skip to main content
Springer logoLink to Springer
. 2018 Jan 30;2018(1):26. doi: 10.1186/s13660-018-1619-9

An alternative extragradient projection method for quasi-equilibrium problems

Haibin Chen 1,, Yiju Wang 1, Yi Xu 2
PMCID: PMC5790870  PMID: 29416291

Abstract

For the quasi-equilibrium problem where the players’ costs and their strategies both depend on the rival’s decisions, an alternative extragradient projection method for solving it is designed. Different from the classical extragradient projection method whose generated sequence has the contraction property with respect to the solution set, the newly designed method possesses an expansion property with respect to a given initial point. The global convergence of the method is established under the assumptions of pseudomonotonicity of the equilibrium function and of continuity of the underlying multi-valued mapping. Furthermore, we show that the generated sequence converges to the nearest point in the solution set to the initial point. Numerical experiments show the efficiency of the method.

Keywords: Quasi-equilibrium problems, Extragradient projection method, Pseudomonotonicity, Multi-valued mapping

Introduction

The equilibrium problem has been considered as an important and general framework for describing various problems arising in different areas of mathematics, including optimization problems, mathematical economic problems and Nash equilibrium problems. As far as we know, this formulation has been followed for a long time by several studies on equilibrium problems considered under different headings such as quasi-equilibrium problem, mixed equilibrium problem, ordered equilibrium problem, vector equilibrium problem and so on [14]. It should be noted that one of the interests of this common formulation is that many techniques developed for a particular case may be extended, with suitable adaptations, to the equilibrium problem, and then they can be applied to other particular cases [516]. In this paper, we mainly deal with existence of solutions and approximate solutions of the quasi-equilibrium problem.

Let XRn be a nonempty closed convex set, K be a point-to-set mapping from X onto itself such that, for any xX, K(x) is a nonempty closed convex set of X, and let f:X×XR be a function such that, for any xX, f(x,x)=0 and f(x,) is convex on X. The quasi-equilibrium problem QEP(K,f) is to find a vector xK(x) such that

f(x,y)0,yK(x). 1.1

Throughout this paper, we denote the solution set by K.

Certainly, when f(x,y)=F(x),yx with F being a vector-valued mapping from X to Rn, then the quasi-equilibrium problem reduces to the generalized variational inequality or quasi-variational inequality problem [1720] which is to find vector xK(x) such that

F(x),yx0,yK(x). 1.2

To move on, we recall the classical equilibrium problem, and classical Nash equilibrium problem (NEP) [21]. Assume the function fi:RnR is continuous, and suppose Ki is a nonempty closed set in Rni for i=1,2,,N with n=i=1Nni. Suppose that there are N players such that each player controls the variables xiRni. Denote x=(x1,,xN), and xi=(x1,,xi1,xi+1,,xN). Player i needs to take an xiKiRni that solves the following optimization problem:

minxiKifi(xi,xi)

based on the other players’ strategies xi. If these N players do not cooperate, then each players’ strategy set may vary with other players’ strategies, that is, the ith player’s strategy set varies with the other players’ strategies xi. In this case, we need to use Ki(xi) instead of Ki to indicate the ith player’s strategy set, and the ith player needs to choose a strategy xiKi(xi) that solves the following optimization problem:

minxiKi(xi)fi(xi,xi).

In [22], the non-cooperative game model is called generalized Nash equilibrium problem (GNEP), which can be formulated as the quasi-equilibrium problem where the involved functions are nondifferentiable [23].

For the problem GNEP, when the functions fi(,xi) are convex and differentiable, then the problem can be equivalently formulated as the quasi-variational inequalities (1.2) by setting

F(x)=(xifi(x))i=1N

and K(x)=i=1NKi(xi). When the function fi(,xi) is convex and nondifferentiable, then the GNEP reduces to the quasi-equilibrium problem (1.1) [24] via the Nikaido Isoda funtion

f(x,y)=i=1N[fi(yi,xi)fi(xi,xi)].

On the other hand, the quasi-equilibrium problem (QEP) has received much attention of researchers in mathematics, economics, engineering, operations research, etc. [17, 22]; for more information see [19, 25, 26]. There are many solution methods for solving QEP. Recently, [27] considered an optimization reformulation approach with the regularized gap function. Different from the variational inequalities problem, the regularized gap function is in general not differentiable, but only directional differentiable. Furthermore, supplementary conditions must be imposed to guarantee that any stationary point of these functions is a global minimum, since the gap functions is nonconvex [28]. It should be noted that such conditions are known for variational inequality problem but not for QEP. So, [23] proposed several projection and extragradient methods rather than methods based on gap functions, which generalized the double-projection methods for variational inequality problem to equilibrium problems with a moving constraint set K(x).

It is well known that the extragradient projection method is an efficient solution method for variational inequalities due to its low memory and low cost of computing [29, 30]. Based on those advantages, it was extended to solve QEP recently [20, 23, 31] and this opened a new approach for solving the problem. An important feature of this method is that it has the contraction property in the sense that the generated sequence has contraction property with respect to the solution set of the problem [29], i.e.

xk+1xxkx,k0,xK.

The numerical experiments given in [20, 23, 31] show that the extragradient projection method is a practical solution method for the QEP.

It should be noted that not all the extragradient projection methods have the contraction property [32]. In that case, it may not slow down the convergence rate significantly. Although the extragradient projection method has no contraction property, it still has a good numerical performance [32]. Now, a question can be posed naturally, can this method be applied to solve the QEP? And if so, how about its performance? This constitutes the main motivation of this paper.

Inspired by the work [23, 32], we propose a new type of extragradient projection method for the QEP in this paper. Different from the extragradient projection method proposed in [23], the generated sequence by the newly designed method possesses an expansion property with respect to the initial point, i.e.,

xk+1xk2+xkx02xk+1x02.

The existence results for (1.1) are established under pseudomonotonicity condition of the equilibrium function and the continuity of the underlying multi-valued mapping. Furthermore, we show that the generated sequence converges to the nearest point in the solution set to the initial point. Numerical experiments show the efficiency of the method.

The remainder of this paper is organized as follows. In Section 2, we recall some concepts and related conclusions to be used in the sequel. The newly designed method and its global convergence are developed in Section 3. Some preliminary computational results and experiments are presented in Section 4.

Preliminaries

Let X be a nonempty closed convex set in Rn. For any xRn, the orthogonal projection of x onto X is defined as

y0=argmin{yxyX},

and denote PX(x)=y0. A basic property of the projection operator is as follows [33].

Lemma 2.1

Suppose X is a nonempty, closed and convex subset in Rn. For any x,yRn, and zX, we have

  • (i)

    PX(x)x,zPX(x)0;

  • (ii)

    PX(x)PX(y)2xy2PX(x)x+yPX(y)2.

Remark 2.1

The first statement in Lemma 2.1 also provides a sufficient condition for vector yX to be the projection of vector x, i.e., y=PX(x) if and only if

yx,zy0,zX.

To proceed, we present the following definitions [34].

Definition 2.1

Suppose X is a nonempty subset of Rn. The bifunction f:X×XR is said to be

  • (i)
    strongly monotone on X with β>0 iff
    f(x,y)+f(y,x)βxy2,x,yX;
  • (ii)
    monotone on X iff
    f(x,y)+f(y,x)0,x,yX;
  • (iii)
    pseudomonotone on X iff
    f(x,y)0f(y,x)0,x,yX.

Definition 2.2

Suppose X is a nonempty, closed and convex subset of Rn. A multi-valued mapping K:X2Rn is said to be

  • (i)

    upper semicontinuous at xX if for any convergent sequence {xk}X with being the limit, and for any convergent sequence {yk} with ykK(xk) and ȳ being the limit, then y¯K(x¯);

  • (ii)

    lower semicontinuous at xX if given any sequence {xk} converging to x and any yK(x), there exists a sequence {yk} with ykK(xk) converges to y;

  • (iii)

    continuous at xX if it is both upper semicontinuous and lower semicontinuous at x.

To end this section, we make the following blanket assumption on bifunction f:X×XR and multi-valued mapping K:X2Rn [20, 23].

Assumption 2.1

For the closed convex set XRn, the bifunction f and multi-valued mapping K satisfy:

  • (i)

    f(x,) is convex for any fixed xX, f is continuous on X×X and f(x,x)=0 for all xX;

  • (ii)

    K is continuous on X and K(x) is a nonempty closed convex subset of X for all xX;

  • (iii)

    xK(x) for all xX;

  • (iv)

    S={xSf(x,y)0,yT} is nonempty for S=xXK(x) and T=xXK(x);

  • (v)

    f is pseudomonotone on X with respect to S i.e. f(x,y)0f(y,x)0, xS, yX.

As noted in [23], the assumption (iv) in Assumption 2.1 guarantees that the solution set of problem (1.1) K is nonempty.

Algorithm and convergence

In this section, we mainly develop a new type of extragradient projection method for solving QEP. The basic idea of the algorithm is as follows. At each step of the algorithm, we obtain a solution yk by solving a convex subproblem. If xk=yk, then stop with xk being a solution of the QEP; otherwise, find a trial point zk by a back-tracking search at xk along the direction xkyk, and the new iterate is obtained by projecting x0 onto the intersection of K(xk) of two halfspaces which are, respectively, associated with zk and xk. Repeat the process until xk=yk. The detailed description of our designed algorithm is as follows.

Algorithm 3.1

Step 0.

Choose c,γ(0,1), x0X, k=0.

Step 1.
For current iterate xk, compute yk by solving the following optimization problem:
minyK(xk){f(xk,y)+12yxk2}.
If xk=yk, then stop. Otherwise, let zk=(1ηk)xk+ηkyk, where ηk=γmk with mk being the smallest nonnegative integer such that
f((1γm)xk+γmyk,xk)f((1γm)xk+γmyk,yk)cxkyk2. 3.1
Step 2.
Compute xk+1=PHk1Hk2X(x0) where
Hk1={xRnf(zk,x)0},Hk2={xRnxxk,x0xk0}.
Set k=k+1 and go to Step 1.

Indeed, for every xkK(xk), since ykK(xk), zkK(xk), so we have K(xk)Hk1 and K(xk)Hk2. To establish the convergence of the algorithm, we first discuss the relationship of the halfspace Hk1 with xk and the solution set K.

Lemma 3.1

If xkyk, then the halfspace Hk1 in Algorithm 3.1 separates the point xk from the set K under Assumption 2.1. Moreover,

KHk1X,k0.

Proof

First, by the fact that f(x,) is convex and

zk=(1ηk)xk+ηkyk,

we obtain

0=f(zk,zk)(1ηk)f(zk,xk)+ηkf(zk,yk),

which can be written as

f(zk,yk)(1ηk1)f(zk,xk).

By (3.1), we have

f(zk,xk)cηkxkyk2>0,

which means xkHk1.

On the other hand, by Assumption 2.1, it follows that K is nonempty. For any xK, from the definition of K and the pseudomonotone property of f, one has

f(zk,x)0,

which implies that the curve Hk1 separates the point xk from the set X. Furthermore, by the definition of K, it is easy to see that

KHk1X,k0,

and the desired result follows. □

The justification of the termination criterion can be seen from Proposition 2 in [23], and the feasibility of the stepsize rule (3.1), i.e., the existence of point mk can be guaranteed from Proposition 7 in [23].

Next, to show that the algorithm is well defined, we will show that the nonempty set K is always contained in Hk1Hk2X for the projection step.

Lemma 3.2

Let Assumption 2.1 be true. Then we have KHk1Hk2X for all k0.

Proof

From the analysis in Lemma 3.1, it is sufficient to prove that KHk2 for all k0. By induction, if k=0, it is obvious that

KH02=Rn.

Suppose that

KHk2

holds for k=l0. Then

KHl1Hl2X.

For any xK, by Lemma 2.1 and the fact that

xl+1=PHl1Hl2X(x0),

we know that

xxl+1,x0xl+10.

Thus KHl+12, which means that KHk2 for all k0 and the desired result follows. □

In the following, we show the expansion property of the algorithm with respect to the initial point.

Lemma 3.3

Suppose {xk} is the generated sequence of Algorithm 3.1, we have

xk+1xk2+xkx02xk+1x02.

Proof

By Algorithm 3.1, one has

xk+1=PHk1Hk2X(x0).

So xk+1Hk2 and

PHk2(xk+1)=xk+1.

By the definition of Hk2, we have

zxk,x0xk0,zHk2.

Thus, xk=PHk2(x0) from the Remark 2.1. Then, from Lemma 2.1, we obtain

PHk2(xk+1)PHk2(x0)xk+1x02PHk2(xk+1)xk+1+x0PHk2(x0)2,

which can be written as

xk+1xk2xk+1x02xkx02,

i.e.,

xk+1xk2+xkx02xk+1x02,

and the proof is completed. □

To prove the boundedness of the generated sequence {xk}, we assume that the algorithm generates an infinite sequence for simple.

Lemma 3.4

Suppose Assumption 2.1 is true. Then the generated sequence {xk} of Algorithm 3.1 is bounded.

Proof

By Assumption 2.1, we know that K. Since xk+1 is the projection of x0 onto Hk1Hk2X, by Lemma 3.2 and the definition of projection, we have

xk+1x0xx0,xK.

So, {xk} is a bounded sequence. □

Since {xk} is bounded, it has an accumulation point. Without loss of generality, assume that the subsequence {xkj} converges to . Then the sequences {ykj}, {zkj} and {gkj} are bounded from the Proposition 10 in [23], where gkjf(zkj,xkj).

Before given the next result, the following lemma is needed (for details see [23]).

Lemma 3.5

For every yK(xk), we have

f(xk,y)f(xk,yk)+xkyk,yyk.

In particular, f(xk,yk)+xkyk20.

Lemma 3.6

Suppose {xkj} is the sequence presented as in Lemma 3.4. If xkjykj, then

xkjykj0

as j.

Proof

We distinguish for the proof two cases.

(1) If lim infkηk>0, by Lemma 3.4, one has

xk+1xk2+xkx02xk+1x02.

Thus, the sequence {xkx0} is nondecreasing and bounded, and hence convergent, which implies that

limkxk+1xk=0.

On the other hand, by Assumption 2.1(i) and the fact that

zkj=(1ηkj)xkj+ηkjykj,

we have

0=f(zkj,zkj)(1ηkj)f(zkj,xkj)+ηkjf(zkj,ykj),

which can be written as

f(zkj,ykj)(1ηkj1)f(zkj,xkj).

By (3.1), one has

f(zkj,xkj)cηkjxkjykj2>0.

Then we will prove that

PHkj1(xkj)=xkjf(zkj,xkj)gkj2gkj, 3.2

where gkjf(zkj,xkj). For all xHkj1, from the remark of Lemma 2.1, we only need to prove

xkj(xkjf(zkj,xkj)gkj2gkj),(xkjf(zkj,xkj)gkj2gkj)x0,

i.e.,

f(zkj,xkj)gkj2gkj,xkjxf2(zkj,xkj)gkj20,

which is equivalent to

gkj,xxkj+f(zkj,xkj)0. 3.3

Since gkjf(zkj,xkj), by the definition of subdifferential we have

f(zkj,x)f(zkj,xkj)+gkj,xxkj,xRn.

So, from the definition of Hkj1, for all xHkj1 we have

f(zkj,x)0,

which implies that (3.3) holds. Moreover, (3.2) is right.

By (3.2) and the fact that there is a constant M>0 such that gkjM, we obtain

xkjxkj+1xkjPHkj1(xkj)=f(zkj,xkj)gkjcηkjMxkjykj2,

which implies that xkjykj0, j, and the desired result holds.

(2) Suppose that lim infkηk=0, and for any subsequence {ηkj} of {ηk}, it satisfies

limjηkj=0.

Let {xkj}x¯ as j, it follows that

z¯kj=(1ηkjγ)xkj+ηkjγykjx¯,j.

By the definition of {ηkj}, one has

f(z¯kj,xkj)f(z¯kj,ykj)<cxkjykj2.

Let ȳ be the limit of {ykj}. By Lemma 3.5 we have

f(z¯kj,xkj)f(z¯kj,ykj)<cxkjykj2cf(xkj,ykj).

Taking j and remembering the fact that f is continuous, we obtain

f(x¯,x¯)f(x¯,y¯)cf(x¯,y¯),

which implies that f(x¯,y¯)0. So xkjykj0, j and the desired result follows. □

Based on the analysis above, we can establish the main results of this section that the generated sequence {xk} globally converge to a solution of the problem (1.1).

Theorem 3.1

Suppose {xk} is an infinite sequence generated in Algorithm 3.1. Let conditions of Lemma 3.6 be true. Then each accumulation point of {xk} is a solution of the QEP under the Assumption 2.1.

Proof

By Lemma 3.4, without loss of generality, assume that the subsequence {xkj} converges to . By Lemma 3.6, one has xkjykj0 and

ykj=ykjxkj+xkjx¯,

where ykjK(xkj) for every j. Thus x¯K(x¯) from the fact that K is upper semicontinuous.

To prove that is a solution of the problem (1.1), since

yk=argminyK(xk)[f(xk,y)+12yxk2],

the optimality condition implies that there exists ωf(xk,yk) such that

0=ω+ykxk+sk,

where skNK(xk)(yk) is a vector in the normal cone to K(xk) at yk. Then we have

ykxk,yykω,yky,yK(xk). 3.4

On the other hand, since ωf(xk,yk) and by the well-known Moreau–Rockafellar theorem [35], one has

f(xk,y)f(xk,yk)ω,yyk. 3.5

By (3.4) and (3.5), we have

f(xk,y)f(xk,yk)xkyk,yyk,yK(xk). 3.6

Letting k=kj in (3.6)

f(xkj,y)f(xkj,ykj)xkjykj,yykj,yK(xkj).

Taking j and remembering that f is continuous, we obtain

f(x¯,y)0,yK(x¯),

that is, is a solution of the QEP and the proof is completed. □

Theorem 3.2

Under the assumption of Theorem 3.1, the generated sequence {xk} converges to a solution x such that

x=PK(x0)

under the Assumption 2.1.

Proof

By Theorem 3.1, we know that the sequence {xk} is bounded and that every accumulation point x of {xk} is a solution of the problem (1.1). Let {xkj} be a convergent subsequence of {xk}, and let xK be its limit. Let x¯=PK(x0). Then by Lemma 3.2,

x¯Hkj11Hkj12X,

for all j. So, from the iterative procedure of Algorithm 3.1,

xkj=PHkj11Hkj12X(x0),

one has

xkjx0x¯x0. 3.7

Thus,

xkjx¯2=xkjx0+x0x¯2=xkjx02+x0x¯2+2xkjx0,x0x¯x¯x02+x0x¯2+2xkjx0,x0x¯,

where the inequality follows from (3.7). Letting j, it follows that

lim supjxkjx¯22x¯x02+2xx0,x0x¯=2xx¯,x0x¯. 3.8

Due to Lemma 2.1 and the fact that x¯=PK(x0) and xK, we have

xx¯,x0x¯0.

Combining this with (3.8) and the fact that x is the limit of {xkj}, we conclude that the sequence {xkj} converges to and

x=x¯=PK(x0).

Since x was taken as an arbitrary accumulation point of {xk}, it follows that is the unique accumulation point of this sequence. Since {xk} is bounded, the whole sequence {xk} converges to . □

Numerical experiment

In this section, we will make some numerical experiments and give a numerical comparison with the method proposed in [23] to test the efficiency of the proposed method. The MATLAB codes are run on a PIV 2.0 GHz personal computer under MATLAB version 7.0.1.24704(R14). In the following, ‘Iter.’ denotes the number of iteration, and ‘CPU’ denotes the running time in seconds. The tolerance ε means the iterative procedure terminates when xkykε.

Example 4.1

The bifunction f of the quasi-equilibrium problem is defined for each x,yR5 by

f(x,y)=Px+Qy+q,yx,

where q, P, Q are chosen as follows:

q=[12121];P=[3.1200023.6000003.5200023.3000003];Q=[1.6100011.6000001.5100011.5000002].

The moving set K(x)=1i5Ki(x) where for each xR5 and each i, the set Ki(x) is defined by

Ki(x)={yiRyi+1j5,jixj1}.

This problem was tested in [36] with initial point x0=(1,3,1,1,2)T. They obtained the appropriate solution after 21 iterates with the tolerance ε=103.

By the algorithm proposed in this paper, the numerical results obtained for this example are listed in Table 1 with c=γ=0.5, ε=103 and X=K(x), and with different initial points.

Table 1.

Numerical results for Example 4.1

Initial point x0
(0,0,0,0,0)T (1,3,1,1,2)T (1,1,1,1,2)T (1,0,1,0,2)T (0,1,1,0,2)T
Iter. 5 13 9 7 8
CPU 0.2060 0.5340 0.3590 0.2190 0.3430

Now, we consider a quasi-variational inequality problems and we solve it by using Algorithm 3.1 with the equilibrium function f(x,y)=F(x),yx.

Example 4.2

Consider a two-person game whose QVI formulation involves the function F=(F1,F2) and the multi-valued mapping K(x)=K1(x2)×K2(x1) for each x=(x1,x2)R2, where

F1(x)=2x1+83x234,F2(x)=2x2+54x124.25,

and

K1(x2)={y1R0y110,y115x2},K2(x1)={y2R0y210,y215x1}.

This problem was tested in [23]. The numerical results of Algorithm 3.1, abbreviated as Alg. 31, for this example are shown in Table 2 with different initial points.

Table 2.

Numerical results for Example 4.2

Alg.31 Initial point x0
(10,10) (10,0) (9,1) (9,3) (9,9) (8,10)
Iter. 53 2 2 8 51 48
CPU(s) 0.7030 0.0160 0.0310 0.2030 0.8440 0.7810

For this example, we choose X=K(x) and take c=γ=0.5. During the experiments, we set the stopping criterion ε=106. The numerical comparison of our proposed method with the algorithms, i.e., Alg.1, Alg.1a, Alg.1b, proposed in [23] are given in Tables 3 and 4.

Table 3.

Iterations from Alg.31, Alg.1, Alg.1a and Alg.1b respectively

Initial point Number of iterations
Alg.31 Alg.1 Alg.1a Alg.1b
(10,0) 2 3 2 2
(10,10) 53 255 120 120

Table 4.

The CPU time from Alg.31, Alg.1, Alg.1a and Alg.1b respectively

Initial point CPU (s)
Alg.31 Alg.1 Alg.1a Alg.1b
(10,0) 0.02 0.26 0.20 0.15
(10,10) 0.70 8.43 3.70 2.57

Conclusions

In this paper, we have proposed a new type of extragradient projection method for the quasi-equilibrium problem. The generated sequence by the newly designed method possesses an expansion property with respect to the initial point. The existence results of the problem is established under pseudomonotonicity condition of the equilibrium function and the continuity of the underlying multi-valued mapping. Furthermore, we have shown that the generated sequence converges to the nearest point in the solution set to the initial point. The given numerical experiments show the efficiency of the proposed method.

Acknowledgements

This work was supported by the National Natural Science Foundation of China (Grant No. 11601261,11671228), and Natural Science Foundation of Shandong Province (Grant No. ZR2016AQ12), and China Postdoctoral Science Foundation (Grant No. 2017M622163).

Authors’ contributions

All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.

Competing interests

The authors declare that they have no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Haibin Chen, Email: chenhaibin508@163.com.

Yiju Wang, Email: wyiju@hotmail.com.

Yi Xu, Email: yi.xu1983@hotmail.com.

References

  • 1.Cho S.Y. Generalized mixed equilibrium and fixed point problems in a Banach space. J. Nonlinear Sci. Appl. 2016;9:1083–1092. doi: 10.22436/jnsa.009.03.36. [DOI] [Google Scholar]
  • 2.Huang N., Long X., Zhao C. Well-posedness for vector quasi-equilibrium problems with applications. J. Ind. Manag. Optim. 2009;5(2):341–349. doi: 10.3934/jimo.2009.5.341. [DOI] [Google Scholar]
  • 3.Li J. Constrained ordered equilibrium problems and applications. J. Nonlinear Var. Anal. 2017;1:357–365. [Google Scholar]
  • 4.Su T.V. A new optimality condition for weakly efficient solutions of convex vector equilibrium problems with constraints. J. Nonlinear Funct. Anal. 2017;2017:7. [Google Scholar]
  • 5.Chen H. A new extra-gradient method for generalized variational inequality in Euclidean space. Fixed Point Theory Appl. 2013;2013:139. doi: 10.1186/1687-1812-2013-139. [DOI] [Google Scholar]
  • 6.Chen H., Wang Y., Zhao H. Finite convergence of a projected proximal point algorithm for generalized variational inequalities. Oper. Res. Lett. 2012;40:303–305. doi: 10.1016/j.orl.2012.03.011. [DOI] [Google Scholar]
  • 7.Chen H., Wang Y., Wang G. Strong convergence of extra-gradient method for generalized variational inequalities in Hilbert space. J. Inequal. Appl. 2014;2014:223. doi: 10.1186/1029-242X-2014-223. [DOI] [Google Scholar]
  • 8.Qin X., Yao J.C. Projection splitting algorithms for nonself operators. J. Nonlinear Convex Anal. 2017;18(5):925–935. [Google Scholar]
  • 9.Xiao Y.B., Huang N.J., Cho Y.J. A class of generalized evolution variational inequalities in Banach spaces. Appl. Math. Lett. 2012;25(6):914–920. doi: 10.1016/j.aml.2011.10.035. [DOI] [Google Scholar]
  • 10.Chen H.B., Qi L.Q., Song Y.S. Column sufficient tensors and tensor complementarity problems. Front. Math. China. 2018 [Google Scholar]
  • 11.Chen H.B., Wang Y.J. A family of higher-order convergent iterative methods for computing the Moore–Penrose inverse. Appl. Math. Comput. 2011;218:4012–4016. [Google Scholar]
  • 12.Sun H.C., Wang Y.J., Qi L.Q. Global error bound for the generalized linear complementarity problem over a polyhedral cone. J. Optim. Theory Appl. 2009;142:417–429. doi: 10.1007/s10957-009-9509-4. [DOI] [Google Scholar]
  • 13.Wang Y.J., Liu W.Q., Cacceta L., Zhou G.L. Parameter selection for nonnegative l1 matrix/tensor sparse decomposition. Oper. Res. Lett. 2015;43:423–426. doi: 10.1016/j.orl.2015.06.005. [DOI] [Google Scholar]
  • 14.Wang Y.J., Cacceta L., Zhou G.L. Convergence analysis of a block improvement method for polynomial optimization over unit spheres. Numer. Linear Algebra Appl. 2015;22:1059–1076. doi: 10.1002/nla.1996. [DOI] [Google Scholar]
  • 15.Wang C.W., Wang Y.J. A superlinearly convergent projection method for constrained systems of nonlinear equations. J. Glob. Optim. 2009;40:283–296. doi: 10.1007/s10898-008-9324-8. [DOI] [Google Scholar]
  • 16.Chen H.B., Chen Y.N., Li G.Y., Qi L.Q. A semidefinite program approach for computing the maximum eigenvalue of a class of structured tensors and its applications in hypergraphs and copositivity test. Numer. Linear Algebra Appl. 2018;25(6):e2125. doi: 10.1002/nla.2125. [DOI] [Google Scholar]
  • 17.Facchinei F., Pang J.-S. Finite-Dimensional Variational Inequalities and Complementarity Problems. New York: Springer; 2003. [Google Scholar]
  • 18.Harker P.T. Generalized Nash games and quasi-variational inequalities. Eur. J. Oper. Res. 1991;54:81–94. doi: 10.1016/0377-2217(91)90325-P. [DOI] [Google Scholar]
  • 19.Pang J.-S., Fukushima M. Quasi-variational inequalities, generalized Nash equilibria, and multi-leaderfollower games. Comput. Manag. Sci. 2005;2:21–56. doi: 10.1007/s10287-004-0010-0. [DOI] [Google Scholar]
  • 20.Zhang J., Qu B., Xiu N. Some projection-like methods for the generalized Nash equilibria. Comput. Optim. Appl. 2010;45:89–109. doi: 10.1007/s10589-008-9173-x. [DOI] [Google Scholar]
  • 21.Nash J. Non-cooperative games. Ann. Math. 1951;54:286–295. doi: 10.2307/1969529. [DOI] [Google Scholar]
  • 22.Facchinei F., Kanzow C. Generalized Nash equilibrium problems. Ann. Oper. Res. 2010;175:177–211. doi: 10.1007/s10479-009-0653-x. [DOI] [Google Scholar]
  • 23.Strodiot J.J., Nguyen T.T.V., Nguyen V.H. A new class of hybrid extragradient algorithms for solving quasi-equilibrium problems. J. Glob. Optim. 2013;56:373–397. doi: 10.1007/s10898-011-9814-y. [DOI] [Google Scholar]
  • 24.Blum E., Oettli W. From optimization and variational inequality to equilibrium problems. Math. Stud. 1994;63:127–149. [Google Scholar]
  • 25.Pang J.-S., Fukushima M. Quasi-variational inequalities, generalized Nash equilibria, and multi-leaderfollower games. Erratum. Comput. Manag. Sci. 2009;6:373–375. doi: 10.1007/s10287-009-0093-8. [DOI] [Google Scholar]
  • 26.Pham H.S., Le A.T., Nguyen B.M. Approximate duality for vector quasi-equilibrium problems and applications. Nonlinear Anal., Theory Methods Appl. 2010;72(11):3994–4004. doi: 10.1016/j.na.2010.01.031. [DOI] [Google Scholar]
  • 27.Taji K. On gap functions for quasi-variational inequalities. Abstr. Appl. Anal. 2008;2008:531361. doi: 10.1155/2008/531361. [DOI] [Google Scholar]
  • 28.Kubota K., Fukushima M. Gap function approach to the generalized Nash equilibrium problem. J. Optim. Theory Appl. 2010;144:511–531. doi: 10.1007/s10957-009-9614-4. [DOI] [Google Scholar]
  • 29.He B.S. A class of projection and contraction methods for monotone variational inequalities. Appl. Math. Optim. 1997;35(1):69–76. doi: 10.1007/BF02683320. [DOI] [Google Scholar]
  • 30.Iusem A.N., Svaiter B.F. A variant of Korpelevich’s method for variational inequalities with a new search strategy. Optimization. 1997;42:309–321. doi: 10.1080/02331939708844365. [DOI] [Google Scholar]
  • 31.Han D.R., Zhang H.C., Qian G., Xu L.L. An improved two-step method for solving generalized Nash equilibrium problems. Eur. J. Oper. Res. 2012;216(3):613–623. doi: 10.1016/j.ejor.2011.08.008. [DOI] [Google Scholar]
  • 32.Wang Y.J., Xiu N.H., Zhang J.Z. Modified extragradient method for variational inequalities and verification of solution existence. J. Optim. Theory Appl. 2003;119:167–183. doi: 10.1023/B:JOTA.0000005047.30026.b8. [DOI] [Google Scholar]
  • 33.Zarantonello E.H. Contributions to Nonlinear Functional Analysis. New York: Academic Press; 1971. Projections on convex sets in Hilbert space and spectral theory. [Google Scholar]
  • 34.Konnov I.V. Combined Relaxation Methods for Variational Inequalities. Berlin: Springer; 2001. [Google Scholar]
  • 35.Rockafellar R.T. Monotone operators and the proximal point algorithm. SIAM J. Control Optim. 1976;14(5):877–898. doi: 10.1137/0314056. [DOI] [Google Scholar]
  • 36.Tran D.Q., LeDung M., Nguyen V.H. Extragradient algorithms extended to equilibrium problems. Optimization. 2008;57:749–776. doi: 10.1080/02331930601122876. [DOI] [Google Scholar]

Articles from Journal of Inequalities and Applications are provided here courtesy of Springer

RESOURCES