Skip to main content
Springer logoLink to Springer
. 2021 Aug 19;80(2):321–346. doi: 10.1007/s10589-021-00305-3

Forward-reflected-backward method with variance reduction

Ahmet Alacaoglu 1,, Yura Malitsky 2, Volkan Cevher 1
PMCID: PMC8550342  PMID: 34720428

Abstract

We propose a variance reduced algorithm for solving monotone variational inequalities. Without assuming strong monotonicity, cocoercivity, or boundedness of the domain, we prove almost sure convergence of the iterates generated by the algorithm to a solution. In the monotone case, the ergodic average converges with the optimal O(1/k) rate of convergence. When strong monotonicity is assumed, the algorithm converges linearly, without requiring the knowledge of strong monotonicity constant. We finalize with extensions and applications of our results to monotone inclusions, a class of non-monotone variational inequalities and Bregman projections.

Keywords: Variational inequalities, Stochastic variance reduction, Finite-sum structure, Saddle point problems, Monotone inclusions

Introduction

We are interested in solving variational inequalities (VI)

FindzZ:F(z),z-z+g(z)-g(z)0,zZ, 1

where g is a proper lower semicontinuous convex function and F is a monotone operator also given as the finite sum F=1ni=1nFi.

A special case of monotone VIs is the structural saddle point problem

minxmaxyΨ(x,y)+f(x)-h(y), 2

where f, h are proper lower semicontinuous convex functions and Ψ is a smooth convex-concave function. Indeed, problem (2) can be formulated as (1) by setting

z=(x,y),F(z)=xΨ(x,y)-yΨ(x,y),g(z)=f(x)+h(y),

and F(z)=1ni=1nFi(z) (see [2, Section 2], [5, 7] for examples).

Another related problem is the monotone inclusion where the aim is to

findzZsuch that0(A+F)(x),

where A:ZZ and F:ZZ are maximally monotone operators and F is Lipschitz continuous with finite sum form. Monotone inclusions generalize (1) and our results also extend to this setting as will be shown in Sect. 4.1. Due to convenient abstraction, it is the problem (1) that will be our main concern.

The case when Ψ in (2) is convex-concave and, in particular when it is bilinear, has found numerous applications in machine learning, image processing and operations research, resulting in efficient methods being developed in the respective areas [6, 14, 15, 33]. As VI methods solve the formulation (1), they seamlessly apply to solve instances of (2) with nonbilinear Ψ.

In addition to the potentially complex structure of Ψ, the size of the data in modern learning tasks lead to development of stochastic variants of VI methods [4, 17, 28]. An important technique on this front is stochastic variance reduction [18] which exploits the finite sum structures in problems to match the convergence rates of deterministic algorithms.

In the specific case of convex minimization, variance reduction has been transformative over the last decade [13, 16, 18, 21]. As a result, there has been several works on developing variance reduced versions of the standard VI methods, including forward-backward [2], extragradient [7, 20], and mirror-prox [5, 27]. Despite recent remarkable advances in this field, these methods rely on strong assumptions such as strong monotonicity [2, 7] or boundedness of the domain [5] and have complicated structures for handling the cases with non-bilinear Ψ [5].

Contributions In this work, we introduce a variance reduced method with a simple single loop structure, for monotone VIs. We prove its almost sure convergence under mere monotonicity; without any of the aforementioned assumptions. The new method achieves the O(1/k) convergence rate in the general monotone case and linear rate of convergence when strong monotonicity is assumed, without using strong monotonicity constant as a parameter. We also consider natural extensions of our algorithm to monotone inclusions, a class of non-monotone problems, and monotone VIs with general Bregman distances.

Related works

Most of the research in variance reduction has focused on convex minimization [13, 16, 18, 21], leading to efficient methods in both theory and practice. On the other hand, variance reduction for solving VIs is started to be investigated recently. One common technique for reducing the variance in stochastic VIs, is to use increasing mini-batch sizes, which leads to high per iteration costs and slower convergence rates in practice [4, 9, 17].

A different approach used in [25] was to use the same sample in both steps of stochastic extragradient method [19] to reduce the variance, which results in a slower O(1/k) rate. The results of [25] for bilinear problems on the other hand are limited to the case when the matrix is full rank. The most related to our work, in the sense how variance reduction is used, are [2, 5, 7] (see Table 1).

Table 1.

We say that the algorithm is μ-adaptive if it does not require strong monotonicity constant as a parameter to obtain linear convergence. [7]) obtains μ-adaptivity if cocoercivity constant of the operator is of the same order as the Lipschitz constant and not in general (see [7, Table 1]). Our complexity matches the rate of deterministic methods [23, 27], however due to worse dependence on n compared to [5], it does not improve deterministic method in bilinear cases

Assumptions for convergence μ-adaptivity Complexity with monotonicity
[2] Strong monotonicity N/A
[7] Strong monotonicity N/A
[5] Monotonicity, bounded domains OnL/ϵ
This work Monotonicity OnL/ϵ

For the specific case of strongly monotone operators, [2] proposed algorithms based on SVRG and SAGA, with linear convergence rates. Two major questions for future work are posed in [2]: (i) obtaining convergence without strong monotonicity assumption and (ii) proving linear convergence without using strong monotonicity constant in the algorithm as a parameter.

The work by [7] proposed an algorithm based on extragradient method [20] and under strong monotonicity assumption, proved linear convergence of the method. The step size in this work depends on cocoercivity constant, which might depend on strong monotonicity constant as discussed in [7, Table 1]. Thus, the result of [7] gave a partial answer to the second question of [2] while leaving the first one unanswered.

An elegant recent work of [5] focused on matrix games and proposed a method based on the mirror prox [27]. The extension of the method of [5] for general min-max problems is also considered there. Unfortunately, this extension not only features a three loop structure, but also uses the bounded domain assumption actively and requires domain diameter as a parameter in the algorithm [5, Corollary 2]. This result has been an important step towards an answer for the first question of [2].

As highlighted in Table 1, our complexity bounds have a worse dependence on n compared to [5], and do not improve the complexity of deterministic VI methods for bilinear games, which was the case in [5]. On the other hand, to our knowledge, our result is the first to show the existence of a variance reduced method that converges under the same set of assumptions as the deterministic methods and also matches the complexity of these deterministic methods. Moreover, our result is also the first variance reduced method to solve monotone inclusions in finite sum form, without strong monotonicity, increasing mini-batch sizes or decreasing step sizes [2].

Finally, our work answers an open problem posed in [23] regarding a stochastic extensions of the forward-reflected-backward method. Our result improves the preliminary result in [23, Section 6], which still requires evaluating the full operator every iteration.

Preliminaries and notation

We work in Euclidean space Z=Rd with scalar product ·,· and induced norm ·. Domain of a function g:ZR{+} is defined as domg={zZ:g(z)<+}. Proximal operator of g is defined as

proxg(u)=argminzZ{g(z)+12z-u2}.

We call an operator F:KZ, where KZ,

  • L-Lipschitz, for L>0, if   F(u)-F(v)Lu-v,u,vK.

  • monotone, if   F(u)-F(v),u-v0,u,vK.

  • ν-cocoercive, for ν>0, if  F(u)-F(v),u-vνF(u)-F(v)2,u,vK.

  • μ-strongly monotone, for μ>0, if  F(u)-F(v),u-vμu-v2,u,vK.

For example, in the context of (2) and (1), F is (strongly) monotone when Ψ is (strongly) convex- (strongly) concave. However, it is worth noting that both cocoercivity and strong monotonicity fail even for the simple bilinear case when Ψ(x,y)=Ax,y in (2).

Given iterates {zk}k1, {wk}k1 and the filtration Fk=σ{z1,,zk,w1,,wk-1}, we define Ek[·]=E[·|Fk] as the conditional expectations with respect to Fk.

Finally, we state our common assumptions for (1).

Assumption 1

  1. g:ZR{+} is proper lower semicontinuous convex.

  2. F:domgZ is monotone.

  3. F=1ni=1nFi, with L-Lipschitz Fi:domgZ, i.

  4. The solution set of (1), denoted by Z, is nonempty.

graphic file with name 10589_2021_305_Figa_HTML.jpg

Algorithm

Our algorithm is a careful mixture of a recent deterministic algorithm for VIs, proposed by [23], with a special technique of using variance reduction in finite sum minimization given in [16] and [21].

It is clear that for n=1 any stochastic variance reduced algorithm for VI reduces to some deterministic one. As a consequence, this immediately rules out the most obvious choice — the well-known forward-backward method (FB)

zk+1=proxτg(zk-τF(zk)), 3

since its convergence requires either strong monotonicity or cocoercivity of F. The classical algorithms that work under mere monotonicity [20, 30, 34] have a more complicated structure, and thus, it is not clear how to meld them with a variance reduction technique for finite sum problems. Instead, we chose the recent forward-reflected-backward method (FoRB) [23]

zk+1=proxτg(zk-τ(2F(zk)-F(zk-1))), 4

which converges under Assumption 1 with n=1.

When g=0, this method takes its origin in the Popov’s algorithm [30]. In this specific case, FoRB is also equivalent to optimistic gradient ascent algorithm [12, 31] which became increasingly popular in machine learning literature recently [11, 12, 24, 26].

Among many variance reduced methods for solving finite sum problems minzf(z):=1ni=1nfi(z) one of the simplest is the Loopless-SVRG method [21] (see also [16]),

zk+1=zk-τf(wk)-τ(fik(zk)-fik(wk))wk+1=zk,with probabilityp,wk,with probability1-p,

which can be seen as a randomized version of the gradient and hence forward-backward methods. The latter is the exact reason why we cannot extend this method directly to the variational inequality setting, without cocoercivity or strong monotonicity.

An accurate blending of [23] and [21], described above, results in Algorithm 1. Compared to Loopless-SVRG, the last evaluation of the operator at step 4 of Algorithm 1 is done at wk-1, instead of wk. In the deterministic case when n=1 or p=1, this modification reduces the method to FoRB (4) and not FB (3). The other change is that we use the most recent iterate zk+1 in the update of wk+1, instead of zk in the Loopless-SVRG. Surprisingly, these two small distinctions result in the method which converges for general VIs without the restrictive assumptions of the previous works.

We note that we use uniform sampling for choosing ik in Algorithm 1 for simplicity. Our arguments directly extend to arbitrary sampling as in [2, 5] which is used for obtaining tighter Lipschitz constants.

Convergence analysis

We start with a key lemma that appeared in [23] for analyzing a general class of VI methods. The proof of this lemma is given in the appendix for completeness. The only change from [23] is that we consider the proximal operator, instead of a more general resolvent.

Lemma 3.1

[23, Proposition 2.3] Let g:ZR{+} be proper lower semicontinuous convex and let x1, U0,U1,V1Z be arbitrary points. Define x2 as

x2=proxg(x1-U1-(V1-U0)). 5

Then for all xZ and V2Z, it holds

x2-x2+2V2-U1,x-x2+2V2,x2-x+2g(x2)-2g(x)x1-x2+2V1-U0,x-x1+2V1-U0,x1-x2-x1-x22. 6

The benefit of Lemma 3.1 is that it gives a candidate for a Lyapunov function that can be used to prove convergence. We will need a slight modification in this function due to randomization in Algorithm 1.

Convergence of the iterates

We start by proving the almost sure convergence of the iterates. Such a result states that the trajectories of the iterates generated by our algorithm converge to a point in the solution set. This type of result is the analogue of sequential convergence results for deterministic methods [23].

For the iterates {zk}, {wk} of Algorithm 1 and any zdomg, β>0 we define

Φk+1(z):=zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+β2zk-wk2+12zk+1-zk2Θk+1(z):=F(zk+1),zk+1-z+g(zk+1)-g(z).

The first expression plays the role of a Lyapunov function and the second is essential for the rate.

Lemma 3.2

Let Assumption 1 hold, τ<1-1-p2L, β=11-p-1, and the iterates {zk} are generated by Algorithm 1. Then for any zdomg,

Ek[Φk+1(z)+2τΘk+1(z)]Φk(z). 7

This lemma is essential in establishing the convergence of iterates and sublinear convergence rates that we will derive in the next section. We now continue with the proof.

Proof

We set in Lemma 3.1U0=τFi(wk-1), U1=τF(wk), V1=τFi(zk), V2=τF(zk+1), and x1=zk, with ik=i. Then by (5) and step 4 of Algorithm 1, x2=zk+1, thus, by (6)

zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+2τ(F(zk+1),zk+1-z+g(zk+1)-g(z))zk-z2+2τFi(zk)-Fi(wk-1),z-zk+2τFi(zk)-Fi(wk-1),zk-zk+1-zk+1-zk2. 8

First, note that by Lipschitzness of Fi, Cauchy-Schwarz and Young’s inequalities,

2τFi(zk)-Fi(wk-1),zk-zk+12τ2L2zk-wk-12+12zk-zk+12. 9

Thus, it follows that

zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+12zk+1-zk2+2τΘk+1(z)zk-z2+2τFi(zk)-Fi(wk-1),z-zk+2τ2L2zk-wk-12. 10

Taking expectation conditioning on the knowledge of zk,wk-1 and using that EkFi(zk)=F(zk), EkFi(wk-1)=F(wk-1), we obtain

Ekzk+1-z2+2τEkF(zk+1)-F(wk),z-zk+1+12Ekzk+1-zk2+2τEkΘk+1(z)zk-z2+2τF(zk)-F(wk-1),z-zk+2τ2L2zk-wk-12. 11

Adding

β2Ekzk-wk2=β(1-p)2zk-wk-12, 12

which follows from the definition of wk, to (11), we obtain

Ek[Φk+1(z)+2τΘk+1(z)]Φk(z)+(2τ2L2+β(1-p)2)zk-wk-12-12zk-zk-12-β2zk-1-wk-12. 13

The proof will be complete, if we can show that the expression in the second and third lines are nonpositive. Due to our choice of β and τ this is a matter of a simple algebra. As β+1=11-p, β1+β=1-1-p, and 2τL<1-1-p=β1+β, we have

2τ2L2+β(1-p)212(β2(1+β)2+β(1+β)2)=β2(1+β). 14

Then we must show that

β1+βzk-wk-12zk-zk-12+βzk-1-wk-12,

which is a direct consequence of u+v2(1+1β)u2+(1+β)v2. The proof is complete.

Theorem 3.1

Let Assumption 1 hold and let τ<1-1-p2L. Then for the iterates {zk} of Algorithm 1, almost surely there exists zZ such that zkz.

Remark 3.1

It is interesting to observe that for p=1, i.e., when the algorithm becomes deterministic, the bound for the stepsize is τ<12L, which coincides with the one in [23] and is known to be tight. In this case analysis will be still valid if for convenience we assume that ·0=0.

For small p we might use a simpler bound for the stepsize, as the following corollary suggests.

Corollary 3.1

Suppose that p=1n and τp4L=14Ln. Then the statement of Theorem 3.1 holds.

Proof

We only have to check that p21-1-p, which follows from 1-p1-p2.

Proof of Theorem 3.1

From Lemma 3.2 we have for any zdomg

Ek[Φk+1(z)+2τΘk+1(z)]Φk(z).

First, we show that Φk+1(z) is nonnegative for all zdomg. This is straightforward but tedious. Recall that 1-1-p=β1+β and hence 2τLβ1+β. Then by Cauchy-Schwarz and Young’s inequalities,

-2τF(zk+1)-F(wk),z-zk+12τLzk+1-wkzk+1-zβ2(1+β)(zk+1-wk2+zk+1-z2)β2(1+β)zk+1-z2+β2(1+β)((1+1β)zk+1-zk2+(1+β)zk-wk2)=β2(1+β)zk+1-z2+12zk+1-zk2+β2zk-wk2. 15

Therefore, we deduce

Φk+1(z)zk+1-z2-β2(1+β)zk+1-z212zk+1-z2. 16

Now let z=z¯Z. Then by monotonicity of F and (1),

Θk+1(z¯)=F(zk+1),zk+1-z¯+g(zk+1)-g(z¯)F(z¯),zk+1-z¯+g(zk+1)-g(z¯)0. 17

Summing up, we have that Θk+1(z¯)0, Φk(z¯)0 and EkΦk+1(z¯)Φk(z¯). Unfortunately, this is still not sufficient for us, so we are going to strengthen this inequality by reexamining the proof of Lemma 3.2. In estimating the second line of inequality (13) we used that 2τL1-1-p, however, both in the statements of Lemma 3.2 and Theorem 3.1 we assumed a strict inequality. Let

δ=β1+β-4τ2L2(1+β)β4τ2L2=β2(1+β)2-δβ1+β. 18

From 2τL<1-1-p=β1+β it follows that δ>0. Now, inequality (14) can be improved to equality as

2τ2L2+β(1-p)2=12(β2(1+β)2-δβ(1+β)+β(1+β)2)=β(1-δ)2(1+β). 19

This change results in a slightly stronger version of (7)

Ek[Φk+1(z¯)+2τΘk+1(z¯)]Φk(z¯)-δ2(zk-zk-12+βzk-1-wk-12). 20

As Φk+1(z¯)0 and Θk+1(z¯)0, we can apply Robbins-Siegmund lemma [32] to conclude that {Φk+1(z¯)} converges almost surely and that

k=1Ezk-zk-12+zk-1-wk-12<. 21

It then follows that almost surely, zk-zk-120 and zk-1-wk-120. Moreover, due to (16), {zk} is almost surely bounded and therefore by the definition of Φk, continuity of F, and (21), we have that zk-z¯2 converges almost surely.

More specifically, this means that for every z¯Z, there exists Ωz¯ with P(Ωz¯)=1 such that ωΩz¯, zk(ω)-z¯2 converges. We can strengthen this result by using the arguments from [3, Proposition 9], [8, Proposition 2.3] to obtain that there exists Ω with P(Ω)=1 such that for every z¯Z and for every ωΩ, zk(ω)-z¯2 converges.

We now pick a realization ωΩ and note that zk(ω)-zk-1(ω)0 and zk-1(ω)-wk-1(ω)0. Let us denote by z~ a cluster point of the bounded sequence zk(ω). By using the definition of zk and convexity of g, as in the proof of Lemma 3.1, we have for any zZ

g(z)g(zk(ω))+1τzk-1(ω)-zk(ω),z-zk(ω)-F(wk-1(ω)),z-zk(ω)-Fik-1(zk-1(ω))-Fik-1(wk-2(ω)),z-zk(ω).

Taking the limit as k and using that g is lower semicontinuous and i, Fi is Lipschitz, zk(ω)-zk-1(ω)0 and zk-1(ω)-wk-1(ω)0, we get that z~Z. Then, as we have that zk(ω)-z~2 converges and we have shown that zk(ω)-z~2 converges to 0 at least on one subsequence, we conclude that the sequence (zk(ω)) converges to some point z~, where z~Z.

Convergence rate for the general case

In this section, we prove that the average of the iterates of the algorithm exhibits O(1/k) convergence rate which is optimal for solving monotone VIs [27]. The standard quantity to show sublinear rates for VIs is gap function which is defined as

G(z¯)=supzZF(z),z¯-z+g(z¯)-g(z).

As this quantity requires taking a supremum over the whole space Z which is potentially unbounded, restricted versions of gap functions are used, for example in [22, 29]

GC(z¯)=supzCF(z),z¯-z+g(z¯)-g(z), 22

where Cdomg is an arbitrary bounded set. It is known that GC(z¯) is a valid merit function, as proven by [29, Lemma 1]. As we are concerned with randomized algorithms, we derive the rate of convergence for the expected gap function EGC(zk).

Theorem 3.2

Given {zk} generated by Algorithm 1, we define the averaged iterate zKav=1Kk=1Kzk. Let Cdomg be an arbitrary bounded set. Then under the hypotheses of Theorem 3.1 it holds that

EGC(zKav)1K1τsupzCz0-z2+2τL2(1+β)δβdist(z0,Z)2,

where δ=β1+β-4τ2L2(1+β)β.

Remark 3.2

If we set p=1n, τ=p32L, and β=11-p-1, the rate will be bounded by nLK32supzCz0-z2+122dist(z0,Z)2, hence it is O(nLK).

The high level idea of the proof is that on top of Lemma 3.2 we sum the resulting inequality and accumulate terms Θk(z). Then we use Jensen’s inequality to obtain the result.

There are two intricate points that need attention in these kind of results. First, the convergence measure is the expected duality gap E[GC(zKav)] that includes the expectation of the supremum. In a standard analysis, it is easy to obtain a bound for the supremum of expectation, however obtaining the former requires a technique, which is common in the literature for saddle point problems [1, 28]. Roughly, the idea is to use an auxiliary iterate to characterize the difference two quantities, and show that the error term does not degrade the rate.

Second, as duality gap requires taking a supremum over the domain, the rate might contain a diameter term as in [5]. The standard way to adjust this result for unbounded domains is to utilize a restricted merit function as in (22) on which the rate is obtained [29]. We note that the result in [5] not only involves the domain diameter in the final bound, but it also requires the domain diameter as a parameter for the algorithm in the general monotone case [5, Corollary 2].

Proof of Theorem 3.2

First, we collect some useful bounds. Consider (20) with a specific choice z¯=PZ(z0). Taking a full expectation and then summing that inequality, we get

δ2k=0Ezk-zk-12+βzk-1-wk-12z0-PZ(z0)2=dist(z0,Z)2, 23

which also implies by Young’s inequality that

βδ2(1+β)k=0Ezk-wk-12dist(z0,Z)2. 24

Next, we rewrite (10) as

2τΘk+1(z)+zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+12zk+1-zk2zk-z2+2τF(zk)-F(wk-1),z-zk+2τ2L2zk-wk-12+2τFik(zk)-Fik(wk-1)-(F(zk)-F(wk-1)),z-zk. 25

Let νk=τ(Fik(zk)-Fik(wk-1)-(F(zk)-F(wk-1))), then Ekνk=0. We define the process {z^k} by z^0=z0 and

z^k+1=z^k+νk. 26

Note that for Fk=σ{z1,,zk,w1,,wk-1}, z^k is Fk-measurable. It also follows that zZ

z^k+1-z2=z^k-z2+2νk,z^k-z+νk2, 27

which after summation over k=0,,K-1 yields

k=0K-12νk,z-z^kz0-z2+k=0K-1νk2. 28

With the definition of νk we can rewrite (25) as

2τΘk+1(z)+zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+12zk+1-zk2zk-z2+2τF(zk)-F(wk-1),z-zk+2τ2L2zk-wk-12+2νk,z-z^k+2νk,z^k-zk.

We use (12), the definition of Φk, and the arguments in Lemma 3.2 to show that the last line of (13) is nonpositive, to obtain

2τΘk+1(z)+Φk+1(z)+β2(Ekzk-wk2-zk-wk2)Φk(z)+2νk,z-z^k+2νk,z^k-zk. 29

Summing this inequality over k=0,,K-1 and using bound (28) yields

2τk=0K-1Θk+1(z)+ΦK(z)+β2k=0K-1(Ekzk-wk2-zk-wk2)Φ0(z)+2k=0K-1νk,z-z^k+2k=0K-1νk,z^k-zkΦ0(z)+z0-z2+2k=0K-1νk2+2k=0K-1νk,z^k-zk=2z0-z2+2k=0K-1νk2+2k=0K-1νk,z^k-zk. 30

We now take the supremum of this inequality over zC and then take a full expectation. As z^k is Fk-measurable, E[Ek[·]]=E[·], and Ekνk=0, we have Ekνk,z^k-zk=0. Using this and that ΦK(z)0 by (16), we arrive at

τEsupzCk=0K-1Θk+1(z)supzCz0-z2+k=0K-1Eνk2. 31

It remains to estimate the last term k=0K-1Eνk2. For this, we use a standard inequality EX-EX2EX2 and Lipschitzness of Fik

k=0K-1Eνk2=k=0K-1Eτ2Fik(zk)-Fik(wk-1)-(F(zk)-F(wk-1))2τ2k=0K-1EFik(zk)-Fik(wk-1)2τ2L2k=0K-1Ezk-wk-12(24)2τ2L2(1+β)δβdist(z0,Z)2. 32

Plugging this bound into (31), we obtain

τEsupzCk=0K-1Θk+1(z)supzCz0-z2+2τ2L2(1+β)δβdist(z0,Z)2. 33

Finally, using monotonicity of F, followed by Jensen inequality, we deduce

supzCk=0K-1Θk+1(z)supzCk=1K(F(z),zk-z+g(zk)-g(z))KGC(zKav),

which combined with (33) finishes the proof.

It is worth mentioning that even though our method is simple and the convergence rate is O(1/k) as in [5], our complexity result has a worse dependence on n, compared to [5]. In particular, our complexity is O(n/ϵ) instead of the O(n/ϵ) of [5]. This is because our step size has the factor of p which is of the order 1n in general and it appears to be tight based on numerical experiments. This seems like the cost of handling a more general problem without bounded domain assumption. We leave it as an open question to derive a method that works under our general assumptions and features favorable complexity guarantees as in [5].

Convergence rate for strongly monotone case

We show that linear convergence is attained when strong monotonicity is assumed.

Theorem 3.3

Let Assumption 1 hold and let F be μ-strongly monotone. Let z be the unique solution of (1). Then for the iterates {zk} generated by Algorithm 1 with τ=p42L, it holds that

Ezk-z21-μp82Lkz0-z2. 34

Remark 3.3

We analyzed the case when F is strongly monotone, however, the same analysis would go through when F is monotone and g is strongly convex. One can transfer strong convexity of g to make F strongly monotone.

Proof of Theorem 3.3

We start from (8) with ik=i,

zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+2τg(zk+1)-2τg(z)+2τF(zk+1),zk+1-zzk-z2+2τFi(zk)-Fi(wk-1),z-zk+2τFi(zk)-Fi(wk-1),zk-zk+1-zk+1-zk2

Setting z=z and using strong monotonicity of F,

F(zk+1),zk+1-z+g(zk+1)-g(z)F(z),zk+1-z+μzk+1-z2+g(zk+1)-g(z)μzk+1-z2.

Hence, we have

(1+2τμ)zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+zk+1-zk2zk-z2+2τFi(zk)-Fi(wk-1),z-zk+2τFi(zk)-Fi(wk-1),zk-zk+1.

Then, we continue as in the proof of Theorem 3.1 until we obtain a stronger version of (20) due to the strong monotonicity term

Ek[(1+2μτ)zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+β2zk-wk2+12zk+1-zk2]zk-z2+2τF(zk)-F(wk-1),z-zk+β2zk-1-wk-12+12zk-zk-12-δ2(zk-zk-12+βzk-1-wk-12). 35

Let ak+1=12zk+1-z2 and

bk+1=12zk+1-z2+2τF(zk+1)-F(wk),z-zk+1+12zk+1-zk2+β2zk-wk2.

Note that we have bk+1+12zk+1-z2=Φk+1(z)12zk+1-z2 by (16), hence bk+10.

Using the definitions of ak and bk in (35), it follows that for any εδ,

Ek[(1+4μτ)ak+1+bk+1]ak+bk-ε2zk-zk-12+βzk-1-wk-12, 36

Next, we derive

RHS of(36)=ak+bk-ε2zk-zk-12-ε2βzk-1-wk-12 37
=1+ε2ak+1-ε2bk-ε4zk-zk-12-εβ4zk-1-wk-12+ετF(zk)-F(wk-1),z-zk1+3ε2ak+1-ε2bk, 38

where the last inequality follows from (15) with a shifted index k. Then, (36) becomes

Ek[(1+4μτ)ak+1+bk+1]1+3ε2ak+1-ε2bk. 39

Since εδ is arbitrary, we can choose ε such that 1+4μτ>1+3ε2. For instance, we can set

ε=minδ,2μτ, 40

that results in

Ek[(1+4μτ)ak+1+bk+1](1+3μτ)ak+1-ε2bk=1-μτ1+4μτ(1+4μτ)ak+1-ε2bk1-minμτ1+4μτ,ε2((1+4μτ)ak+bk). 41

Taking a full expectation and using that ε2=min{δ2,μτ} and b0=0, we obtain

E[(1+4μτ)ak+1+bk+1]1-minμτ1+4μτ,δ2E[(1+4μτ)ak+bk]1-minμτ1+4μτ,δ2k+1(1+4μτ)a0.

Now it only remains to compute the contraction factor. By our choice of τ, we have τL=p421-1-p22=β22(1+β), and hence,

δ=β1+β-4τ2L2(1+β)ββ2(1+β)1-1-p2p4. 42

From μL it follows that 4μτ=μp2Lp2<1 and, hence, μτ1+4μτμτ2=μp82L. Thus, we obtain

minμτ1+4μτ,δ2minμp82L,p8=μp82L,

which finally implies

Ezk+1-z21-μp82Lk+1z0-z2.

A key characteristic of our result is that strong monotonicity constant is not required in the algorithm as a parameter to obtain the rate. This has been raised as an open question by [2] and a partial answer is studied by [7] (see Table 1). Our result gives a full answer to this question without using strong monotonicity constant in all cases.

We next discuss the dependence of μ in the convergence rate. Our rate has a dependence of 1μ compared to 1μ2 of non-accelerated methods of [2] and the method of [7]. This difference is important especially when μ is small. On the other hand, in terms of n, our complexity has a worse dependence compared to [5] and accelerated method of [2] as discussed before (see the discussions in Sect. 1.1 and Section 3.2).

Beyond monotonicity

Lastly, we illustrate that our method has convergence guarantees for a class of non-monotone problems. There exist several relaxations of monotonicity that are used in the literature [10, 17, 22, 24]. Among these, we assume the existence of the solutions to Minty variational inequality given as

z^Z:F(z),z-z^+g(z)-g(z^)0,zZ. 43

Under (43), we can drop the monotonicity assumption and show almost sure subsequential convergence of the iterates of our method. Naturally, in this case one can no longer show sequential convergence as with monotonicity (see Theorem 3.1).

Theorem 3.4

Suppose that Assumption 1 (a), (c), (d) and the condition (43) hold. Then almost surely all cluster points of the sequence {zk} generated by Algorithm 1 are in Z.

Proof

We will proceed as in Theorem 3.1 and [22, Theorem 6]. We note that Lemma 3.2 does not use monotonicity of F, thus its result follows in this case. In the inequality

Ek[Φk+1(z)+2τΘk+1(z)]Φk(z).

we plug in z=z^ for a point satisfying (43).

Then, by (43), we have

Θk+1(z^)=F(zk+1),zk+1-z^+g(zk+1)-g(z^)0.

We then argue the same way as in Theorem 3.1 to conclude that almost surely, {zk} is bounded and cluster points of {zk} are in Z.

Note that the steps in Theorem 3.1 for showing sequential convergence relies on the choice of z as an arbitrary point in Z, which is not the case here, therefore, we can only use the arguments from Theorem 3.1 for showing subsequential convergence.

Extensions

We illustrate extensions of our results to monotone inclusions and Bregman projections. The proofs for this section are given in the appendix in Section 7.

Monotone inclusions

We have chosen to focus on monotone VIs in the main part of the paper for being able to derive sublinear rates for the gap function. In this section, we show that our analysis extends directly for solving monotone inclusions. In this case, we are interested in finding z such that 0(A+F)(z), where AF are monotone operators and each Fi is Lipschitz with the form F=1ni=1nFi. In this case, one changes the prox operator in the algorithm, to resolvent operator of A which is defined as JτA(z)=(I+τA)-1(z). Then, one can use Lemma 3.1 as directly given in [23, Proposition 2.3] to prove an analogous result of Theorem 3.1 for solving monotone inclusions. Moreover, when A+F is strongly monotone, one can prove an analogue of Theorem 3.3. We prove the former result and we note that the latter can be shown by applying the steps in Theorem 3.3 on top of Theorem 4.1, which we do not repeat for brevity.

Theorem 4.1

Let A:ZZ be maximally monotone and F:ZZ be monotone with F=1ni=1nFi, where Fi is L-Lipschitz for all i. Assume that (A+F)-1(0) is nonempty and let the iterates {zk} be generated by Algorithm 1 with the update for zk+1

zk+1=JτA(zk-τF(wk)-τ(Fik(zk)-Fik(wk-1))). 44

Then, for τ<1-1-p2L, almost surely there exist z(A+F)-1(0) such that zkz.

Bregman distances

We developed our analysis in the Euclidean setting, relying on 2-norm for simplicity. However, we can also generalize it to proximal operators involving Bregman distances. In this setting, we have a distance generating function h:ZR, which is 1-strongly convex and continuous. We follow the standard convention to assume that subdifferential of h admits a continuous selection, which means that there exists a continuous function h such that h(x)h(x) for all xdomh. We define the Bregman distance as Dh(z,z¯)=h(z)-h(z¯)-h(z¯),z-z¯. Then, we will change the proximal step 4 of Algorithm 1 with

zk+1=argminz{g(z)+F(wk)+Fik(zk)-Fik(wk-1),z-zk+1τDh(z,zk)}. 45

We prove an analogue of Lemma 3.2 with Bregman distances from which the convergence rate results will follow.

Lemma 4.1

Let Assumption 1 hold and

Φk+1(z):=Dh(z,zk+1)+τF(zk+1)-F(wk),z-zk+1+β4zk-wk2+12Dh(zk+1,zk).

Moreover, suppose τ<1-1-p2L, β=11-p-1, and the iterates {zk} are generated by Algorithm 1 with the update (45) for zk+1. Then for any zdomg,

Ek[Φk+1(z)+τΘk+1(z)]Φk(z).

Numerical verification

In this section, we include preliminary experimental results for our algorithm. We would like to note that these results are mainly for verifying our theoretical results and are not intended to serve as complete benchmarks. We suspect that for an extensive practical comparison, some practical enhancements of our method similar to proximal-point acceleration from [2] or restarting from [7] may be useful. We leave such investigations for future work.

First, we apply our method to the unconstrained bilinear problem. It was shown in [7] that this simple problem is particularly challenging for stochastic methods, due to unboundedness of the domain, where the standard methods, such as stochastic extragradient method [19], diverges. Our assumptions are general enough to cover this case and we now verify in practice that our method indeed converges for this problem by setting d=n=100 and generating AiRd×d randomly with distribution N(0,1)

minxRdmaxyRd1ni=1nAix,y. 46

For this experiment, we test the tightness of our step size rule by progressively increasing it. Recall that our step size is τ=pcL, where c=4 is suggested in our analysis, see Corollary 3.1. We try the values of c={0.5,1,2,4} and observe that for c=0.5 the algorithm diverges, see Fig. 1(left). The message of this experiment is that even though slightly higher step sizes than what is allowed in our theory might work, it is not possible to significantly increase it.

Fig. 1.

Fig. 1

Left: bilinear problem. Middle: Constrained minimization with data generated by normal distribution. Right: Constrained minimization with data generated by uniform distribution

The second problem we consider is constrained minimization, which is an instance where the dual domain is not necessarily bounded. We want to solve

minxCf(x)s.t.hi(x)0,i=1,,m,

where f(x)=12x-u2 for some uZ and hi(x)=Aix-bi2-δi for AiRd×d, biRd, δiR++, and C is a unit ball. In other words, we want to find a projection of u onto the intersection given by C and the constraint inequalities {x:hi(x)0}.

Introducing Lagrange multipliers yi for each constraint, we obtain (see Section 7.3 for further details)

minxRdmaxyR+mf(x)+i=1myihi(x).

As the Lipschitz constant in this problem does not admit a closed-form expression, we first estimate the Lipschitz constant by finding an L such that deterministic method converges. Next, we note that even though we analyzed the algorithm with a single step size τ for both primal and dual variables xy, one can use different step sizes for primal and dual variables (see [22, Section 4.1]). Therefore, we tuned the scaling of primal and dual step sizes for both methods with one random instance and we used the same scaling for all tests for both methods.

We set p=1/m. Every iteration, the deterministic method needs to go through all m constraints to compute i=1myihi(x), whereas our method computes hi(x) for only one i. First setup is with m=400, d=100, and the data is generated with the normal distribution N(0,1). Second setup is with m=400, d=50, and the data is generated with the uniform distribution U(-1,1). We ran both setups with 10 different instances of randomly generated data and plotted all results, see Fig. 1. We observe that in one instance, the tuned scaling diverges for deterministic method, whereas our method with the same tuning converged in all cases.

Conclusion

In this work, we proposed a variance reduced algorithm for solving monotone VIs without assuming bilinearity, strong monotonicity, cocoercivity or boundedness. Even though our method is the first to converge under the same set of assumptions as deterministic methods, a drawback of our approach is the lack of complexity improvements.

In particular, previous approach of [5] showed complexity improvements for bilinear games, while needing more assumptions than deterministic methods to converge. Thus, an important open problem is to obtain a method that i) converges under the minimal set of assumptions as our algorithm, ii) features improved complexity guarantees compared to deterministic methods, while solving structured problems such as bilinear games such as [5] to obtain the best of both worlds.

7. Appendix

7.1 Proofs for Sect. 3

Proof of Lemma 3.1

By using the definition of proximal operator and convexity of g, we have for all xZ

g(x)g(x2)+x1-U1-(V1-U0)-x2,x-x2=g(x2)+x1-x2,x-x2-U1,x-x2-V1-U0,x-x2. 47

Since 2a,b=a2+b2-a-b2, a,b, it follows that

2x1-x2,x-x2=x1-x22+x-x22-x-x12.

Simple rearrangements give

-U1,x-x2=V2-U1,x-x2-V2,x-x2

and

-V1-U0,x-x2=-V1-U0,x-x1-V1-U0,x1-x2.

Using the last three equalities in (47) completes the result.

7.2 Proofs for Sect. 4

We first need a generalized version of Lemma 3.1. In fact, this is the exact form proven in [23], therefore we do not provide its proof.

Lemma 7.1

[23, Proposition 2.3] Let A:ZZ be maximally monotone and let x1,U0,U1, V1Z be arbitrary points. Define x2 as

x2=JA(x1-U1-(V1-U0)). 48

Then for all xZ, V2Z, and U-A(x), we have

x2-x2+2V2-U1,x-x2+2V2-U,x2-xx1-x2+2V1-U0,x-x1+2V1-U0,x1-x2-x1-x22. 49
7.2.1 Proof of Theorem 4.1
Proof

We will start similar to Lemma 3.2. After setting U0=τFi(wk-1), U1=τF(wk), V1=τFi(zk), V2=τF(zk+1), x1=zk, x2=zk+1 with ik=i and plugging into Lemma 7.1, we have

zk+1-z2+2τF(zk+1)-F(wk),z-zk+1zk-z2+2τFi(zk)-Fi(wk-1),z-zk+2τFi(zk)-Fi(wk-1),zk-zk+1-zk+1-zk2-2τF(zk+1)-F(z),zk+1-z. 50

We use monotonicity for the last term and get

zk+1-z2+2τF(zk+1)-F(wk),z-zk+1zk-z2-zk+1-zk2+2τFi(zk)-Fi(wk-1),z-zk+2τFi(zk)-Fi(wk-1),zk-zk+1. 51

The rest of Lemma 3.2 follows in this case the same way with the lack of the terms with Θk+1(z). Then, similar arguments as in Theorem 3.1 with the changes of i) not having Θk+1(z), ii) using the definition of resolvent instead of proximal operator to show cluster points are solutions, will give the result (see also [23, Theorem 2.5]).

We now present a version of Lemma 3.1 with the proximal operator using Bregman distance.

Lemma 7.2

Let g:ZR{+} be proper lower semicontinuous convex and let x1,U0,U1, V1Z be arbitrary points. Define x2 as

x2=argminzZ{g(z)+U1+(V1-U0),z-x1+Dh(z,x1)}. 52

Then, for all xZ, V2Z we have

Dh(x,x2)+V2-U1,z-x2+V2,x2-x+g(x2)-g(x)Dh(x,x1)+V1-U0,x-x1+V1-U0,x1-x2-Dh(x2,x1). 53
Proof

By the definition of x2, it follows from [35, Property 1] that

g(x)g(x2)-U1+V1-U0,x-x2-Dh(x,x1)+Dh(x,x2)+Dh(x2,x1).

For the bilinear term, we argue the same as Lemma 3.1.

7.2.2 Proof of Lemma 4.1
Proof

We will follow the proof of Lemma 3.2 with suitable changes for Bregman distances.

First, set U0=τFi(wk-1), U1=τF(wk), V1=τFi(zk), V2=τF(zk+1), x1=zk, then x2=zk+1 with ik=i and we plug these into (53) to get

Dh(z,zk+1)+τF(zk+1)-F(wk),z-zk+1+τ(F(zk+1),zk+1-z+g(zk+1)-g(z))Dh(z,zk)+τFi(zk)-Fi(wk-1),z-zk+τFi(zk)-Fi(wk-1),zk-zk+1-Dh(zk+1,zk).

First, note that by Lipschitzness of Fi, Cauchy-Schwarz, Young’s inequalities, and since 12zk-zk-12Dh(zk,zk-1),

τFi(zk)-Fi(wk-1),zk-zk+1τ2L2zk-wk-12+14zk-zk+12τ2L2zk-wk-12+12Dh(zk+1,zk) 54

Thus, it follows that

Dh(z,zk+1)+τF(zk+1)-F(wk),z-zk+1+12Dh(zk+1,zk)+τΘk+1(z)Dh(z,zk)+τFi(zk)-Fi(wk-1),z-zk+τ2L2zk-wk-12. 55

Taking expectation conditioning on the knowledge of zk,wk-1 and using that EkFi(zk)=F(zk), EkFi(wk-1)=F(wk-1), we obtain

EkDh(z,zk+1)+τEkF(zk+1)-F(wk),z-zk+1+12EkDh(zk+1,zk)+τEkΘk+1(z)Dh(z,zk)+τF(zk)-F(wk-1),z-zk+τ2L2zk-wk-12. 56

Adding

β4Ekzk-wk2=β(1-p)4zk-wk-12, 57

which follows by the definition of wk, to (56), we obtain

Ek[Φk+1(z)+Θk+1(z)]Φk(z)+(τ2L2+β(1-p)4)zk-wk-12-12Dh(zk,zk-1)-β4zk-1-wk-12. 58

To show that the last line is nonpositive, we use (14), Young’s inequality as in Lemma 3.2 and 12zk-zk-12Dh(zk,zk-1).

Nonnegativity of Φk follows as in Theorem 3.1 after using 12zk-zk-12Dh(zk,zk-1).

7.3 Experiment details

Only for this section, we will use superscripts for iterates rather than subscripts that we have used up to now. Recall that our problem is

minxRdmaxyR+mf(x)+i=1myihi(x).

This problem is equivalent to the following variational inequality

F(z),z-z0zC×R+m,

where

z=(x,y),F(z)=f(x)+i=1myihi(x)-h(x)=F(1)(z)F(2)(z)

The notation reads: h(x)=(h1(x),,hm(x)) and h(x)=(h1(x),,hm(x)). Let us note h:RdRm, h(x)Rm×d. We note that the residual in y-axes of Fig. 1 is computed as xt-proxg(xt-F(xt)).

We split F as follows

F(z)=1mi=1mFi(z)withFi(z)=f(x)+myihi(x)-mhi(x)ei=Fi(1)(z)Fi(2)(z),

where (ei)i=1m is a standard basis in Rm.

Hence, Algorithm 1, with different step sizes for primal and dual, will be

xk+1=PC(xk-τF(1)(uk,vk)-τ(Fi(1)(xk,yk)-Fi(1)(uk-1,vk-1)))yk+1=PRm+(yk-σF(2)(uk,vk)-σ(Fi(2)(xk,yk)-Fi(2)(uk-1,vk-1)))(uk+1,vk+1)=(xk+1,yk+1)with probabilityp(uk,vk)with probability1-p 59

Funding

Open Access funding provided by EPFL Lausanne. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement no 725594 - time-data), the Swiss National Science Foundation (SNSF) under grant number 200021_178865/1, the Department of the Navy, Office of Naval Research (ONR) under a grant number N62909-17-1-211, the Hasler Foundation Program: Cyber Human Systems (project number 16066), the Wallenberg Al, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation with project number 305286.

Declarations

Conflict of interest

The author declared that there is no conflict of interest.

Data availability

Not applicable.

Footnotes

Part of the work was done while Y. Malitsky was at EPFL.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Ahmet Alacaoglu, Email: ahmet.alacaoglu@epfl.ch.

Yura Malitsky, Email: yurii.malitskyi@liu.se.

Volkan Cevher, Email: volkan.cevher@epfl.ch.

References

  • 1.Alacaoglu, A., Fercoq, O., Cevher, V. On the convergence of stochastic primal-dual hybrid gradient. arXiv preprint arXiv:1911.00799, (2019)
  • 2.Balamurugan, P., and Bach. F. Stochastic variance reduction methods for saddle-point problems. In Advances in Neural Information Processing Systems, pages 1416–1424, (2016)
  • 3.Bertsekas DP. Incremental proximal methods for large scale convex optimization. Math. Prog. 2011;129(2):163. doi: 10.1007/s10107-011-0472-0. [DOI] [Google Scholar]
  • 4.Boţ R. I., Mertikopoulos, P., Staudigl, M., and Vuong, P. T. Forward-backward-forward methods with variance reduction for stochastic variational inequalities. arXiv preprint arXiv:1902.03355, (2019)
  • 5.Carmon, Y., Jin, Y., Sidford, A., and Tian. K. Variance reduction for matrix games. In Advances in Neural Information Processing Systems, pages 11377–11388, (2019)
  • 6.Chambolle A, Pock T. A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imag. Vis. 2011;40(1):120–145. doi: 10.1007/s10851-010-0251-1. [DOI] [Google Scholar]
  • 7.Chavdarova, T., Gidel G., Fleuret., F and  Lacoste-Julien., S. Reducing noise in GAN training with variance reduced extragradient. In Advances in Neural Information Processing Systems, pages 391–401, (2019)
  • 8.Combettes PL, Pesquet J-C. Stochastic quasi-Fejér block-coordinate fixed point iterations with random sweeping. SIAM J. Optim. 2015;25(2):1221–1248. doi: 10.1137/140971233. [DOI] [Google Scholar]
  • 9.Cui., S and Shanbhag., UV. On the analysis of variance-reduced and randomized projection variants of single projection schemes for monotone stochastic variational inequality problems. arXiv preprint arXiv:1904.11076, (2019)
  • 10.Dang CD, Lan G. On the convergence properties of non-euclidean extragradient methods for variational inequalities with generalized monotone operators. Comput. Opt. Appl. 2015;60(2):277–310. doi: 10.1007/s10589-014-9673-9. [DOI] [Google Scholar]
  • 11.Daskalakis. C., and Panageas. I. The limit points of (optimistic) gradient descent in min-max optimization. In Advances in Neural Information Processing Systems, pages 9236–9246, (2018)
  • 12.Daskalakis, C., Ilyas, A., Syrgkanis, V., and Zeng, H. Training GANs with optimism. In International Conference on Learning Representations, (2018)
  • 13.Defazio, A., Bach, F., and Lacoste-Julien, S. Saga: A fast incremental gradient method with support for non-strongly convex composite objectives. In Advances in Neural Information Processing Systems, pages 1646–1654, (2014)
  • 14.Esser E, Zhang X, Chan TF. A general framework for a class of first order primal-dual algorithms for convex optimization in imaging science. SIAM J. Imag. Sci. 2010;3(4):1015–1046. doi: 10.1137/09076934X. [DOI] [Google Scholar]
  • 15.Hamedani, E. Y., and Aybat, N. S. A primal-dual algorithm for general convex-concave saddle point problems. arXiv:1803.01401, (2018)
  • 16.Hofmann, T., Lucchi, A., Lacoste-Julien, S., and McWilliams, B. Variance reduced stochastic gradient descent with neighbors. In Advances in Neural Information Processing Systems, pages 2305–2313, (2015)
  • 17.Iusem AN, Jofré A, Oliveira RI, Thompson P. Extragradient method with variance reduction for stochastic variational inequalities. SIAM J. Opt. 2017;27(2):686–724. doi: 10.1137/15M1031953. [DOI] [Google Scholar]
  • 18.Johnson, R., and Zhang. T. Accelerating stochastic gradient descent using predictive variance reduction. In Advances in Neural Information Processing Systems, pages 315–323, (2013)
  • 19.Juditsky A, Nemirovski A, Tauvel C. Solving variational inequalities with stochastic mirror-prox algorithm. Stoch. Syst. 2011;1(1):17–58. doi: 10.1287/10-SSY011. [DOI] [Google Scholar]
  • 20.Korpelevich G. The extragradient method for finding saddle points and other problems. Ekon. Mat. Metody. 1976;12:747–756. [Google Scholar]
  • 21.Kovalev, D., Horvath, S., and Richtarik. P. Don’t jump through hoops and remove those loops: SVRG and Katyusha are better without the outer loop. In Proceedings of the 31st International Conference on Algorithmic Learning Theory, pages 451–467, (2020)
  • 22.Malitsky Y. Golden ratio algorithms for variational inequalities. Math. Prog. 2019;184:383–410. doi: 10.1007/s10107-019-01416-w. [DOI] [Google Scholar]
  • 23.Malitsky Y, Tam MK. A forward-backward splitting method for monotone inclusions without cocoercivity. SIAM J. Optim. 2020;30(2):1451–1472. doi: 10.1137/18M1207260. [DOI] [Google Scholar]
  • 24.Mertikopoulos, P., Lecouat, B., Zenati, H., Foo, C. S., Chandrasekhar, V., and Piliouras. G. Optimistic mirror descent in saddle-point problems: Going the extra(-gradient) mile. In International Conference on Learning Representations, (2019)
  • 25.Mishchenko, K., Kovalev, D., Shulgin, E., Richtárik, P., and Malitsky. Y.: Revisiting stochastic extragradient. In The 23rd International Conference on Artificial Intelligence and Statistics, pages 4573–4582. PMLR, (2020)
  • 26.Mokhtari, A., Ozdaglar, A., and Pattathil. S.: A unified analysis of extra-gradient and optimistic gradient methods for saddle point problems: Proximal point approach. In International Conference on Artificial Intelligence and Statistics, pages 1497–1507. PMLR, (2020)
  • 27.Nemirovski A. Prox-method with rate of convergence O(1/t) for variational inequalities with lipschitz continuous monotone operators and smooth convex-concave saddle point problems. SIAM J. Opt. 2004;15(1):229–251. doi: 10.1137/S1052623403425629. [DOI] [Google Scholar]
  • 28.Nemirovski A, Juditsky A, Lan G, Shapiro A. Robust stochastic approximation approach to stochastic programming. SIAM J. Opt. 2009;19(4):1574–1609. doi: 10.1137/070704277. [DOI] [Google Scholar]
  • 29.Nesterov Y. Dual extrapolation and its applications to solving variational inequalities and related problems. Math. Prog. 2007;109(2–3):319–344. doi: 10.1007/s10107-006-0034-z. [DOI] [Google Scholar]
  • 30.Popov LD. A modification of the Arrow-Hurwicz method for search of saddle points. Math. Notes Acad. Sci. USSR. 1980;28(5):845–848. [Google Scholar]
  • 31.Rakhlin, S., and Sridharan. K, Optimization, learning, and games with predictable sequences. In Advances in Neural Information Processing Systems, pages 3066–3074, (2013)
  • 32.Robbins, H., and Siegmund. D.: A convergence theorem for non negative almost supermartingales and some applications. In Optimizing methods in statistics, pages 233–257. Elsevier, (1971)
  • 33.Shalev-Shwartz, S., and Zhang. T: Stochastic dual coordinate ascent methods for regularized loss minimization. J. Mach. Learn. Res., 14(Feb):567–599, (2013)
  • 34.Tseng P. A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control Opt. 2000;38(2):431–446. doi: 10.1137/S0363012998338806. [DOI] [Google Scholar]
  • 35.Tseng. P.: On accelerated proximal gradient methods for convex-concave optimization. submitted to SIAM J. Opt., 1, (2008)

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Computational Optimization and Applications are provided here courtesy of Springer

RESOURCES