Abstract
We propose a variance reduced algorithm for solving monotone variational inequalities. Without assuming strong monotonicity, cocoercivity, or boundedness of the domain, we prove almost sure convergence of the iterates generated by the algorithm to a solution. In the monotone case, the ergodic average converges with the optimal O(1/k) rate of convergence. When strong monotonicity is assumed, the algorithm converges linearly, without requiring the knowledge of strong monotonicity constant. We finalize with extensions and applications of our results to monotone inclusions, a class of non-monotone variational inequalities and Bregman projections.
Keywords: Variational inequalities, Stochastic variance reduction, Finite-sum structure, Saddle point problems, Monotone inclusions
Introduction
We are interested in solving variational inequalities (VI)
| 1 |
where g is a proper lower semicontinuous convex function and F is a monotone operator also given as the finite sum .
A special case of monotone VIs is the structural saddle point problem
| 2 |
where f, h are proper lower semicontinuous convex functions and is a smooth convex-concave function. Indeed, problem (2) can be formulated as (1) by setting
and (see [2, Section 2], [5, 7] for examples).
Another related problem is the monotone inclusion where the aim is to
where and are maximally monotone operators and F is Lipschitz continuous with finite sum form. Monotone inclusions generalize (1) and our results also extend to this setting as will be shown in Sect. 4.1. Due to convenient abstraction, it is the problem (1) that will be our main concern.
The case when in (2) is convex-concave and, in particular when it is bilinear, has found numerous applications in machine learning, image processing and operations research, resulting in efficient methods being developed in the respective areas [6, 14, 15, 33]. As VI methods solve the formulation (1), they seamlessly apply to solve instances of (2) with nonbilinear .
In addition to the potentially complex structure of , the size of the data in modern learning tasks lead to development of stochastic variants of VI methods [4, 17, 28]. An important technique on this front is stochastic variance reduction [18] which exploits the finite sum structures in problems to match the convergence rates of deterministic algorithms.
In the specific case of convex minimization, variance reduction has been transformative over the last decade [13, 16, 18, 21]. As a result, there has been several works on developing variance reduced versions of the standard VI methods, including forward-backward [2], extragradient [7, 20], and mirror-prox [5, 27]. Despite recent remarkable advances in this field, these methods rely on strong assumptions such as strong monotonicity [2, 7] or boundedness of the domain [5] and have complicated structures for handling the cases with non-bilinear [5].
Contributions In this work, we introduce a variance reduced method with a simple single loop structure, for monotone VIs. We prove its almost sure convergence under mere monotonicity; without any of the aforementioned assumptions. The new method achieves the O(1/k) convergence rate in the general monotone case and linear rate of convergence when strong monotonicity is assumed, without using strong monotonicity constant as a parameter. We also consider natural extensions of our algorithm to monotone inclusions, a class of non-monotone problems, and monotone VIs with general Bregman distances.
Related works
Most of the research in variance reduction has focused on convex minimization [13, 16, 18, 21], leading to efficient methods in both theory and practice. On the other hand, variance reduction for solving VIs is started to be investigated recently. One common technique for reducing the variance in stochastic VIs, is to use increasing mini-batch sizes, which leads to high per iteration costs and slower convergence rates in practice [4, 9, 17].
A different approach used in [25] was to use the same sample in both steps of stochastic extragradient method [19] to reduce the variance, which results in a slower rate. The results of [25] for bilinear problems on the other hand are limited to the case when the matrix is full rank. The most related to our work, in the sense how variance reduction is used, are [2, 5, 7] (see Table 1).
Table 1.
We say that the algorithm is -adaptive if it does not require strong monotonicity constant as a parameter to obtain linear convergence. [7]) obtains -adaptivity if cocoercivity constant of the operator is of the same order as the Lipschitz constant and not in general (see [7, Table 1]). Our complexity matches the rate of deterministic methods [23, 27], however due to worse dependence on n compared to [5], it does not improve deterministic method in bilinear cases
For the specific case of strongly monotone operators, [2] proposed algorithms based on SVRG and SAGA, with linear convergence rates. Two major questions for future work are posed in [2]: (i) obtaining convergence without strong monotonicity assumption and (ii) proving linear convergence without using strong monotonicity constant in the algorithm as a parameter.
The work by [7] proposed an algorithm based on extragradient method [20] and under strong monotonicity assumption, proved linear convergence of the method. The step size in this work depends on cocoercivity constant, which might depend on strong monotonicity constant as discussed in [7, Table 1]. Thus, the result of [7] gave a partial answer to the second question of [2] while leaving the first one unanswered.
An elegant recent work of [5] focused on matrix games and proposed a method based on the mirror prox [27]. The extension of the method of [5] for general min-max problems is also considered there. Unfortunately, this extension not only features a three loop structure, but also uses the bounded domain assumption actively and requires domain diameter as a parameter in the algorithm [5, Corollary 2]. This result has been an important step towards an answer for the first question of [2].
As highlighted in Table 1, our complexity bounds have a worse dependence on n compared to [5], and do not improve the complexity of deterministic VI methods for bilinear games, which was the case in [5]. On the other hand, to our knowledge, our result is the first to show the existence of a variance reduced method that converges under the same set of assumptions as the deterministic methods and also matches the complexity of these deterministic methods. Moreover, our result is also the first variance reduced method to solve monotone inclusions in finite sum form, without strong monotonicity, increasing mini-batch sizes or decreasing step sizes [2].
Finally, our work answers an open problem posed in [23] regarding a stochastic extensions of the forward-reflected-backward method. Our result improves the preliminary result in [23, Section 6], which still requires evaluating the full operator every iteration.
Preliminaries and notation
We work in Euclidean space with scalar product and induced norm . Domain of a function is defined as . Proximal operator of g is defined as
We call an operator , where ,
L-Lipschitz, for , if .
monotone, if .
-cocoercive, for , if .
-strongly monotone, for , if .
For example, in the context of (2) and (1), F is (strongly) monotone when is (strongly) convex- (strongly) concave. However, it is worth noting that both cocoercivity and strong monotonicity fail even for the simple bilinear case when in (2).
Given iterates , and the filtration , we define as the conditional expectations with respect to .
Finally, we state our common assumptions for (1).
Assumption 1
is proper lower semicontinuous convex.
is monotone.
, with L-Lipschitz , .
The solution set of (1), denoted by , is nonempty.
Algorithm
Our algorithm is a careful mixture of a recent deterministic algorithm for VIs, proposed by [23], with a special technique of using variance reduction in finite sum minimization given in [16] and [21].
It is clear that for any stochastic variance reduced algorithm for VI reduces to some deterministic one. As a consequence, this immediately rules out the most obvious choice — the well-known forward-backward method (FB)
| 3 |
since its convergence requires either strong monotonicity or cocoercivity of F. The classical algorithms that work under mere monotonicity [20, 30, 34] have a more complicated structure, and thus, it is not clear how to meld them with a variance reduction technique for finite sum problems. Instead, we chose the recent forward-reflected-backward method (FoRB) [23]
| 4 |
which converges under Assumption 1 with .
When , this method takes its origin in the Popov’s algorithm [30]. In this specific case, FoRB is also equivalent to optimistic gradient ascent algorithm [12, 31] which became increasingly popular in machine learning literature recently [11, 12, 24, 26].
Among many variance reduced methods for solving finite sum problems one of the simplest is the Loopless-SVRG method [21] (see also [16]),
which can be seen as a randomized version of the gradient and hence forward-backward methods. The latter is the exact reason why we cannot extend this method directly to the variational inequality setting, without cocoercivity or strong monotonicity.
An accurate blending of [23] and [21], described above, results in Algorithm 1. Compared to Loopless-SVRG, the last evaluation of the operator at step 4 of Algorithm 1 is done at , instead of . In the deterministic case when or , this modification reduces the method to FoRB (4) and not FB (3). The other change is that we use the most recent iterate in the update of , instead of in the Loopless-SVRG. Surprisingly, these two small distinctions result in the method which converges for general VIs without the restrictive assumptions of the previous works.
We note that we use uniform sampling for choosing in Algorithm 1 for simplicity. Our arguments directly extend to arbitrary sampling as in [2, 5] which is used for obtaining tighter Lipschitz constants.
Convergence analysis
We start with a key lemma that appeared in [23] for analyzing a general class of VI methods. The proof of this lemma is given in the appendix for completeness. The only change from [23] is that we consider the proximal operator, instead of a more general resolvent.
Lemma 3.1
[23, Proposition 2.3] Let be proper lower semicontinuous convex and let , be arbitrary points. Define as
| 5 |
Then for all and , it holds
| 6 |
The benefit of Lemma 3.1 is that it gives a candidate for a Lyapunov function that can be used to prove convergence. We will need a slight modification in this function due to randomization in Algorithm 1.
Convergence of the iterates
We start by proving the almost sure convergence of the iterates. Such a result states that the trajectories of the iterates generated by our algorithm converge to a point in the solution set. This type of result is the analogue of sequential convergence results for deterministic methods [23].
For the iterates , of Algorithm 1 and any , we define
The first expression plays the role of a Lyapunov function and the second is essential for the rate.
Lemma 3.2
Let Assumption 1 hold, , , and the iterates are generated by Algorithm 1. Then for any ,
| 7 |
This lemma is essential in establishing the convergence of iterates and sublinear convergence rates that we will derive in the next section. We now continue with the proof.
Proof
We set in Lemma 3.1, , , , and , with . Then by (5) and step 4 of Algorithm 1, , thus, by (6)
| 8 |
First, note that by Lipschitzness of , Cauchy-Schwarz and Young’s inequalities,
| 9 |
Thus, it follows that
| 10 |
Taking expectation conditioning on the knowledge of and using that , , we obtain
| 11 |
Adding
| 12 |
which follows from the definition of , to (11), we obtain
| 13 |
The proof will be complete, if we can show that the expression in the second and third lines are nonpositive. Due to our choice of and this is a matter of a simple algebra. As , , and , we have
| 14 |
Then we must show that
which is a direct consequence of . The proof is complete.
Theorem 3.1
Let Assumption 1 hold and let . Then for the iterates of Algorithm 1, almost surely there exists such that .
Remark 3.1
It is interesting to observe that for , i.e., when the algorithm becomes deterministic, the bound for the stepsize is , which coincides with the one in [23] and is known to be tight. In this case analysis will be still valid if for convenience we assume that .
For small p we might use a simpler bound for the stepsize, as the following corollary suggests.
Corollary 3.1
Suppose that and . Then the statement of Theorem 3.1 holds.
Proof
We only have to check that , which follows from .
Proof of Theorem 3.1
From Lemma 3.2 we have for any
First, we show that is nonnegative for all . This is straightforward but tedious. Recall that and hence . Then by Cauchy-Schwarz and Young’s inequalities,
| 15 |
Therefore, we deduce
| 16 |
Now let . Then by monotonicity of F and (1),
| 17 |
Summing up, we have that , and . Unfortunately, this is still not sufficient for us, so we are going to strengthen this inequality by reexamining the proof of Lemma 3.2. In estimating the second line of inequality (13) we used that , however, both in the statements of Lemma 3.2 and Theorem 3.1 we assumed a strict inequality. Let
| 18 |
From it follows that . Now, inequality (14) can be improved to equality as
| 19 |
This change results in a slightly stronger version of (7)
| 20 |
As and , we can apply Robbins-Siegmund lemma [32] to conclude that converges almost surely and that
| 21 |
It then follows that almost surely, and . Moreover, due to (16), is almost surely bounded and therefore by the definition of , continuity of F, and (21), we have that converges almost surely.
More specifically, this means that for every , there exists with such that , converges. We can strengthen this result by using the arguments from [3, Proposition 9], [8, Proposition 2.3] to obtain that there exists with such that for every and for every , converges.
We now pick a realization and note that and . Let us denote by a cluster point of the bounded sequence . By using the definition of and convexity of g, as in the proof of Lemma 3.1, we have for any
Taking the limit as and using that g is lower semicontinuous and , is Lipschitz, and , we get that . Then, as we have that converges and we have shown that converges to 0 at least on one subsequence, we conclude that the sequence converges to some point , where .
Convergence rate for the general case
In this section, we prove that the average of the iterates of the algorithm exhibits O(1/k) convergence rate which is optimal for solving monotone VIs [27]. The standard quantity to show sublinear rates for VIs is gap function which is defined as
As this quantity requires taking a supremum over the whole space which is potentially unbounded, restricted versions of gap functions are used, for example in [22, 29]
| 22 |
where is an arbitrary bounded set. It is known that is a valid merit function, as proven by [29, Lemma 1]. As we are concerned with randomized algorithms, we derive the rate of convergence for the expected gap function .
Theorem 3.2
Given generated by Algorithm 1, we define the averaged iterate . Let be an arbitrary bounded set. Then under the hypotheses of Theorem 3.1 it holds that
where .
Remark 3.2
If we set , , and , the rate will be bounded by , hence it is .
The high level idea of the proof is that on top of Lemma 3.2 we sum the resulting inequality and accumulate terms . Then we use Jensen’s inequality to obtain the result.
There are two intricate points that need attention in these kind of results. First, the convergence measure is the expected duality gap that includes the expectation of the supremum. In a standard analysis, it is easy to obtain a bound for the supremum of expectation, however obtaining the former requires a technique, which is common in the literature for saddle point problems [1, 28]. Roughly, the idea is to use an auxiliary iterate to characterize the difference two quantities, and show that the error term does not degrade the rate.
Second, as duality gap requires taking a supremum over the domain, the rate might contain a diameter term as in [5]. The standard way to adjust this result for unbounded domains is to utilize a restricted merit function as in (22) on which the rate is obtained [29]. We note that the result in [5] not only involves the domain diameter in the final bound, but it also requires the domain diameter as a parameter for the algorithm in the general monotone case [5, Corollary 2].
Proof of Theorem 3.2
First, we collect some useful bounds. Consider (20) with a specific choice . Taking a full expectation and then summing that inequality, we get
| 23 |
which also implies by Young’s inequality that
| 24 |
Next, we rewrite (10) as
| 25 |
Let , then . We define the process by and
| 26 |
Note that for , is -measurable. It also follows that
| 27 |
which after summation over yields
| 28 |
With the definition of we can rewrite (25) as
We use (12), the definition of , and the arguments in Lemma 3.2 to show that the last line of (13) is nonpositive, to obtain
| 29 |
Summing this inequality over and using bound (28) yields
| 30 |
We now take the supremum of this inequality over and then take a full expectation. As is -measurable, , and , we have . Using this and that by (16), we arrive at
| 31 |
It remains to estimate the last term . For this, we use a standard inequality and Lipschitzness of
| 32 |
Plugging this bound into (31), we obtain
| 33 |
Finally, using monotonicity of F, followed by Jensen inequality, we deduce
which combined with (33) finishes the proof.
It is worth mentioning that even though our method is simple and the convergence rate is O(1/k) as in [5], our complexity result has a worse dependence on n, compared to [5]. In particular, our complexity is instead of the of [5]. This is because our step size has the factor of p which is of the order in general and it appears to be tight based on numerical experiments. This seems like the cost of handling a more general problem without bounded domain assumption. We leave it as an open question to derive a method that works under our general assumptions and features favorable complexity guarantees as in [5].
Convergence rate for strongly monotone case
We show that linear convergence is attained when strong monotonicity is assumed.
Theorem 3.3
Let Assumption 1 hold and let F be -strongly monotone. Let be the unique solution of (1). Then for the iterates generated by Algorithm 1 with , it holds that
| 34 |
Remark 3.3
We analyzed the case when F is strongly monotone, however, the same analysis would go through when F is monotone and g is strongly convex. One can transfer strong convexity of g to make F strongly monotone.
Proof of Theorem 3.3
We start from (8) with ,
Setting and using strong monotonicity of F,
Hence, we have
Then, we continue as in the proof of Theorem 3.1 until we obtain a stronger version of (20) due to the strong monotonicity term
| 35 |
Let and
Note that we have by (16), hence .
Using the definitions of and in (35), it follows that for any ,
| 36 |
Next, we derive
| 37 |
| 38 |
where the last inequality follows from (15) with a shifted index k. Then, (36) becomes
| 39 |
Since is arbitrary, we can choose such that . For instance, we can set
| 40 |
that results in
| 41 |
Taking a full expectation and using that and , we obtain
Now it only remains to compute the contraction factor. By our choice of , we have , and hence,
| 42 |
From it follows that and, hence, . Thus, we obtain
which finally implies
A key characteristic of our result is that strong monotonicity constant is not required in the algorithm as a parameter to obtain the rate. This has been raised as an open question by [2] and a partial answer is studied by [7] (see Table 1). Our result gives a full answer to this question without using strong monotonicity constant in all cases.
We next discuss the dependence of in the convergence rate. Our rate has a dependence of compared to of non-accelerated methods of [2] and the method of [7]. This difference is important especially when is small. On the other hand, in terms of n, our complexity has a worse dependence compared to [5] and accelerated method of [2] as discussed before (see the discussions in Sect. 1.1 and Section 3.2).
Beyond monotonicity
Lastly, we illustrate that our method has convergence guarantees for a class of non-monotone problems. There exist several relaxations of monotonicity that are used in the literature [10, 17, 22, 24]. Among these, we assume the existence of the solutions to Minty variational inequality given as
| 43 |
Under (43), we can drop the monotonicity assumption and show almost sure subsequential convergence of the iterates of our method. Naturally, in this case one can no longer show sequential convergence as with monotonicity (see Theorem 3.1).
Theorem 3.4
Suppose that Assumption 1 (a), (c), (d) and the condition (43) hold. Then almost surely all cluster points of the sequence generated by Algorithm 1 are in .
Proof
We will proceed as in Theorem 3.1 and [22, Theorem 6]. We note that Lemma 3.2 does not use monotonicity of F, thus its result follows in this case. In the inequality
we plug in for a point satisfying (43).
Then, by (43), we have
We then argue the same way as in Theorem 3.1 to conclude that almost surely, is bounded and cluster points of are in .
Note that the steps in Theorem 3.1 for showing sequential convergence relies on the choice of z as an arbitrary point in , which is not the case here, therefore, we can only use the arguments from Theorem 3.1 for showing subsequential convergence.
Extensions
We illustrate extensions of our results to monotone inclusions and Bregman projections. The proofs for this section are given in the appendix in Section 7.
Monotone inclusions
We have chosen to focus on monotone VIs in the main part of the paper for being able to derive sublinear rates for the gap function. In this section, we show that our analysis extends directly for solving monotone inclusions. In this case, we are interested in finding z such that , where A, F are monotone operators and each is Lipschitz with the form . In this case, one changes the prox operator in the algorithm, to resolvent operator of A which is defined as . Then, one can use Lemma 3.1 as directly given in [23, Proposition 2.3] to prove an analogous result of Theorem 3.1 for solving monotone inclusions. Moreover, when is strongly monotone, one can prove an analogue of Theorem 3.3. We prove the former result and we note that the latter can be shown by applying the steps in Theorem 3.3 on top of Theorem 4.1, which we do not repeat for brevity.
Theorem 4.1
Let be maximally monotone and be monotone with , where is L-Lipschitz for all i. Assume that is nonempty and let the iterates be generated by Algorithm 1 with the update for
| 44 |
Then, for , almost surely there exist such that .
Bregman distances
We developed our analysis in the Euclidean setting, relying on -norm for simplicity. However, we can also generalize it to proximal operators involving Bregman distances. In this setting, we have a distance generating function , which is 1-strongly convex and continuous. We follow the standard convention to assume that subdifferential of h admits a continuous selection, which means that there exists a continuous function such that for all . We define the Bregman distance as . Then, we will change the proximal step 4 of Algorithm 1 with
| 45 |
We prove an analogue of Lemma 3.2 with Bregman distances from which the convergence rate results will follow.
Lemma 4.1
Let Assumption 1 hold and
Moreover, suppose , , and the iterates are generated by Algorithm 1 with the update (45) for . Then for any ,
Numerical verification
In this section, we include preliminary experimental results for our algorithm. We would like to note that these results are mainly for verifying our theoretical results and are not intended to serve as complete benchmarks. We suspect that for an extensive practical comparison, some practical enhancements of our method similar to proximal-point acceleration from [2] or restarting from [7] may be useful. We leave such investigations for future work.
First, we apply our method to the unconstrained bilinear problem. It was shown in [7] that this simple problem is particularly challenging for stochastic methods, due to unboundedness of the domain, where the standard methods, such as stochastic extragradient method [19], diverges. Our assumptions are general enough to cover this case and we now verify in practice that our method indeed converges for this problem by setting and generating randomly with distribution
| 46 |
For this experiment, we test the tightness of our step size rule by progressively increasing it. Recall that our step size is , where is suggested in our analysis, see Corollary 3.1. We try the values of and observe that for the algorithm diverges, see Fig. 1(left). The message of this experiment is that even though slightly higher step sizes than what is allowed in our theory might work, it is not possible to significantly increase it.
Fig. 1.
Left: bilinear problem. Middle: Constrained minimization with data generated by normal distribution. Right: Constrained minimization with data generated by uniform distribution
The second problem we consider is constrained minimization, which is an instance where the dual domain is not necessarily bounded. We want to solve
where for some and for , , , and C is a unit ball. In other words, we want to find a projection of u onto the intersection given by C and the constraint inequalities .
Introducing Lagrange multipliers for each constraint, we obtain (see Section 7.3 for further details)
As the Lipschitz constant in this problem does not admit a closed-form expression, we first estimate the Lipschitz constant by finding an L such that deterministic method converges. Next, we note that even though we analyzed the algorithm with a single step size for both primal and dual variables x, y, one can use different step sizes for primal and dual variables (see [22, Section 4.1]). Therefore, we tuned the scaling of primal and dual step sizes for both methods with one random instance and we used the same scaling for all tests for both methods.
We set . Every iteration, the deterministic method needs to go through all m constraints to compute , whereas our method computes for only one i. First setup is with , , and the data is generated with the normal distribution . Second setup is with , , and the data is generated with the uniform distribution . We ran both setups with 10 different instances of randomly generated data and plotted all results, see Fig. 1. We observe that in one instance, the tuned scaling diverges for deterministic method, whereas our method with the same tuning converged in all cases.
Conclusion
In this work, we proposed a variance reduced algorithm for solving monotone VIs without assuming bilinearity, strong monotonicity, cocoercivity or boundedness. Even though our method is the first to converge under the same set of assumptions as deterministic methods, a drawback of our approach is the lack of complexity improvements.
In particular, previous approach of [5] showed complexity improvements for bilinear games, while needing more assumptions than deterministic methods to converge. Thus, an important open problem is to obtain a method that i) converges under the minimal set of assumptions as our algorithm, ii) features improved complexity guarantees compared to deterministic methods, while solving structured problems such as bilinear games such as [5] to obtain the best of both worlds.
7. Appendix
7.1 Proofs for Sect. 3
Proof of Lemma 3.1
By using the definition of proximal operator and convexity of g, we have for all
| 47 |
Since , , it follows that
Simple rearrangements give
and
Using the last three equalities in (47) completes the result.
7.2 Proofs for Sect. 4
We first need a generalized version of Lemma 3.1. In fact, this is the exact form proven in [23], therefore we do not provide its proof.
Lemma 7.1
[23, Proposition 2.3] Let be maximally monotone and let , be arbitrary points. Define as
| 48 |
Then for all , , and , we have
| 49 |
7.2.1 Proof of Theorem 4.1
Proof
We will start similar to Lemma 3.2. After setting , , , , , with and plugging into Lemma 7.1, we have
| 50 |
We use monotonicity for the last term and get
| 51 |
The rest of Lemma 3.2 follows in this case the same way with the lack of the terms with . Then, similar arguments as in Theorem 3.1 with the changes of i) not having , ii) using the definition of resolvent instead of proximal operator to show cluster points are solutions, will give the result (see also [23, Theorem 2.5]).
We now present a version of Lemma 3.1 with the proximal operator using Bregman distance.
Lemma 7.2
Let be proper lower semicontinuous convex and let , be arbitrary points. Define as
| 52 |
Then, for all , we have
| 53 |
Proof
By the definition of , it follows from [35, Property 1] that
For the bilinear term, we argue the same as Lemma 3.1.
7.2.2 Proof of Lemma 4.1
Proof
We will follow the proof of Lemma 3.2 with suitable changes for Bregman distances.
First, set , , , , , then with and we plug these into (53) to get
First, note that by Lipschitzness of , Cauchy-Schwarz, Young’s inequalities, and since ,
| 54 |
Thus, it follows that
| 55 |
Taking expectation conditioning on the knowledge of and using that , , we obtain
| 56 |
Adding
| 57 |
which follows by the definition of , to (56), we obtain
| 58 |
To show that the last line is nonpositive, we use (14), Young’s inequality as in Lemma 3.2 and .
Nonnegativity of follows as in Theorem 3.1 after using .
7.3 Experiment details
Only for this section, we will use superscripts for iterates rather than subscripts that we have used up to now. Recall that our problem is
This problem is equivalent to the following variational inequality
where
The notation reads: and . Let us note , . We note that the residual in y-axes of Fig. 1 is computed as .
We split F as follows
where is a standard basis in .
Hence, Algorithm 1, with different step sizes for primal and dual, will be
| 59 |
Funding
Open Access funding provided by EPFL Lausanne. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement no 725594 - time-data), the Swiss National Science Foundation (SNSF) under grant number , the Department of the Navy, Office of Naval Research (ONR) under a grant number N62909-17-1-211, the Hasler Foundation Program: Cyber Human Systems (project number 16066), the Wallenberg Al, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation with project number 305286.
Declarations
Conflict of interest
The author declared that there is no conflict of interest.
Data availability
Not applicable.
Footnotes
Part of the work was done while Y. Malitsky was at EPFL.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Ahmet Alacaoglu, Email: ahmet.alacaoglu@epfl.ch.
Yura Malitsky, Email: yurii.malitskyi@liu.se.
Volkan Cevher, Email: volkan.cevher@epfl.ch.
References
- 1.Alacaoglu, A., Fercoq, O., Cevher, V. On the convergence of stochastic primal-dual hybrid gradient. arXiv preprint arXiv:1911.00799, (2019)
- 2.Balamurugan, P., and Bach. F. Stochastic variance reduction methods for saddle-point problems. In Advances in Neural Information Processing Systems, pages 1416–1424, (2016)
- 3.Bertsekas DP. Incremental proximal methods for large scale convex optimization. Math. Prog. 2011;129(2):163. doi: 10.1007/s10107-011-0472-0. [DOI] [Google Scholar]
- 4.Boţ R. I., Mertikopoulos, P., Staudigl, M., and Vuong, P. T. Forward-backward-forward methods with variance reduction for stochastic variational inequalities. arXiv preprint arXiv:1902.03355, (2019)
- 5.Carmon, Y., Jin, Y., Sidford, A., and Tian. K. Variance reduction for matrix games. In Advances in Neural Information Processing Systems, pages 11377–11388, (2019)
- 6.Chambolle A, Pock T. A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imag. Vis. 2011;40(1):120–145. doi: 10.1007/s10851-010-0251-1. [DOI] [Google Scholar]
- 7.Chavdarova, T., Gidel G., Fleuret., F and Lacoste-Julien., S. Reducing noise in GAN training with variance reduced extragradient. In Advances in Neural Information Processing Systems, pages 391–401, (2019)
- 8.Combettes PL, Pesquet J-C. Stochastic quasi-Fejér block-coordinate fixed point iterations with random sweeping. SIAM J. Optim. 2015;25(2):1221–1248. doi: 10.1137/140971233. [DOI] [Google Scholar]
- 9.Cui., S and Shanbhag., UV. On the analysis of variance-reduced and randomized projection variants of single projection schemes for monotone stochastic variational inequality problems. arXiv preprint arXiv:1904.11076, (2019)
- 10.Dang CD, Lan G. On the convergence properties of non-euclidean extragradient methods for variational inequalities with generalized monotone operators. Comput. Opt. Appl. 2015;60(2):277–310. doi: 10.1007/s10589-014-9673-9. [DOI] [Google Scholar]
- 11.Daskalakis. C., and Panageas. I. The limit points of (optimistic) gradient descent in min-max optimization. In Advances in Neural Information Processing Systems, pages 9236–9246, (2018)
- 12.Daskalakis, C., Ilyas, A., Syrgkanis, V., and Zeng, H. Training GANs with optimism. In International Conference on Learning Representations, (2018)
- 13.Defazio, A., Bach, F., and Lacoste-Julien, S. Saga: A fast incremental gradient method with support for non-strongly convex composite objectives. In Advances in Neural Information Processing Systems, pages 1646–1654, (2014)
- 14.Esser E, Zhang X, Chan TF. A general framework for a class of first order primal-dual algorithms for convex optimization in imaging science. SIAM J. Imag. Sci. 2010;3(4):1015–1046. doi: 10.1137/09076934X. [DOI] [Google Scholar]
- 15.Hamedani, E. Y., and Aybat, N. S. A primal-dual algorithm for general convex-concave saddle point problems. arXiv:1803.01401, (2018)
- 16.Hofmann, T., Lucchi, A., Lacoste-Julien, S., and McWilliams, B. Variance reduced stochastic gradient descent with neighbors. In Advances in Neural Information Processing Systems, pages 2305–2313, (2015)
- 17.Iusem AN, Jofré A, Oliveira RI, Thompson P. Extragradient method with variance reduction for stochastic variational inequalities. SIAM J. Opt. 2017;27(2):686–724. doi: 10.1137/15M1031953. [DOI] [Google Scholar]
- 18.Johnson, R., and Zhang. T. Accelerating stochastic gradient descent using predictive variance reduction. In Advances in Neural Information Processing Systems, pages 315–323, (2013)
- 19.Juditsky A, Nemirovski A, Tauvel C. Solving variational inequalities with stochastic mirror-prox algorithm. Stoch. Syst. 2011;1(1):17–58. doi: 10.1287/10-SSY011. [DOI] [Google Scholar]
- 20.Korpelevich G. The extragradient method for finding saddle points and other problems. Ekon. Mat. Metody. 1976;12:747–756. [Google Scholar]
- 21.Kovalev, D., Horvath, S., and Richtarik. P. Don’t jump through hoops and remove those loops: SVRG and Katyusha are better without the outer loop. In Proceedings of the 31st International Conference on Algorithmic Learning Theory, pages 451–467, (2020)
- 22.Malitsky Y. Golden ratio algorithms for variational inequalities. Math. Prog. 2019;184:383–410. doi: 10.1007/s10107-019-01416-w. [DOI] [Google Scholar]
- 23.Malitsky Y, Tam MK. A forward-backward splitting method for monotone inclusions without cocoercivity. SIAM J. Optim. 2020;30(2):1451–1472. doi: 10.1137/18M1207260. [DOI] [Google Scholar]
- 24.Mertikopoulos, P., Lecouat, B., Zenati, H., Foo, C. S., Chandrasekhar, V., and Piliouras. G. Optimistic mirror descent in saddle-point problems: Going the extra(-gradient) mile. In International Conference on Learning Representations, (2019)
- 25.Mishchenko, K., Kovalev, D., Shulgin, E., Richtárik, P., and Malitsky. Y.: Revisiting stochastic extragradient. In The 23rd International Conference on Artificial Intelligence and Statistics, pages 4573–4582. PMLR, (2020)
- 26.Mokhtari, A., Ozdaglar, A., and Pattathil. S.: A unified analysis of extra-gradient and optimistic gradient methods for saddle point problems: Proximal point approach. In International Conference on Artificial Intelligence and Statistics, pages 1497–1507. PMLR, (2020)
- 27.Nemirovski A. Prox-method with rate of convergence for variational inequalities with lipschitz continuous monotone operators and smooth convex-concave saddle point problems. SIAM J. Opt. 2004;15(1):229–251. doi: 10.1137/S1052623403425629. [DOI] [Google Scholar]
- 28.Nemirovski A, Juditsky A, Lan G, Shapiro A. Robust stochastic approximation approach to stochastic programming. SIAM J. Opt. 2009;19(4):1574–1609. doi: 10.1137/070704277. [DOI] [Google Scholar]
- 29.Nesterov Y. Dual extrapolation and its applications to solving variational inequalities and related problems. Math. Prog. 2007;109(2–3):319–344. doi: 10.1007/s10107-006-0034-z. [DOI] [Google Scholar]
- 30.Popov LD. A modification of the Arrow-Hurwicz method for search of saddle points. Math. Notes Acad. Sci. USSR. 1980;28(5):845–848. [Google Scholar]
- 31.Rakhlin, S., and Sridharan. K, Optimization, learning, and games with predictable sequences. In Advances in Neural Information Processing Systems, pages 3066–3074, (2013)
- 32.Robbins, H., and Siegmund. D.: A convergence theorem for non negative almost supermartingales and some applications. In Optimizing methods in statistics, pages 233–257. Elsevier, (1971)
- 33.Shalev-Shwartz, S., and Zhang. T: Stochastic dual coordinate ascent methods for regularized loss minimization. J. Mach. Learn. Res., 14(Feb):567–599, (2013)
- 34.Tseng P. A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control Opt. 2000;38(2):431–446. doi: 10.1137/S0363012998338806. [DOI] [Google Scholar]
- 35.Tseng. P.: On accelerated proximal gradient methods for convex-concave optimization. submitted to SIAM J. Opt., 1, (2008)
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Not applicable.


