Abstract
We present a new theoretical analysis of local superlinear convergence of classical quasi-Newton methods from the convex Broyden class. As a result, we obtain a significant improvement in the currently known estimates of the convergence rates for these methods. In particular, we show that the corresponding rate of the Broyden–Fletcher–Goldfarb–Shanno method depends only on the product of the dimensionality of the problem and the logarithm of its condition number.
Keywords: Quasi-Newton methods, Convex Broyden class, DFP, BFGS, Superlinear convergence, Local convergence, Rate of convergence
Introduction
We study local superlinear convergence of classical quasi-Newton methods for smooth unconstrained optimization. These algorithms can be seen as an approximation of the standard Newton method, in which the exact Hessian is replaced by some operator, which is updated in iterations by using the gradients of the objective function. The two most famous examples of quasi-Newton algorithms are the Davidon–Fletcher–Powell (DFP) [1, 2] and the Broyden–Fletcher–Goldfarb–Shanno (BFGS) [3–7] methods, which together belong to the Broyden family [8] of quasi-Newton algorithms. For an introduction into the topic, see [9] and [10, Chapter 6]. See also [11] for the discussion of quasi-Newton algorithms in the context of nonsmooth optimization.
The superlinear convergence of quasi-Newton methods was established as early as in 1970s, firstly by Powell [12] and Dixon [13, 14] for the methods with exact line search, and then by Broyden, Dennis and Moré [15] and Dennis and Moré [16] for the methods without line search. The latter two approaches have been extended onto more general methods under various settings (see, e.g., [17–25]).
However, explicit rates of superlinear convergence for quasi-Newton algorithms were obtained only recently. The first results were presented in [26] for the greedy quasi-Newton methods. After that, in [27], the classical quasi-Newton methods were considered, for which the authors established certain superlinear convergence rates, depending on the problem dimension and its condition number. The analysis was based on the trace potential function, which was then augmented by the logarithm of determinant of the inverse Hessian approximation to extend the proof onto the general nonlinear case.
In this paper, we further improve the results of [27]. For the classical quasi-Newton methods, we obtain new convergence rate estimates, which have better dependency on the condition number of the problem. In particular, we show that the superlinear convergence rate of BFGS depends on the condition number only through the logarithm. As compared to the previous work, the main difference in the analysis is the choice of the potential function: now the main part is formed by the logarithm of determinant of Hessian approximation, which is then augmented by the trace of inverse Hessian approximation.
It is worth noting that recently, in [28], another analysis of local superlinear convergence of the classical DFP and BFGS methods was presented with the resulting rate, which is independent of the dimensionality of the problem and its condition number. However, to obtain such a rate, the authors had to make an additional assumption that the methods start from a sufficiently good initial Hessian approximation. Without this assumption, to our knowledge, their proof technique, based on the Frobenius-norm potential function, leads only to the rates, which are weaker than those in [27].
This paper is organized as follows. In Sect. 2, we introduce our notation. In Sect. 3, we study the convex Broyden class of quasi-Newton updates for approximating a self-adjoint positive definite operator. In Sect. 4, we analyze the rate of convergence of the classical quasi-Newton methods from the convex Broyden class as applied to minimizing a quadratic function. On this simple example, where the Hessian is constant, we illustrate the main ideas of our analysis. In Sect. 5, we consider the general unconstrained optimization problem. Finally, in Sect. 6, we discuss why the new superlinear convergence rates, obtained in this paper, are better than the previously known ones.
Notation
In what follows, denotes an n-dimensional real vector space. Its dual space, composed of all linear functionals on , is denoted by . The value of a linear function , evaluated at a point , is denoted by .
For a smooth function , we denote by and its gradient and Hessian, respectively, evaluated at a point . Note that , and is a self-adjoint linear operator from to .
The partial ordering of self-adjoint linear operators is defined in the standard way. We write for , if for all , and for , if for all .
Any self-adjoint positive definite linear operator induces in the spaces and the following pair of conjugate Euclidean norms:
| 1 |
When , where is a smooth function with positive definite Hessian, and , we prefer to use notation and , provided that there is no ambiguity with the reference function f.
Sometimes, in the formulas, involving products of linear operators, it is convenient to treat as a linear operator from to , defined by , and as a linear operator from to , defined by . Likewise, any can be treated as a linear operator from to , defined by , and as a linear operator from to , defined by . In this case, and are rank-one self-adjoint linear operators from to and from to , respectively, acting as follows: and for and .
Given two self-adjoint linear operators and , we define the trace and the determinant of A with respect to H as follows: , and . Note that HA is a linear operator from to itself, and hence, its trace and determinant are well defined by the eigenvalues (they coincide with the trace and determinant of the matrix representation of HA with respect to an arbitrary chosen basis in the space , and the result is independent of the particular choice of the basis). In particular, if H is positive definite, then and are, respectively, the sum and the product of the eigenvalues of A relative to . Observe that is a bilinear form, and for any , we have . When A is invertible, we also have and for any . Also recall the following multiplicative formula for the determinant: , which is valid for any invertible linear operator . If the operator H is positive semidefinite, and for some self-adjoint linear operators , then and . Similarly, if A is positive semidefinite and for some self-adjoint linear operators , then and .
Convex Broyden Class
Let A and G be two self-adjoint positive definite linear operators from to , where A is the target operator, which we want to approximate, and G is its current approximation. The Broyden class of quasi-Newton updates of G with respect to A along a direction is the following family of updating formulas, parameterized by a scalar :
| 2 |
where
| 3 |
If the denominator in (3) is zero, we left both and undefined. For the sake of convenience, we also set for .
In this paper, we are interested in the convex Broyden class, which is described by the values of . Note that for all such the denominator in (3) is always positive for any , so both and are well defined; moreover, . For , we have , and (2) becomes the DFP update; for , we have , and (2) becomes the BFGS update.
Remark 3.1
Usually the Broyden class is defined directly in terms of the parameter . However, in the context of this paper, it is more convenient to work with instead of . As can be seen from (66), is exactly the weight of the DFP component in the updating formula for the inverse operator.
A basic property of an update from the convex Broyden class is that it preserves the bounds on the eigenvalues with respect to the target operator.
Lemma 3.1
(see [27, Lemma 2.1]) If for some , then, for any , and any , we have .
Consider the measure of closeness of G to A along direction :
| 4 |
Let us present two potential functions, whose improvement after one update from the convex Broyden class can be bounded from below by a certain nonnegative monotonically increasing function of , vanishing at zero.
First, consider the log-det barrier
| 5 |
It will be useful when . Note that in this case .
Lemma 3.2
Let be self-adjoint positive definite linear operators, for some . Then, for any and :
Proof
Indeed, denoting , we obtain
| 6 |
Since1, we have
| 7 |
Therefore, denoting , we can write that
and, since , that
Substituting the above two inequalities into (6), we obtain the claim.
Now consider another potential function, the augmented log-det barrier:
| 8 |
As compared to the log-det barrier, this potential function is more universal since it works even if the condition is violated. Note that the augmented log-det barrier is in fact the Bregman divergence, generated by the strictly convex function , defined on the set of self-adjoint positive definite linear operators from to , where is an arbitrary fixed self-adjoint positive definite linear operator. Indeed,
| 9 |
Remark 3.2
The idea of combining the trace with the logarithm of determinant to form a potential function for the analysis of quasi-Newton methods can be traced back to [29]. Note also that in [27], the authors studied the evolution of , i.e. the Bregman divergence was centered at A instead of G.
Lemma 3.3
For any real , we have , and
| 10 |
Proof
We only need to prove the first inequality in (10) since the second one follows from it and the fact that (since ).
Let be fixed, and let be the function, defined by . Note that the domain of includes the point since . Let us show that increases on the interval . Indeed, for any , we have
Thus, it is sufficient to prove (10) only in the case when . Equivalently, we need to show that the function , defined by the formula , is nonnegative. Differentiating, we find that, for all , we have
Hence, for , and for . Thus, the minimum of is attained at . Consequently, for all .
It turns out that, up to some constants, the improvement in the augmented log-det barrier can be bounded from below by exactly the same logarithmic function of , which was used for the simple log-det barrier.
Lemma 3.4
Let be self-adjoint positive definite linear operators, for some . Then, for any and :
Proof
Indeed, denoting , we obtain
and
Thus,
| 11 |
where we denote , , , , , . Note that and by the Cauchy–Schwartz inequality. At the same time, by the convexity of the inverse function . Hence, we can apply Lemma 3.3 to estimate (11) from below. Note that
The measure , defined in (4), is the ratio of the norm of , measured with respect to G, and the norm of u, measured with respect to A. It is important that we can change the corresponding metrics to and G, respectively, by paying only with the minimal eigenvalue of G relative to A.
Lemma 3.5
Let be self-adjoint positive definite linear operators such that for some . Then, for any , any , and , we have
Proof
From (66), it is easy to see that . Hence,
| 12 |
Since for all , we further have
| 13 |
Denote . Then,
| 14 |
Consequently,
| 15 |
Thus,
where we have used the Cauchy–Schwartz inequality .
Unconstrained Quadratic Minimization
Let us study the convergence properties of the classical quasi-Newton methods from the convex Broyden class, as applied to minimizing the quadratic function
| 16 |
where is a self-adjoint positive definite linear operator, and .
Let be a fixed self-adjoint positive definite linear operator, and let be such that
| 17 |
Thus, is the strong convexity parameter of f, and L is the constant of Lipschitz continuity of the gradient of f, both measured relative to B.
Consider the following standard quasi-Newton process for minimizing (16):
For measuring its rate of convergence, we use the norm of the gradient, taken with respect to the Hessian:
It is known that the process () has at least a linear convergence rate of the standard gradient method:
Theorem 4.1
(see [27, Theorem 3.1]) In scheme (), for all :
| 19 |
Let us establish the superlinear convergence. According to (19), for the quadratic function, we have for all . Therefore, in our analysis, we can use both potential functions: the log-det barrier and the augmented log-det barrier. Let us consider both options. We start with the first one.
Theorem 4.2
In scheme (), for all , we have
| 20 |
Proof
Without loss of generality, we can assume that for all . Denote , , , for any . By Lemma 3.2 and (19), for all , we have . Summing up, we obtain
| 21 |
Hence, by the convexity of function , we get
| 22 |
But, for all , we have by Lemma 3.5, (19), and since , . Hence, , and so . Rearranging, we obtain . It remains to note that and in view of (19).
Remark 4.1
As can be seen from (21), the factor in (20) can be improved up to , where are the eigenvalues of A relative to B. This improved factor can be significantly smaller than the original one if the majority of the eigenvalues are much larger than .
Let us briefly present another approach, which is based on the augmented log-det barrier. The resulting efficiency estimate will be the same as in Theorem 4.2 up to a slightly worse absolute constant under the exponent. However, this proof can be extended onto general nonlinear functions.
Theorem 4.3
In scheme (), for all , we have
Proof
Without loss of generality, we can assume that for all . Denote , , , for all . By Lemma 3.4 and (19), for all , we have . Hence,
| 23 |
and we can continue exactly as in the proof of Theorem 4.2.
Minimization of General Functions
In this section, we consider the general unconstrained minimization problem:
| 24 |
where is a twice continuously differentiable function with positive definite second derivative. Our goal is to study the convergence properties of the following standard quasi-Newton scheme for solving (24):
Here, is a self-adjoint positive definite linear operator, and L is a positive constant, which together define the initial Hessian approximation .
We assume that there exist constants and , such that
| 26 |
| 27 |
for all . The first assumption (26) specifies that, relative to the operator B, the objective function f is -strongly convex and its gradient is L-Lipschitz continuous. The second assumption (27) means that f is M-strongly self-concordant. This assumption was recently introduced in [26] as a convenient affine-invariant alternative to the standard assumption of the Lipschitz second derivative and is satisfied at least for any strongly convex function with Lipschitz continuous Hessian (see [26, Example 4.1]). The main facts, which we use about strongly self-concordant functions, are summarized in the following lemma (see [26, Lemma 4.1]):
Lemma 5.1
For any , , :
| 28 |
| 29 |
Note that for a quadratic function, we have .
For measuring the convergence rate of (), we use the local gradient norm:
| 30 |
The local convergence analysis of the scheme () is, in general, the same as the corresponding analysis in the quadratic case. However, it is much more technical due to the fact that, in the nonlinear case, the Hessian is no longer constant. This causes a few problems.
First, there are now several different ways how one can treat the Hessian approximation . One can view it as an approximation to the Hessian at the current iterate , to the Hessian at the minimizer , to the integral Hessian , etc. Of course, locally, due to strong self-concordancy, all these variants are equivalent since the corresponding Hessians are close to each other. Nevertheless, from the viewpoint of technical simplicity of the analysis, some options are slightly more preferable than others. We find it to be the most convenient to always think of as an approximation to the integral Hessian .
The second issue is as follows. Suppose we already know what is the connection between our current Hessian approximation and the actual integral Hessian , e.g., in terms of the relative eigenvalues and the value of the augmented log-det barrier potential function (8). Naturally, we want to know how these quantities change after we update into at Step 4 of the scheme (). For this, we apply Lemma 3.1 and Lemma 3.4, respectively. However, the problem is that both of these lemmas will provide us only with the information on the connection between the update result and the current integral Hessian (which was used for performing the update), not the next one . Therefore, we need to additionally take into account the errors, resulting from approximating by .
For estimating the errors, which accumulate as a result of approximating one Hessian by another, it is convenient to introduce the following quantities2:
| 31 |
Remark 5.1
The general framework of our analysis is the same as in the previous paper [27]. The main difference is that now another potential function is used for establishing the rate of superlinear convergence (Lemma 5.4). However, in order to properly incorporate the new potential function into the analysis, many parts in the proof had to be appropriately modified, most notably the part, related to estimating the region of local convergence. In any case, the analysis, presented below, is fully self-contained and does not require the reader first go through [27].
We analyze the method () in several steps. The first step is to establish the bounds on the relative eigenvalues of the Hessian approximations with respect to the corresponding Hessians.
Lemma 5.2
For all , we have
| 32 |
| 33 |
Proof
For , (32) follows from (26) and the fact that and . Now suppose that , and that (32) has already been proved for all indices up to k. Then, applying Lemma 5.1 to (32), we obtain
| 34 |
Since by (31), this proves (33) for the index k. Applying Lemma 3.1 to (34), we get , and so
This proves (32) for the index , and we can continue by induction.
Corollary 5.1
For all , we have
| 35 |
Proof
Indeed,
The second step in our analysis is to establish a preliminary version of the linear convergence theorem for the scheme ().
Lemma 5.3
For all , we have
| 36 |
where
| 37 |
Proof
Let be arbitrary. By Taylor’s formula, we have
| 38 |
Hence,
| 39 |
Note that . Therefore,
Thus, in view of (39) and (30). Consequently,
Next, we establish a preliminary version of the theorem on superlinear convergence of the scheme (). The proof uses the augmented log-det barrier potential function and is essentially a generalization of the corresponding proof of Theorem 4.3.
Lemma 5.4
For all , we have
| 40 |
Proof
Without loss of generality, assume that for all . Denote , , , , and for any .
Let be arbitrary. By Lemma 3.4 and (33), we have
| 41 |
where
| 42 |
Note that in view of (29) and (28). In particular, . Therefore, , and so
Consequently,
| 43 |
Summing up (41), we thus obtain
By the convexity of function , it follows that
| 44 |
At the same time, in view of Lemma 3.5, (33) and since , . Hence, we can write . Consequently, . Rearranging, we obtain that . But by (32), and in view of (26) and the fact that .
In the quadratic case (), we have (see (31)), and Lemmas 5.2 and 5.3 reduce to the already known Theorem 4.1, and Lemma 5.4 reduces to the already known Theorem 4.2. In the general case, the quantities can grow with iterations. However, as we will see in a moment, by requiring the initial point in the scheme () to be sufficiently close to the solution, we can still ensure that stay uniformly bounded by a sufficiently small absolute constant. This allows us to recover all the main results of the quadratic case.
To write down the region of local convergence of (), we need to introduce one more quantity, related to the starting moment of superlinear convergence3:
| 45 |
For DFP () and BFGS (, we have, respectively,
| 46 |
Now we are ready to prove the main result of this section.
Theorem 5.1
Suppose that, in scheme (), we have
| 47 |
Then, for all ,
| 48 |
| 49 |
and, for all ,
| 50 |
Proof
Let us prove by induction that, for all , we have
| 51 |
Clearly, (51) is satisfied for since . It is also satisfied for since .
Now let , and suppose that (51) has already been proved for all indices up to . Then, applying Lemma 5.2, we obtain (48) for all indices up to . Applying now Lemma 5.3 and using for all the relation , we obtain (49) for all indices up to . Finally, if , then, applying Lemma 5.4 and using that for all , we obtain (50) for all indices up to k. Thus, at this moment, (48) and (49) are proved for all indices up to , while (50) is proved only up to k.
To finish the inductive step, it remains to prove that (51) is satisfied for the index , or, equivalently, in view of (31), that . Since in view of (35) and (51), respectively, it suffices to show that .
Note that
| 52 |
Therefore, if we could prove that
| 53 |
then, combining (52) and (53), we would obtain
which is exactly what we need. Let us prove (53). If , in view of (49), we have , and (53) follows. Therefore, from now on, we can assume that . Then4,
It remains to show . We can do this using (50).
First, let us make some estimations. Clearly, for all , we have . Hence, for all , we obtain , and so
| 54 |
At the same time, . Hence,
| 55 |
Thus, for all , and :
Hence, .
Remark 5.2
In accordance with Theorem 5.1, the parameter M of strong self-concordancy affects only the size of the region of local convergence of the process (), and not its rate of convergence. We do not know whether this is an artifact of the analysis or not, but it might be an interesting topic for future research. For a quadratic function, we have , and so the scheme () is globally convergent.
The region of local convergence, specified by (47), depends on the maximum of two quantities: and . For DFP, the part in this maximum is in fact redundant, and its region of local convergence is simply inversely proportional to the condition number: . However, for BFGS, the part does not disappear, and we obtain the following region of local convergence:
Clearly, the latter region can be much bigger than the former when the condition number is significantly larger than the dimension n.
Remark 5.3
The previous estimate of the size of the region of local convergence, established in [27], was for both DFP and BFGS.
Example 5.1
Consider the functions
where , , , , and is the Euclidean norm, induced by the operator B. Let be such that
where is the norm conjugate to . Define
Clearly, , for all , . It is not difficult to check that, for all , we have5
Thus, is a convex function with -Lipschitz gradient and -Lipschitz Hessian. Consequently, the function f is -strongly convex with L-Lipschitz gradient, -Lipschitz Hessian, and, in view of [26, Example 4.1], M-strongly self-concordant, where
Let the regularization parameter be sufficiently small, namely . Denote . Then, , , so, according to (47), the region of local convergence of BFGS can be described as follows:
Discussion
Let us compare the new convergence rates, obtained in this paper for the classical DFP and BFGS methods, with the previously known ones from [27]. Since the estimates for the general nonlinear case differ from those for the quadratic one just in absolute constants, we only discuss the latter case.
In what follows, we use our standard notation: n is the dimension of the space, is the strong convexity parameter, L is the Lipschitz constant of the gradient, and is the local norm of the gradient at the kth iteration.
For BFGS, the previously known rate (see [27, Theorem 3.2]) is
| 56 |
Although (56) is formally valid for all , it becomes useful6 only after
| 57 |
iterations. Thus, can be thought of as the starting moment of the superlinear convergence, according to the estimate (56).
In this paper, we have obtained a new estimate (Theorem 4.2):
| 58 |
Its starting moment of superlinear convergence can be described as follows:
| 59 |
Indeed, since for any , we have, for all ,
| 60 |
At the same time, for all :
| 61 |
Hence, according the new estimate (58), for all :
| 62 |
Comparing the previously known efficiency estimate (56) and its starting moment of superlinear convergence (57) with the new ones (62), (59), we thus conclude that we manage to put the condition number under the logarithm.
For DFP, the previously known rate (see [27, Theorem 3.2]) is
with the following starting moment of the superlinear convergence:
| 63 |
The new rate, which we have obtained in this paper (Theorem 4.2), is
| 64 |
Repeating the same reasoning as above, we can easily obtain that the new starting moment of the superlinear convergence can be described as follows:
| 65 |
and, for all , the new estimate (64) takes the following form:
Thus, compared to the old result, we have improved the factor up to . Interestingly enough, the ratio between the old starting moments (63), (57) of the superlinear convergence of DFP and BFGS and the new ones (65), (59) have remained the same, , although the both estimates have been improved.
It is also interesting whether the results, obtained in this paper, can be applied to limited-memory quasi-Newton methods such as L-BFGS [30]. Unfortunately, it seems like the answer is negative. The main problem is that we cannot say anything interesting about just a few iterations of BFGS. Indeed, according to our main result, after k iterations of BFGS, the initial residual is contracted by the factor of the form . For all values , this contraction factor is in fact bigger than 1, so the result becomes useless.
Conclusions
We have presented a new theoretical analysis of local superlinear convergence of classical quasi-Newton methods from the convex Broyden class. Our analysis has been based on the potential function involving the logarithm of determinant of Hessian approximation and the trace of inverse Hessian approximation. Compared to the previous works, we have obtained new convergence rate estimates, which have much better dependency on the condition number of the problem.
Note that all our results are local, i.e. they are valid under the assumption that the starting point is sufficiently close to a minimizer. In particular, there is no contradiction between our results and the fact that the DFP method is not known to be globally convergent with inexact line search (see, e.g., [31]).
Let us mention several open questions. First, looking at the starting moment of superlinear convergence of the BFGS method, in addition to the dimension of the problem, we see the presence of the logarithm of its condition number. Although typically such logarithmic factors are considered small, it is still interesting to understand whether this factor can be completely removed.
Second, all the superlinear convergence rates, which we have obtained for the convex Broyden class in this paper, are expressed in terms of the parameter , which controls the weight of the DFP component in the updating formula for the inverse operator. At the same time, in [27], the corresponding estimates were presented in terms of the parameter , which controls the weight of the DFP component in the updating formula for the primal operator. Of course, for the extreme members of the convex Broyden class, DFP and BFGS, and coincide. However, in general, they could be quite different. We do not know if it is possible to express the results of this paper in terms of instead of .
Finally, in all the methods, which we considered, the initial Hessian approximation was LB, where L is the Lipschitz constant of the gradient, measured relative to the operator B. We always assume that this constant is known. Of course, it is interesting to develop some adaptive algorithms, which could start from any initial guess for the constant L, and then somehow dynamically adjust the Hessian approximations in iterations, yet retaining all the original efficiency estimates.
Acknowledgements
The presented results were supported by ERC Advanced Grant 788368. The authors are thankful to the anonymous reviewers for their valuable time and comments.
Appendix
Lemma A.1
Let be self-adjoint positive definite linear operators, let be nonzero, and let be such that is well defined. Then,
| 66 |
and
| 67 |
Proof
Denote . According to Lemma 6.2 in [27], we have
This proves (67) since . Let us prove (66). Denote
| 68 |
Note that
| 69 |
Let and be the identity operators in and . Since , we have
Hence, we can conclude that
Thus, we see that the right-hand side of (66) equals
| 70 |
where
| 71 |
It remains to verify that . Clearly,
| 72 |
Hence,
| 73 |
Consequently,
| 74 |
Thus,
Footnotes
This is obvious when is nondegenerate. The general case follows by continuity.
We follow the standard convention that the sum over the empty set is defined as 0, so . Similarly, the product over the empty set is defined as 1.
Hereinafter, for denotes the smallest positive integer greater or equal to t.
We will estimate the second sum using (50). However, recall that, at this moment, (50) is proved only up to the index k. This is the reason why we move into the first sum.
is the third derivative of f along the direction h.
Indeed, according to Theorem 4.1, we have at least for all .
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Anton Rodomanov, Email: anton.rodomanov@uclouvain.be.
Yurii Nesterov, Email: yurii.nesterov@uclouvain.be.
References
- 1.Davidon, W.: Variable metric method for minimization. Argonne National Laboratory Research and Development Report 5990, (1959)
- 2.Fletcher R, Powell M. A rapidly convergent descent method for minimization. Comput. J. 1963;6(2):163–168. doi: 10.1093/comjnl/6.2.163. [DOI] [Google Scholar]
- 3.Broyden C. The convergence of a class of double-rank minimization algorithms: 1. General considerations. IMA J. Appl. Math. 1970;6(1):76–90. doi: 10.1093/imamat/6.1.76. [DOI] [Google Scholar]
- 4.Broyden C. The convergence of a class of double-rank minimization algorithms: 2. The new algorithm. IMA J. Appl. Math. 1970;6(3):222–231. doi: 10.1093/imamat/6.3.222. [DOI] [Google Scholar]
- 5.Fletcher R. A new approach to variable metric algorithms. Comput. J. 1970;13(3):317–322. doi: 10.1093/comjnl/13.3.317. [DOI] [Google Scholar]
- 6.Goldfarb D. A family of variable-metric methods derived by variational means. Math. Comput. 1970;24(109):23–26. doi: 10.1090/S0025-5718-1970-0258249-6. [DOI] [Google Scholar]
- 7.Shanno D. Conditioning of quasi-Newton methods for function minimization. Math. Comput. 1970;24(111):647–656. doi: 10.1090/S0025-5718-1970-0274029-X. [DOI] [Google Scholar]
- 8.Broyden C. Quasi-Newton methods and their application to function minimization. Math. Comput. 1967;21(99):368–381. doi: 10.1090/S0025-5718-1967-0224273-2. [DOI] [Google Scholar]
- 9.Dennis J, Moré J. Quasi-Newton methods, motivation and theory. SIAM Rev. 1977;19(1):46–89. doi: 10.1137/1019005. [DOI] [Google Scholar]
- 10.Nocedal J, Wright S. Numerical Optimization. New York: Springer; 2006. [Google Scholar]
- 11.Lewis A, Overton M. Nonsmooth optimization via quasi-Newton methods. Math. Program. 2013;141(1–2):135–163. doi: 10.1007/s10107-012-0514-2. [DOI] [Google Scholar]
- 12.Powell M. On the convergence of the variable metric algorithm. IMA J. Appl. Math. 1971;7(1):21–36. doi: 10.1093/imamat/7.1.21. [DOI] [Google Scholar]
- 13.Dixon L. Quasi-Newton algorithms generate identical points. Math. Program. 1972;2(1):383–387. doi: 10.1007/BF01584554. [DOI] [Google Scholar]
- 14.Dixon L. Quasi Newton techniques generate identical points II: the proofs of four new theorems. Math. Program. 1972;3(1):345–358. doi: 10.1007/BF01585007. [DOI] [Google Scholar]
- 15.Broyden C, Dennis J, Moré J. On the local and superlinear convergence of quasi-Newton methods. IMA J. Appl. Math. 1973;12(3):223–245. doi: 10.1093/imamat/12.3.223. [DOI] [Google Scholar]
- 16.Dennis J, Moré J. A characterization of superlinear convergence and its application to quasi-Newton methods. Math. Comput. 1974;28(126):549–560. doi: 10.1090/S0025-5718-1974-0343581-1. [DOI] [Google Scholar]
- 17.Stachurski A. Superlinear convergence of Broyden’s bounded -class of methods. Math. Program. 1981;20(1):196–212. doi: 10.1007/BF01589345. [DOI] [Google Scholar]
- 18.Griewank A, Toint P. Local convergence analysis for partitioned quasi-Newton updates. Numer. Math. 1982;39(3):429–448. doi: 10.1007/BF01407874. [DOI] [Google Scholar]
- 19.Engels J, Martínez H. Local and superlinear convergence for partially known quasi-Newton methods. SIAM J. Optim. 1991;1(1):42–56. doi: 10.1137/0801005. [DOI] [Google Scholar]
- 20.Byrd R, Liu D, Nocedal J. On the behavior of Broyden’s class of quasi-Newton methods. SIAM J. Optim. 1992;2(4):533–557. doi: 10.1137/0802026. [DOI] [Google Scholar]
- 21.Yabe H, Yamaki N. Local and superlinear convergence of structured quasi-Newton methods for nonlinear optimization. J. Oper. Res. Soc. Jpn. 1996;39(4):541–557. [Google Scholar]
- 22.Wei Z, Yu G, Yuan G, Lian Z. The superlinear convergence of a modified BFGS-type method for unconstrained optimization. Comput. Optim. Appl. 2004;29(3):315–332. doi: 10.1023/B:COAP.0000044184.25410.39. [DOI] [Google Scholar]
- 23.Yabe H, Ogasawara H, Yoshino M. Local and superlinear convergence of quasi-Newton methods based on modified secant conditions. J. Comput. Appl. Math. 2007;205(1):617–632. doi: 10.1016/j.cam.2006.05.018. [DOI] [Google Scholar]
- 24.Mokhtari A, Eisen M, Ribeiro A. IQN: an incremental quasi-Newton method with local superlinear convergence rate. SIAM J. Optim. 2018;28(2):1670–1698. doi: 10.1137/17M1122943. [DOI] [Google Scholar]
- 25.Gao W, Goldfarb D. Quasi-Newton methods: superlinear convergence without line searches for self-concordant functions. Optim. Methods Softw. 2019;34(1):194–217. doi: 10.1080/10556788.2018.1510927. [DOI] [Google Scholar]
- 26.Rodomanov, A., Nesterov, Y.: Greedy quasi-Newton methods with explicit superlinear convergence. CORE Discussion Papers. 06 (2020)
- 27.Rodomanov, A., Nesterov, Y.: Rates of Superlinear Convergence for Classical Quasi-Newton Methods. CORE Discussion Papers. 11 (2020) [DOI] [PMC free article] [PubMed]
- 28.Jin, Q., Mokhtari, A.: Non-asymptotic Superlinear Convergence of Standard Quasi-Newton Methods. arXiv preprint arXiv:2003.13607 (2020)
- 29.Byrd R, Nocedal J. A tool for the analysis of quasi-Newton methods with application to unconstrained minimization. SIAM J. Numer. Anal. 1989;26(3):727–739. doi: 10.1137/0726042. [DOI] [Google Scholar]
- 30.Liu D, Nocedal J. On the limited memory BFGS method for large scale optimization. Math. Program. 1989;45(1–3):503–528. doi: 10.1007/BF01589116. [DOI] [Google Scholar]
- 31.Byrd R, Nocedal J, Yuan Y. Global convergence of a class of quasi-Newton methods on convex problems. SIAM J. Numer. Anal. 1987;24(5):1171–1190. doi: 10.1137/0724077. [DOI] [Google Scholar]
