Skip to main content
PLOS One logoLink to PLOS One
. 2015 Oct 26;10(10):e0140071. doi: 10.1371/journal.pone.0140071

Two New PRP Conjugate Gradient Algorithms for Minimization Optimization Models

Gonglin Yuan 1,2, Xiabin Duan 1,*, Wenjie Liu 2,3, Xiaoliang Wang 1, Zengru Cui 1, Zhou Sheng 1
Editor: Yongtang Shi4
PMCID: PMC4621041  PMID: 26502409

Abstract

Two new PRP conjugate Algorithms are proposed in this paper based on two modified PRP conjugate gradient methods: the first algorithm is proposed for solving unconstrained optimization problems, and the second algorithm is proposed for solving nonlinear equations. The first method contains two aspects of information: function value and gradient value. The two methods both possess some good properties, as follows: 1)β k ≥ 0 2) the search direction has the trust region property without the use of any line search method 3) the search direction has sufficient descent property without the use of any line search method. Under some suitable conditions, we establish the global convergence of the two algorithms. We conduct numerical experiments to evaluate our algorithms. The numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems and that the second algorithm is effective for solving large-scale nonlinear equations.

Introduction

As we know, the conjugate gradient method is very popular and effective for solving the following unconstrained optimization problem

minxnf(x) (1)

where f : ℜn → ℜ is continuously differentiable and g(x) denotes the gradient of f(x) at x, the problem Eq (1) also can be applied to model some other problems [15]. The iterative formula used in the conjugate gradient method is usually given by

xk+1=xk+αkdk (2)

and

dk={gkifk=1gk+βkdk1ifk2 (3)

where g k = g(x k), β k ∈ ℜ is a scalar, α k > 0 is a step length that is determined by some line search, and d k denotes the search direction. Different conjugate methods have different choices for β k. Some of the popular methods [612] used to compute β k are the DY conjugate gradient method [6], FR conjugate gradient method [7], PRP conjugate gradient method [8, 9], HS conjugate gradient method [10], LS conjugate gradient method [11], and CD conjugate gradient method [12]. β k [8, 9] is defined by

βkPRP=gkTyk1gk12 (4)

where · denotes the Euclidean norm, y k−1 = g kg k−1. The PRP conjugate gradient method is currently considered to have the best numerical performance, but it does not have good convergence. With an exact line search, the global convergence of the PRP conjugate gradient method has been established by Polak and Ribière [8] for convex objective functions. However, Powell [13] proposed a counter example that proved the existence of nonconvex functions on which the PRP conjugate gradient method does not have global convergence, even with an exact line search. With the weak Wolfe-Powell line search, Gilbert and Nocedal [14] proposed a modified PRP conjugate gradient method by restricting β k to be not less than zero and proved that it has global convergence, with the hypothesis that it satisfies the sufficient descent condition. Gilbert and Nocedal [14] also gave an example showing that β k may be negative even though the objective function is uniformly convex. When the Strong Wolfe-Powell line search was used, Dai [15] gave a example showing that the PRP method cannot guarantee that every step search direction is the descent direction, even if the objective function is uniformly convex.

Through the above observations and [13, 14, 1618], we know that the following sufficient descent condition

gkTdkbgk2,   b>0 (5)

and the condition β k is not less than zero are very important for establishing the global convergence of the conjugate gradient method.

The weak Wolfe-Powell (WWP) line search is designed to compute α k and is usually used for the global convergence analysis. The WWP line search is as follows

f(xk+αkdk)f(xk)+δ1αkgkTdk (6)

and

g(xk+αkdk)Tdkδ2gkTdk (7)

where δ1(0,12),δ2(δ1,1).

Recently, many new conjugate gradient methods ([1928] etc.) that possess some good properties have been proposed for solving unconstrained optimization problems.

In Section 2, we state the motivation behind our approach and give a new modified PRP conjugate gradient method and new algorithm for solving problem Eq (1). In Section 3, we prove that the search direction of our new algorithm satisfies the sufficient descent property and trust region property; moreover, we establish the global convergence of the new algorithm with the WWP line search. In Section 4, we provide numerical experiment results for some test problems.

New algorithm for unconstrained optimization

Wei et al. [29] give a new PRP conjugate gradient method usually called the WYL method. When the WWP line search is used, this WYL method has global convergence under the sufficient descent condition. Zhang [30] give a modified WYL method called the NPRP method as follows

βkNPRP=gk2gkgk1|gkTgk1|gk12

The NPRP method possesses better convergence properties. The above formula for y k−1 contains only gradient value information, but some new y k−1 formulas [31, 32] contain information on gradient value and function value. Yuan et al.[32] propose a new y k−1 formula as follows

yk-1m=yk-1+max{ρk-1,0}sk-12sk-1,

and

ρk-1=2[f(xk-1)-f(xk)]+(g(xk)+g(xk-1))Tsk-1.

Where s k−1 = x kx k−1.

Li and Qu [33] give a modified PRP conjugate method as follows

βk=gkTyk-1max{tdk-1,gk-12},t>0

and

dk=-gk-βkgkTdk-1gk2gk+βkdk-1,d0=-g0.

Under suitable conditions, Li and Qu [33] prove that the modified PRP conjugate method has global convergence.

Motivated by the above discussions, we propose a new modified PRP conjugate method as follows

βkBPRP=min{|gkTyk1m|,  u1(gk2gkgk1|gkTgk1|)}u2dk1yk1+gk12 (8)

and

dk={-gkifk=1-gk-βkBPRPgkTdk-1gk2gk+βkBPRPdk-1ifk2 (9)

where u 1 > 0, u 2 > 0, yk-1m is the yk-1m of [32].

As gk2-gkgk-1|gkTgk-1|0, it follows directly from the above formula that βkBPRP0. Next, we present a new algorithm and it’s diagram (Fig 1) as follows.

Fig 1. The diagram about Algorithm 2.1.

Fig 1

Algorithm 2.1

Step 0: Given the initial point x1n,u1>0,u2>0,ε10,0<δ1<12,δ1<δ2<1, set d 1 = −∇f(x 1) = −g 1, k: = 1.

Step 1: Calculate gk; if gkε1, stop; otherwise, go to step 2.

Step 2: Calculate step length α k by the WWP line search.

Step 3: Set x k+1 = x k + α k d k, then calculate gk+1; if gk+1ε1, stop; otherwise, go to step 4.

Step 4: Calculate the scalar β k+1 by Eq (8) and calculate the search direction d k+1 by Eq (9).

Step 5: Set k: = k + 1; go to step 2.

Global convergence analysis

Some suitable assumptions are often used to analyze the global convergence of the conjugate gradient method. Here, we state it as follows

Assumption 3.1

  1. The level set Ω = {x ∈ ℜnf(x) ≤ f(x 1)} is bounded.

  2. In some neighborhood H of Ω, f is a continuously differentiable function, and the gradient function g of f is Lipschitz continuous, namely, there exists a constant L > 0 such that
    g(x)g(y)Lxy,x,yH (10)

By Assumption 3.1, it is easy to obtain that there exist two constants A > 0 and η 1 > 0 satisfying

xA,   g(x)η1,      xΩ (11)

Lemma 0.1 Let the sequence {d k} be generated by Eq (9); then, we have

gkTdk=gk2,       k1 (12)

Proof When k = 1, we can obtain g1Td1=g12 by Eq (9), so Eq (12) holds. When k ≥ 2, we can obtain

gkTdk=gkT(gkβkBPRPgkTdk1gk2gk+βkBPRPdk1)              =  gk2

The proof is achieved.

We know directly from above Lemma that our new method has the sufficient descent property.

Lemma 0.2 Let the sequence {x k} and {d k, g k} be generated by Algorithm 2.1, and suppose that Assumption 3.1 holds; then, we can obtain

k=1(gkTdk)2dk2<+ (13)

Proof By Eq (7) and the Cauchy-Schwarz inequality, we have

(1δ2)gkTdk  gk+1gkdk

Combining the above inequality with Assumption 3.1 ii) generates

(1δ2)gkTdkLαkdk2

it is easy to know gkTdk0 by lemma 0.1. By combining the above inequality with Eq (6), we obtain

fk-fk+1δ1(1-δ2)L(gkTdk)2dk2.

Summing up the above inequalities from k = 1 to k = ∞, we can deduce that

δ1(1-δ2)Lk=1(gkTdk)2dk2f1-f.

By Eq (6), Assumption 3.1 and lemma 0.1, we know that {f k} is bounded below, so we obtain

k=1(gkTdk)2dk2<+.

This finishes the proof.

The Eq (13) is usually called the Zoutendijk condition [34], and it is very important for establishing global convergence.

Lemma 0.3 Let the sequence {β k, d k} be generated by Algorithm 2.1, we have

dkNgk (14)

where N=1+4u1u2.

Proof When d k = 0, we directly get g k = 0 from Eq (12). When d k ≠ 0, by the Cauchy-Schwarz inequality, we can easily obtain

gk2gkgk1|gkTgk1|gkT(gkgkgk1gk1)

and

gkT(gkgkgk1gk1)gk(gkgk1)+(gk1gkgk1gk1)2gk  gkgk1

We can obtain

gk2gkgk1  |gkTgk1|  2  gk  yk1

Using Eq (8), we have

|βkBPRP|u1(gk2gkgk1  |gkTgk1|)u2dk1yk12u1u2gkdk1

Finally, when k ≥ 2 by Eq (9), we have

dkgk+|βkBPRP|gkdk1gk2gk+|βkBPRP|dk1gk+2u1u2gk+2u1u2gk(1+4u1u2)gk

Let N=1+4u1u2; we obtain dkNgk. This finishes the proof.

This lemma also shows that the search direction of our algorithm has the trust region property.

Theorem 0.1 Let the sequence {d k, g k, β k} and {x k} be generated by Algorithm 2.1. Suppose that Assumption 3.1 holds; then

limkgk=0 (15)

Proof By Eqs (12) and (13), we obtain

k=1gk4dk2<+ (16)

By Eq (14), we have dk2N2gk2; then, we obtain

gk2N2gk4dk2,

which together with Eq (16) can yield

k=1gk2N2k=1gk4dk2<+.

From the above inequality, we can obtain limkgk=0. The proof is finished.

Numerical Results

When β k+1 and d k+1 are calculated by Eqs (4) and (3), respectively, in step 4 of Algorithm 2.1, we call it the PRP conjugate gradient algorithm. We test Algorithm 2.1 and the PRP conjugate gradient algorithm using some benchmark problems. The test environment is MATLAB 7.0, on a Windows 7 system. The initial parameters are given by

u1=1,u2=2,δ1=0.2,δ2=0.8,ε1=10-6.

We use the following Himmeblau stop rule, which satisfies

If ∣f(x k)∣ ≤ ɛ 2, let stop1 = stop1=|f(xk)f(xk+1)|; otherwise, let stop1=|f(xk)f(xk+1)||f(xk)|. The test program will be stopped if stop1 < ɛ 3 or g(xk)<ε1 is satisfied, where ɛ 2 = ɛ 3 = 10−6. When the total number of iterations is greater than one thousand, the test program will be stopped. The test results are given in Tables 1 and 2: x 1 denotes the initial point, Dim denotes the dimension of test function, NI denotes the the total number of iterations, and NFG = NF+NG (NF and NG denote the number of the function evaluations and the number of the gradient evaluations, respectively). f denotes the function value when the program is stopped. The test problems are defined as follows.

  1. Schwefel function:
    fSch(x)=418.9829n+i=1nxisin|xi|,xi[-512.03,511.97],
    x*=(-420.9687,-420.9687,...,-420.9687),fSch(x*)=0.
  2. Langerman function:
    fLan(x)=-i=1mcie-1πj=1n(xj-aij)2cos(πj=1n(xj-aij)2),xi[0,10],m=n,
    x*=random,fLan(x*)=random.
  3. Schwefel′s function
    fSchDS(x)=i=1n(j=1ixj)2,xi[65.536,65.536],
    x*=(0,0,...,0),fSchDS(x*)=0.
  4. Sphere function:
    fSph(x)=i=1nxi2,xi[-5.12,5.12],
    x*=(0,0,...,0),fSph(x*)=0.
  5. Griewangk function:
    fGri(x)=1+i=1nxi24000-i=1ncos(xii),xi[-600,600],
    x*=(0,0,...,0),fGri(x*)=0.
  6. Rosenbrock function:
    fRos(x)=i=1n-1[100(xi+1-xi2)2+(xi-1)2],xi[-2.048,2.048],
    x*=(1,...,1),fRos(x*)=0.
  7. Ackley function:
    fAck(x)=20+e-20e-0.21ni=1nxi2-e1ni=1ncos(2πxi),xi[-30,30],
    x*=(0,0,...,0),fAck(x*)=0.
  8. Rastrigin function:
    fRas(x)=10n+i=1n(xi2-10cos(2πxi)),xi[-5.12,5.12],
    x*=(0,0,...,0),fRas(x*)=0.

Table 1. Test results for Algorithm 2.1.

Problems Dim x 1 NI/NFG f′
1 50 (-426,-426,…,-426) 2/9 6.363783e-004
120 (-426,-426,…,-426) 2/9 1.527308e-003
200 (-426,-426,…,-426) 2/9 2.545514e-003
1000 (-410,-410,…,-410) 3/12 1.272757e-002
2 50 (3,3,…,3) 0/2 -1.520789e-060
120 (5,5,…,5) 0/2 0.000000e+000
200 (6,6,…,6) 0/2 0.000000e+000
1000 (1,1,…,1) 0/2 -7.907025e-136
3 50 (-0.00001,0,-0.00001,0,…) 2/8 1.561447e-009
120 (-0.00001,0,-0.00001,0,…) 2/8 1.769900e-008
200 (-0.00001,0,-0.00001,0,…) 2/8 7.906818e-008
1000 (0.000001,0,0.000001,0,…) 2/8 9.619586e-008
4 50 (-4,-4,…,-4) 1/6 1.577722e-028
120 (-2,-2,…,-2) 1/6 3.786532e-028
200 (1,1,…,1) 1/6 7.730837e-027
1000 (3,3,…,3) 1/6 1.079951e-024
5 50 (-7,0,-7,0,…) 2/10 0.000000e+000
120 (0.592,0,0.592,0,…) 4/14 3.183458e-007
200 (0.451,0,0.451,0,…) 4/14 3.476453e-007
1000 (0.38,0,0.38,0,…) 1/6 0.000000e+000
6 50 (1.001,1.001,…,1.001) 2/36 4.925508e-003
120 (1.001,1.001,…,1.001) 2/36 1.198551e-002
200 (1.001,1.001,…,1.001) 2/36 2.006158e-002
1000 (1.001,1.001,…,1.001) 2/36 1.009107e-001
7 50 (0.01,0,0.01,0,…) 0/2 3.094491e-002
120 (-0.05,0,-0.05,0,…) 0/2 2.066363e-001
200 (0.01,0,0.01,0,…) 0/2 3.094491e-002
1000 (0.07,0,0.07,0,…) 0/2 3.233371e-001
8 50 (0.003,0.003,…,0.003) 3/26 0.000000e+000
120 (0.005,0.005,…,0.005) 2/9 0.000000e+000
200 (0.006,0,0.006,0,…) 2/9 0.000000e+000
1000 (0.015,0.015,…,0.015) 2/8 0.000000e+000

Table 2. Test results for the PRP conjugate gradient algorithm.

Problems Dim x 1 NI/NFG f′
1 50 (-426,-426,…,-426) 2/24 6.363783e-004
120 (-426,-426,…,-426) 2/11 1.527308e-003
200 (-426,-426,…,-426) 3/41 2.545514e-003
1000 (-410,-410,…,-410) 3/41 1.272757e-002
2 50 (3,3,…,3) 0/2 -1.520789e-060
120 (5,5,…,5) 0/2 0.000000e+000
200 (6,6,…,6) 0/2 0.000000e+000
1000 (1,1,…,1) 0/2 -7.907025e-136
3 50 (-0.00001,0,-0.00001,0,…) 2/8 1.516186e-009
120 (-0.00001,0,-0.00001,0,…) 2/8 1.701075e-008
200 (-0.00001,0,-0.00001,0,…) 2/8 7.579825e-008
1000 (0.000001,0,0.000001,0,…) 2/8 9.198262e-008
4 50 (-4,-4,…,-4) 1/6 1.577722e-028
120 (-2,-2,…,-2) 1/6 3.786532e-028
200 (1,1,…,1) 1/6 7.730837e-027
1000 (3,3,…,3) 1/6 1.079951e-024
5 50 (-7,0,-7,0,…) 4/16 3.597123e-013
120 (0.592,0,0.592,0,…) 5/17 3.401145e-007
200 (0.451,0,0.451,0,…) 5/17 4.566281e-007
1000 (0.38,0,0.38,0,…) 1/6 0.000000e+000
6 50 (1.001,1.001,…,1.001) 2/36 4.925508e-003
120 (1.001,1.001,…,1.001) 2/36 1.198551e-002
200 (1.001,1.001,…,1.001) 2/36 2.006158e-002
1000 (1.001,1.001,…,1.001) 2/36 1.009107e-001
7 50 (0.01,0,0.01,0,…) 0/2 3.094491e-002
120 (-0.05,0,-0.05,0,…) 0/2 2.066363e-001
200 (0.01,0,0.01,0,…) 0/2 3.094491e-002
1000 (0.07,0,0.07,0,…) 0/2 3.233371e-001
8 50 (0.003,0.003,…,0.003) 2/10 0.000000e+000
120 (0.005,0.005,…,0.005) 2/10 0.000000e+000
200 (0.006,0,0.006,0,…) 2/10 0.000000e+000
1000 (0.015,0.015,…,0.015) 2/22 3.636160e-009

It is easy to see that the two algorithms are effective for the above eight test problems listed in Tables 1 and 2. We use the tool of Dolan and Morè [35] to analyze the numerical performance of the two algorithms.

For the above eight test problems, Fig 2 shows the numerical performance of the two algorithms when the information of NI is considered, and Fig 3 shows the the numerical performance of the two algorithms when the information of NFG is considered. From the above two figures, it is easy to see that Algorithm 2.1 yields a better numerical performance than the PRP conjugate gradient algorithm on the whole. From Tables 1 and 2 and the two figures, we can conclude that Algorithm 2.1 is effective and competitive for solving unconstrained optimization problems.

Fig 2. Performance profiles of the two algorithms (NI).

Fig 2

Fig 3. Performance profiles of the two algorithms (NFG).

Fig 3

A new algorithm is given for solving nonlinear equations in the next section. The sufficient descent property and the trust region property of the new algorithm are proved in Section 6; moreover, we establish the global convergence of the new algorithm. In Section 7, the numerical results are presented.

New algorithm for nonlinear equations

We consider the system of nonlinear equations

q(x)=0,xn. (17)

where q : ℜn → ℜn is a continuously differentiable and monotonic function. ∇q(x) denotes the Jacobian matrix of q(x); if ∇q(x) is symmetric, we call Eq (17) symmetric nonlinear equations. As q(x) is monotonic, the following inequality

(q(x)-q(y))T(x-y)0,x,yn

holds. If a norm function is defined as follows

h(x)=12q(x)2

and we define the unconstrained optimization problem as follows,

minh(x),xn (18)

We know directly that the problem Eq (17) is equivalent to the problem Eq (18).

The iterative formula Eq (2) is also usually used in many algorithms for solving problem Eq (17). Many algorithms ([3641], etc.) have been proposed for solving special classes of nonlinear equations. We are more interested in the process of dealing with large-scale nonlinear equations. By Eq (2), it is easy to see that the two factors of stepsize α k and search direction d k are very important for dealing with large-scale problems. When dealing with large-scale nonlinear equations and unconstrained optimization problems, there are many popular methods ([38, 4246] etc.) for computing d k, such as conjugate gradient methods, spectral gradient methods, and limited-memory quasi-Newton approaches. Some new line search methods [37, 47] have been proposed for calculating α k. Li and Li [48] provide the following new derivative-free line search method

q(xk+αkdk)Tdkσ3αkq(xk+αkdk)dk2, (19)

where α k = max{γ, ργ, ρ 2 γ, …}, ρ ∈ (0,1), σ 3 > 0 and γ > 0. This line search method is very effective for solving large-scale nonlinear monotonic equations.

Solodov and Svaiter [49] presented a hybrid projection-proximal point algorithm that could conquer some drawbacks when the form Eq (18) is used with nonlinear equations. Yuan et al.[50] proposed a three-term PRP conjugate gradient algorithm by using the projection-based technique, which was introduced by Solodov et al.[51] for optimization problems. The projection-based technique is very effective for solving nonlinear equations. It involves certain methods to compute search direction d k and certain line search methods to calculate α k, which satisfies

q(wk)T(xk-wk)>0

in which w k = x k + α k d k. For any x* that satisfies q(x*) = 0, considering that q(x) is monotonic, we can obtain

q(wk)T(x*-wk)0.

Thus, it is easy to obtain the current iterate x k, which is strictly separated from the zeros of the system of equations Eq (17) by the following hyperplane

Tk={xn|q(wk)T(x-wk)=0}

Then, the iterate x k+1 can be obtained by projecting x k onto the above hyperplane. The projection formula can be set as follows

xk+1=xk-q(wk)T(xk-wk)q(wk)2q(wk) (20)

Yuan et al. [50] present a three-term Polak-Ribière-Polyak conjugate gradient algorithm in which the search direction d k is defined as follows

dk={-qkifk=0-qk+qkTyk-1dk-1-qkTdk-1yk-1max{μdk-1yk-1,qk-12}ifk1

where y k−1 = q kq k−1. The derivative-free line search method [48] and the projection-based techniques are used by the algorithm [50], proved to be very suitable for solving large-scale nonlinear equations. The most attractive property of algorithm [50] is the the trust region property of d k.

Motivated by our new modified PRP conjugate gradient formula, proposed in Section 2, we proposed the following modified PRP conjugate gradient formula

βk*=min{|qkT(qkqk1)|,  u3(qk2qkqk1|qkTqk1|)}u4dk1qkqk1+qk12 (21)

and

dk={-qk,ifk=1-qk-βk*qkTdk-1qk2qk+βk*dk-1,ifk2 (22)

Where u 3 > 0, u 4 > 0. It is easy to see that βk*0, motivated by the above observation and [50]. We present a new algorithm for solving problem Eq (17): it uses our modified PRP conjugate gradient formula Eqs (21) and (22). Here, we list the new algorithm and it’s diagram (Fig 4) as follows.

Fig 4. The diagram about Algorithm 5.1.

Fig 4

Algorithm 5.1

Step 1: Given the initial point x 1 ∈ ℜn,ɛ 4 > 0,ρ ∈ (0,1), σ 3 > 0, γ > 0,u 3 > 0, u 4 > 0, and k: = 1.

Step 2: If q(xk)ε4, stop; otherwise, go to step 3.

Step 3: Compute d k by Eq (22) and calculate α k by Eq (19)

Step 4: Set the next iterate to be w k = x k + α k d k;

Step 5: If q(wk)ε4, stop and set x k+1 = w k; otherwise, calculate x k+1 by Eq (20)

Step 6: Set k: = k + 1; go to step 2.

Convergence Analysis

When we analyze the global convergence of Algorithm 5.1, we require the following suitable assumptions.

Assumption 6.1

  1. The solution set of the problem Eq (17) is nonempty.

  2. q(x) is Lipschitz continuous, namely, there exists a constant E > 0 such that
    q(x)-q(y)Ex-y,x,yn.

By Assumption 6.1, it is easy to obtain that there exists a positive constant ζ that satisfies

q(xk)ζ (23)

Lemma 0.4 Let the sequence {d k} be generated by Eq (22) ; then, we can obtain

qkTdk=qk2 (24)

and

qkdk(1+4u3u4)qk (25)

Proof As the proof is similar to Lemma 0.1 and Lemma 0.3 of this paper, we omit it here.

Similar to Lemma 3.1 of [50] and theorem 2.1 of [51], it is easy to obtain the following lemma. Here, we omit this proof and only list it.

Lemma 0.5 Suppose that Assumption 6.1 holds and x* is a solution of problem Eq (17) that satisfies g(x*) = 0. Let the sequence {x k} be obtained by Algorithm 5.1; then, the {x k} is a bounded sequence and

xk+1x*2xkx*2xk+1xk2

holds. Moreover, either {x k} is a infinite sequence and

k=1xk+1-xk2<

or the {x k} is a finite sequence and a solution of problem Eq (17) is the last iteration.

Lemma 0.6 Suppose that Assumption 6.1 holds, then, an iteration x k+1 = x k + α k d k will be generated by Algorithm 5.1 in a finite number of backtracking steps.

Proof We will obtain this conclusion by contradiction: suppose that qk0 does not hold; then, there exists a positive constant ɛ 5 that satisfies

qkε5,k1 (26)

suppose that there exist some iterate indexes k that do not satisfy the condition Eq (19). We let αk(c)=ρ(c)γ then it can obtain

q(xk+αk(c)dk)Tdk<σ3αk(c)q(xk+αk(c)dk)dk2,cN*{0}.

By Assumption 6.1 (b) and Eq (24), we find

dk2=qkTdk=[q(xk+αk(c)dk)q(xk)]Tdkq(xk+αk(c)dk)Tdk<[E+σ3q(xk+αk(c)dk)]αk(c)dk2

By Eqs (23) and (25), we can obtain

q(xk+αk(c)dk)q(xk+αk(c)dk)-qk+qkEαk(c)dk+ζEγζ(1+4u3u4)+ζ

Thus, we obtain

αk(c)>ε52u42[E+σ3(Eγζ(1+4u3u4)+ζ)](u4+4u3)2ζ2,cN*{0}

which shows that αk(c) is bounded below. This contradicts the definition of αk(c); so, the lemma holds.

Similar to Theorem 3.1 of [50], we list the following theorem but omit its proof.

Theorem 0.2 Let the sequence {x k+1, q k+1} and {α k, d k} be generated by Algorithm 5.1. Suppose that Assumption 6.1 holds; then, we have

limkinfqk=0. (27)

Numerical results

When the following d k formula of the famous PRP conjugate gradient method [8, 9]

dk={-qkifk=1-qk+qkT(qk-qk-1)qk-12dk-1ifk2

is used to compute d k in step 3 of Algorithm 5.1, then it is called PRP algorithm. We test Algorithm 5.1 and the PRP algorithm for some problems in this section. The test environment is MATLAB 7.0 on a Windows 7 system. The initial parameters are given by

σ3=u4=0.02,γ=1,ρ=0.1,u3=1,ε4=10-5.

When the number of iterations is greater than or equal to one thousand and five hundred, the test program will also be stopped. The test results are given in Tables 3 and 4. As we know, when the line search cannot guarantee that d k satisfies qkTdk<0, some uphill search direction may be produced; the line search method possibly fails in this case. In order to prevent this situation, when the search time is greater than or equal to fifteen in the inner cycle of our program, we set α k that is acceptable. NG, NI stand for the number of gradient evaluations and iterations respectively. Dim denotes the dimension of the testing function, and cputime denotes the cpu time in seconds. GF denotes the evaluation of the final function norm when the program terminates. The test functions all have the following form

q(x)=(f1(x),f2(x),...,fn(x))T

the concrete function definitions are given as follows.

  • Function 1. Exponential function 2
    f1(x)=ex1-1,fi(x)=i10(exi+xi-1-1),i=2,3,,n
    Initial guess: x0=(1n2,1n2,,1n2)T.
  • Function 2. Trigonometric function
    fi(x)=2(n+i(1-cos(xi))-sin(xi)-k=1ncos(xk))(2sin(xi)-cos(xi)),i=1,2,,n
    Initial guess: x0=(101100n,101100n,,101100n)T.
  • Function 3. Logarithmic function
    fi(x)=ln(xi+1)-xin,i=1,2,3,,n.
    Initial guess: x 0 = (1,1,⋯,1)T.
  • Function 4. Broyden Tridiagonal function [[52], pp. 471–472]
    f1(x)=(3-0.5x1)x1-2x2+1,fi(x)=(3-0.5xi)xi-xi-1+2xi+1+1,i=2,3,,n-1,fn(x)=(3-0.5xn)xn-xn-1+1.
    Initial guess: x 0 = (−1,−1,⋯,−1)T.
  • Function 5. Strictly convex function 1 [[44], p. 29]

    q(x) is the gradient of h(x)=i=1n(exi-xi).
    fi(x)=exi-1,i=1,2,3,,n
    Initial guess: x0=(1n,2n,,1)T.
  • Function 6. Variable dimensioned function
    fi(x)=xi-1,i=1,2,3,,n-2,fn-1(x)=j=1n-2j(xj-1),fn(x)=(j=1n-2j(xj-1))2.
    Initial guess: x0=(1-1n,1-2n,,0)T.
  • Function 7. Discrete boundary value problem [53].
    f1(x)=2x1+0.5h2(x1+h)3-x2,fi(x)=2xi+0.5h2(xi+hi)3-xi-1+xi+1,i=2,3,,n-1fn(x)=2xn+0.5h2(xn+hn)3-xn-1,h=1n+1.
    Initial guess: x 0 = (h(h−1), h(2h−1),⋯,h(nh−1))T.
  • Function 8. Troesch problem [54]
    f1(x)=2x1+ϱh2sinh(ϱx1)-x2fi(x)=2xi+ϱh2sinh(ϱx1)-xi-1-xi+1,i=2,3,,n-1fn(x)=2xn+ϱh2sinh(ϱxn)-xn-1,h=1n+1,ϱ=10.
    Initial guess: x 0 = (0, 0, ⋯, 0)T.

Table 3. Test results for Algorithm 5.1.

Function Dim NI/NG cputime GF
1 3000 55/209 2.043613 9.850811e-006
5000 8/33 0.858005 6.116936e-006
30000 26/127 100.792246 8.983556e-006
45000 7/36 62.681202 7.863794e-006
50000 5/26 56.659563 5.807294e-006
2 3000 43/86 1.076407 8.532827e-006
5000 42/84 2.745618 8.256326e-006
30000 38/76 73.039668 8.065468e-006
45000 37/74 164.284653 8.064230e-006
50000 36/72 201.288090 9.519786e-006
3 3000 5/6 0.093601 1.009984e-008
5000 5/6 0.249602 6.263918e-009
30000 18/33 32.775810 2.472117e-009
45000 21/39 91.229385 2.840234e-010
50000 21/39 108.202294 2.661223e-010
4 3000 95/190 2.137214 9.497689e-006
5000 97/194 5.834437 9.048858e-006
30000 103/206 194.954450 8.891642e-006
45000 104/208 446.568463 9.350859e-006
50000 104/208 549.529123 9.856874e-006
5 3000 64/128 1.497610 9.111464e-006
5000 65/130 4.102826 9.525878e-006
30000 70/140 132.117247 8.131796e-006
45000 70/140 297.868309 9.959279e-006
50000 71/142 374.964004 8.502923e-006
6 3000 1/2 0.031200 0.000000e+000
5000 1/2 0.062400 0.000000e+000
30000 1/2 1.918812 0.000000e+000
45000 1/2 4.258827 0.000000e+000
50000 1/2 5.194833 0.000000e+000
7 3000 35/71 0.842405 9.291878e-006
5000 34/69 2.121614 8.658237e-006
30000 30/61 58.391174 8.288490e-006
45000 29/59 135.627269 8.443996e-006
50000 29/58 153.801386 9.993530e-006
8 3000 0/1 0.015600 0.000000e+000
5000 0/1 0.046800 0.000000e+000
30000 0/1 1.326008 0.000000e+000
45000 0/1 2.917219 0.000000e+000
50000 0/1 3.510022 0.000000e+000

Table 4. Test results for PRP algorithm.

Function Dim NI/NG cputime GF
1 3000 58/220 2.043613 9.947840e-006
5000 24/97 2.496016 9.754454e-006
30000 29/141 109.668703 9.705424e-006
45000 13/66 118.108357 9.450575e-006
50000 10/51 112.383120 9.221806e-006
2 3000 48/95 1.138807 8.647042e-006
5000 46/91 2.932819 9.736889e-006
30000 41/81 78.733705 9.983531e-006
45000 40/79 181.709965 9.632281e-006
50000 40/79 212.832164 9.121412e-006
3 3000 11/12 0.171601 1.012266e-008
5000 11/12 0.530403 8.539532e-009
30000 23/38 39.749055 2.574915e-009
45000 26/44 100.542645 2.931611e-010
50000 26/44 123.864794 2.838473e-010
4 3000 104/208 2.246414 9.243312e-006
5000 106/212 6.193240 9.130520e-006
30000 113/226 219.821009 8.747379e-006
45000 114/228 487.908728 9.368026e-006
50000 114/228 611.976323 9.874918e-006
5 3000 35/53 0.561604 2.164559e-006
5000 35/53 1.716011 1.291210e-006
30000 35/53 55.926358 1.336971e-006
45000 33/49 116.361146 2.109293e-006
50000 33/49 147.452145 2.225071e-006
6 3000 1/2 0.031200 0.000000e+000
5000 1/2 0.062400 0.000000e+000
30000 1/2 1.965613 0.000000e+000
45000 1/2 4.290028 0.000000e+000
50000 1/2 5.257234 0.000000e+000
7 3000 40/80 0.904806 9.908999e-006
5000 39/78 2.386815 9.198351e-006
30000 34/68 66.440826 9.515010e-006
45000 33/66 140.026498 9.366998e-006
50000 33/66 173.597913 8.886013e-006
8 3000 0/1 0.015600 0.000000e+000
5000 0/1 0.031200 0.000000e+000
30000 0/1 1.279208 0.000000e+000
45000 0/1 2.808018 0.000000e+000
50000 0/1 3.432022 0.000000e+000

By Tables 3 and 4, we see that Algorithm 5.1 and the PRP algorithm are effective for solving the above eight problems.

We use the tool of Dolan and Morè [35] to analyze the numerical performance of the two algorithms when NI, NG and cputime are considered, for which we generate three figures.

Fig 5 shows that the numerical performance of Algorithm 5.1 is slightly better than that of the PRP algorithm when NI is considered. It is easy to see that the numerical performance of Algorithm 5.1 is better than that of the PRP algorithm from Figs 6 and 7 because the PRP algorithm requires a bigger horizontal axis when the problems are completely solved.

Fig 5. Performance profiles of the two algorithms (NI).

Fig 5

Fig 6. Performance profiles of the two algorithms (NG).

Fig 6

Fig 7. Performance profiles of the two algorithms (cputime).

Fig 7

From the above two tables and three figures, we see that Algorithm 5.1 is effective and competitive for solving large-scale nonlinear equations.

Conclusion

(i) This paper provides the first new algorithm based on the first modified PRP conjugate gradient method in Sections 1–4. The β k formula of the method includes the gradient value and function value. The global convergence of the algorithm is established under some suitable conditions. The trust region property and sufficient descent property of the method have been proved without the use of any line search method. For some test functions, the numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems.

(ii) The second new algorithm based on the second modified PRP conjugate gradient method is presented in Sections 5-7. The new algorithm has global convergence under suitable conditions. The trust region property and the sufficient descent property of the method are proved without the use of any line search method. The numerical results of some tests function are demonstrated. The numerical results show that the second algorithm is very effective for solving large-scale nonlinear equations.

Acknowledgments

This work is supported by China NSF (Grant No. 11261006 and 11161003), NSFC No. 61232016, NSFC No. U1405254, the Guangxi Science Fund for Distinguished Young Scholars (No. 2015GXNSFGA139001) and PAPD issue of Jiangsu advantages discipline. The authors wish to thank the editor and the referees for their useful suggestions and comments which greatly improve this paper.

Data Availability

All data are available in the paper.

Funding Statement

This work is supported by China NSF (Grant No. 11261006 and 11161003), NSFC No. 61232016, NSFC No. U1405254, the Guangxi Science Fund for Distinguished Young Scholars (No. 2015GXNSFGA139001) and PAPD issue of Jiangsu advantages discipline.

References

  • 1. Gu B, Sheng V, Feasibility and finite convergence analysis for accurate on-line v-support vector learning, IEEE Transactions on Neural Networks and Learning Systems. 24 (2013) 1304–1315. 10.1109/TNNLS.2013.2250300 [DOI] [PubMed] [Google Scholar]
  • 2. Li J, Li X, Yang B, Sun X, Segmentation-based Image Copy-move Forgery Detection Scheme, IEEE Transactions on Information Forensics and Security. 10 (2015) 507–518. 10.1109/TIFS.2014.2381872 [DOI] [Google Scholar]
  • 3. Wen X, Shao L, Fang W, Xue Y, Efficient Feature Selection and Classification for Vehicle Detection, IEEE Transactions on Circuits and Systems for Video Technology. 2015, 10.1109/TCSVT.2014.2358031 [DOI] [Google Scholar]
  • 4. Zhang H, Wu J, Nguyen TM, Sun M, Synthetic Aperture Radar Image Segmentation by Modified Student’s t-Mixture Model, IEEE Transaction on Geoscience and Remote Sensing. 52 (2014) 4391–4403. 10.1109/TGRS.2013.2281854 [DOI] [Google Scholar]
  • 5. Fu Z, Achieving Efficient Cloud Search Services: Multi-keyword Ranked Search over Encrypted Cloud Data Supporting Parallel Computing, IEICE Transactions on Communications. E98-B (2015) 190–200. 10.1587/transcom.E98.B.190 [DOI] [Google Scholar]
  • 6. Dai Y, Yuan Y, A nonlinear conjugate gradient with a strong global convergence properties, SIAM J. Optim. 10 (2000) 177–182. 10.1137/S1052623497318992 [DOI] [Google Scholar]
  • 7. Fletcher R, Reeves C, Function minimization by conjugate gradients, Comput. J. 7 (1964) 149–154. 10.1093/comjnl/7.2.149 [DOI] [Google Scholar]
  • 8. Polak E, Ribière G, Note sur la convergence de directions conjugees, Rev. Fran. Inf. Rech. Opérat. 3 (1969) 35–43. [Google Scholar]
  • 9. Polyak BT, The conjugate gradient method in extreme problems, USSR Comput. Math. Math.Phys. 9 (1969) 94–112. 10.1016/0041-5553(69)90035-4 [DOI] [Google Scholar]
  • 10. Hestenes MR, Stiefel EL, Methods of conjugate gradients for solving linear systems, J. Res. Natl.Bur. Stand. Sect. B. 49 (1952) 409–432. 10.6028/jres.049.044 [DOI] [Google Scholar]
  • 11. Liu Y, Storey C, Effcient generalized conjugate gradient algorithms part 1: theory. J. Comput. Appl.Math. 69 (1992) 17–41. [Google Scholar]
  • 12. Fletcher R, Practical method of optimization, vol I: unconstrained optimization, 2nd edn Wiley,New York, 1997. [Google Scholar]
  • 13. Powell MJD, Nonconvex minimization calculations and the conjugate gradient method, Lecture Notes in Mathematics, vol. 1066, Spinger, Berlin, 1984, pp. 122–141. [Google Scholar]
  • 14. Gibert JC, Nocedal J, Global convergence properties of conjugate gradient methods for optimization. SIAM J. Optim. 2 (1992) 21–42. 10.1137/0802003 [DOI] [Google Scholar]
  • 15.Dai Y, Analyses of conjugate gradient methods, PH.D.thesis, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Sciences (in Chinese), 1997.
  • 16. Ahmed T, Storey D, Efficient hybrid conjugate gradient techniques, Journal of Optimization Theory and Applications. 64 (1990) 379–394. 10.1007/BF00939455 [DOI] [Google Scholar]
  • 17. Al-Baali A, Descent property and global convergence of the Flecher-Reeves method with inexact line search, IMA Journal of Numerical Analysis. 5 (1985) 121–124. 10.1093/imanum/5.1.121 [DOI] [Google Scholar]
  • 18. Hu Y F, Storey C, Global convergence result for conjugate gradient methods, Journal of Optimization Theory and Applications. 71 (1991) 399–405. 10.1007/BF00939927 [DOI] [Google Scholar]
  • 19. Yuan G, Wei Z, Zhao Q, A modified Polak-Ribière-Polyak conjugate gradient algorithm for large-scale optimization problems, IIE Transactions. 46 (2014), 397–413. 10.1080/0740817X.2012.726757 [DOI] [Google Scholar]
  • 20. Yuan G, Zhang M, A modified Hestenes-Stiefel conjugate gradient algorithm for large-scale optimization, Numerical Functional Analysis and Optimization. 34 (2013), 914–937. 10.1080/01630563.2013.777350 [DOI] [Google Scholar]
  • 21. Yuan G, Lu X, Wei Z, A conjugate gradient method with descent direction for unconstrained optimization, Journal of Computational and Applied Mathematics, 233 (2009), 519–530. 10.1016/j.cam.2009.08.001 [DOI] [Google Scholar]
  • 22. Yuan G, Lu X, A modified PRP conjugate gradient method, Annals of Operations Research. 166 (2009),73–90. 10.1007/s10479-008-0420-4 [DOI] [Google Scholar]
  • 23. Li X, Zhao X, A hybrid conjugate gradient method for optimization problems, Natural Science. 3(2011): 85–90. 10.4236/ns.2011.31012 [DOI] [Google Scholar]
  • 24. Yuan G, Modified nonlinear conjugate gradient methods with sufficient descent property for large-scale optimization problems, Optimization Letters. 3 (2009) 11–21. 10.1007/s11590-008-0086-5 [DOI] [Google Scholar]
  • 25. Huang H, Lin S, A modified Wei-Yao-Liu conjugate gradient method for unconstrained optimization, Applied Mathematics and Computation. 231 (2014) 179–186. 10.1016/j.amc.2014.01.012 [DOI] [Google Scholar]
  • 26. Yu G, Zhao Y, Wei Z, A descent nonlinear conjugate gradient method for large-scale unconstrained optimization, Applied mathematics and computation. 187 (2) (2007) 636–643. 10.1016/j.amc.2006.08.087 [DOI] [Google Scholar]
  • 27. Yao S, Wei Z, Huang H, A note about WYL’s conjugate gradient method and its applications, Applied Mathematics and computation. 191 (2) (2007) 381–388. 10.1016/j.amc.2007.02.094 [DOI] [Google Scholar]
  • 28. Hager WW, Zhang H, A new conjugate gradient method with guaranteed descent and an efficient line search, SIAM Journal on Optimization. 16(1) (2005) 170–192. 10.1137/030601880 [DOI] [Google Scholar]
  • 29. Wei Z, Yao S, Liu L, The convergence properties of some new conjugate gradient methods, Applied Mathematics and Computation. 183 (2006) 1341–1350. 10.1016/j.amc.2006.05.150 [DOI] [Google Scholar]
  • 30. Zhang L, An improved Wei-Yao-Liu nonlinear conjugate gradient method for optimization computation, Applied Mathematics and computation. 215 (6) (2009) 2269–2274. 10.1016/j.amc.2009.08.016 [DOI] [Google Scholar]
  • 31. Wei Z, Yu G, Yuan G, Lian Z, The superlinear convergence of a modified BFGS-type method for unconstrained optimization, Comput. Optim. Appl. 29 (2004) 315–332. 10.1023/B:COAP.0000044184.25410.39 [DOI] [Google Scholar]
  • 32. Yuan G, Wei Z, Convergence analysis of a modified BFGS method on convex minimizations, Computational Optimization and Applications. 47 (2) (2010) 237–255. 10.1007/s10589-008-9219-0 [DOI] [Google Scholar]
  • 33. Li M, Qu A, Some sufficient descent conjugate gradient methods and their global convergence, Computational and Applied Mathematics. 33 (2) (2014) 333–347. 10.1007/s40314-013-0064-0 [DOI] [Google Scholar]
  • 34. Zoutendijk Z, Nonlinear programming computational methods In: Abadie J. (ed.) Integer and nonlinear programming, North-Holland, Amsterdam, 1970, 37–86. [Google Scholar]
  • 35. Dolan ED, Morè JJ, Benchmarking optimization software with performance profiles, Math.Program. 91 (2002) 201–213. 10.1007/s101070100263 [DOI] [Google Scholar]
  • 36. Buhmiler S, Krejic´ N, Luzanin Z, Practical quasi-Newton algorithms for singular nonlinear systems, Numer. Algorithms 55 (2010) 481–502. 10.1007/s11075-010-9367-z [DOI] [Google Scholar]
  • 37. Gu G, Li D, Qi L, Zhou S, Descent directions of quasi-Newton methods for symmetric nonlinear equations, SIAMJ.Numer.Anal. 40 (2002) 1763–1774. 10.1137/S0036142901397423 [DOI] [Google Scholar]
  • 38. La Cruz W, Martínez. M, Raydan M. Spectral residual method without gradient information for solving large-scale nonlinear systems of equations, Math. Comp. 75 (2006) 1429–1448. 10.1090/S0025-5718-06-01840-0 [DOI] [Google Scholar]
  • 39. Kanzow C, Yamashita N, Fukushima M, Levenberg-Marquardt methods for constrained nonlinear equations with strong local convergence properties, J. Comput. Appl. Math. 172 (2004) 375–397. 10.1016/j.cam.2004.02.013 [DOI] [Google Scholar]
  • 40. Zhang J, Wang Y, A new trust region method for nonlinear equations, Math. Methods Oper. Res,58 (2003) 283–298. 10.1007/s001860300302 [DOI] [Google Scholar]
  • 41. Grippo L, Sciandrone M, Nonmonotone derivative-free methods for nonlinear equations, Comput. Optim. Appl. 37 (2007) 297–328. 10.1007/s10589-007-9028-x [DOI] [Google Scholar]
  • 42. Cheng W, A PRP type method for systems of monotone equations, Math. Comput. Modelling. 50 (2009) 15–20. 10.1016/j.mcm.2009.04.007 [DOI] [Google Scholar]
  • 43. La Cruz W, Raydan M, Nonmonotone spectral methods for large-scale nonlinear systems, Optim. Methods Softw. 18 (2003) 583–599. 10.1080/10556780310001610493 [DOI] [Google Scholar]
  • 44. Raydan M, The Barzilai and Borwein gradient method for the large scale unconstrained minimization problem, SIAM J. Optim. 7 (1997) 26–33. 10.1137/S1052623494266365 [DOI] [Google Scholar]
  • 45. Yu G, Guan L, Chen W, Spectral conjugate gradient methods with sufficient descent property for large-scale unconstraned optimization, Optim.Methods Softw. 23 (2) (2008) 275–293. 10.1080/10556780701661344 [DOI] [Google Scholar]
  • 46. Yuan G, Wei Z, Lu S, Limited memory BFGS method with backtracking for symmetric nonlinear equations, Math. Comput. Modelling. 54 (2011) 367–377. 10.1016/j.mcm.2011.02.021 [DOI] [Google Scholar]
  • 47. Li D, Fukushima M, A global and superlinear convergent Gauss-Newton -based BFGS method for symmetric nonlinear equations, SIAMJ.Numer.Anal. 37 (1999) 152–172. 10.1137/S0036142998335704 [DOI] [Google Scholar]
  • 48. Li Q, Li D, A class of derivative-free methods for large-scale nonlinear monotone equations, IMA J. Numer. Anal. 31 (2011) 1625–1635. 10.1093/imanum/drq015 [DOI] [Google Scholar]
  • 49. Solodov MV, Svaiter BF, A hybrid projection-proximal point algorithm, J. Convex Anal.6 (1999) 59–70. [Google Scholar]
  • 50. Yuan G, Zhang M, A three-terms Polak-Ribière-Polyak conjugate gradient algorithm for large-scale nonlinear equations, Journal of Computational and Applied Mathematics. 286 (2015) 186–195. 10.1016/j.cam.2015.03.014 [DOI] [Google Scholar]
  • 51. Solodov MV, Svaiter BF, A globally convergent inexact Newton method for systems of monotone equations, in: Fukushima M., Qi L. (Eds.),Reformulation:Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods, Kluwer Academic Publishers; 1998, pp. 355–369. [Google Scholar]
  • 52. Gomez-Ruggiero M, Martinez J, Moretti A, Comparing algorithms for solving sparse nonlinear systems of equations, SIAM J. Sci. Comput. 23 (1992)459–483. 10.1137/0913025 [DOI] [Google Scholar]
  • 53. Morè J, Garbow B, Hillström K, Testing unconstrained optimization software, ACM Trans. Math. Softw. 7 (1981) 17–41. [Google Scholar]
  • 54. Roberts SM, Shipman JJ, On the closed form solution of Troesch’s problem, J. Comput. Phys. 21 (1976) 291–304. 10.1016/0021-9991(76)90026-7 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All data are available in the paper.


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES