Abstract
The paper describes an application of the p-regularity theory to Quadratic Programming (QP) and nonlinear equations with quadratic mappings. In the first part of the paper, a special structure of the nonlinear equation and a construction of the 2-factor operator are used to obtain an exact formula for a solution to the nonlinear equation. In the second part of the paper, the QP problem is reduced to a system of linear equations using the 2-factor operator. The solution to this system represents a local minimizer of the QP problem along with its corresponding Lagrange multiplier. An explicit formula for the solution of the linear system is provided. Additionally, the paper outlines a procedure for identifying active constraints, which plays a crucial role in constructing the linear system.
Keywords: quadratic programming, singular problems, p-regularity, 2-factor-operator
MSC: 65H10, 90C20, 65K05, 90C30
1. Introduction
Consider the nonlinear equation with the mapping F defined by:
| (1) |
where is a matrix, is a vector, and is the map defined for by:
| (2) |
where is an symmetric matrix for .
We also consider the quadratic programming (QP) problem with inequality constraints:
| (3) |
where Q is an symmetric matrix, A is an matrix, , and .
The paper describes an application of the p-regularity theory to nonlinear equations with the mapping F introduced in (1) and to the quadratic programming problem (3).
In recent years, there has been growing interest in nonlinear problems, including quadratic and polynomial equations, as well as nonlinear optimization problems, attracting specialists from various disciplines (see, for example, refs. [1,2,3,4] and references therein). Furthermore, it was observed that nonlinear problems are closely related to singular problems, as demonstrated in [5]. In fact, it has been discovered that essentially nonlinear problems and singular problems are locally equivalent. In this work, we aim to provide a theoretical foundation for this claim by introducing several auxiliary concepts as proposed in [5].
Definition 1.
Let V be a neighborhood of in , and let be a neighborhood of 0. A mapping , where , is considered essentially nonlinear at if there exists a perturbation of the form:
such that no nondegenerate transformation of coordinates , where , satisfies , , where is the identity map in , and:
Definition 2.
We say that the mapping F is singular (or degenerate) at if it fails to be regular, meaning its derivative is not onto:
The relationship between the notions of essential nonlinearity and singularity is established in Theorem 1, which was derived in [5].
Theorem 1.
Let V be a neighborhood of in . Suppose is and that is a solution of . Then F is essentially nonlinear at the point if and only if F is singular at the point .
The work presented in [5] primarily focuses on the construction of p-regularity and its applications in various areas of mathematics. However, it does not specifically cover quadratic nonlinear equations and quadratic programming problems. The current paper builds upon the foundation of the p-regularity theory established in [5] but introduces novel results. The main objective of this paper is to explore the key aspects of nonlinear problems, with a particular emphasis on systems of quadratic equations and quadratic programming problems that may involve singular solutions.
Specifically, we begin by considering the nonlinear equation . One of the main goals of the paper is to derive the exact formula for a solution of the nonlinear equation using the special form of the quadratic mapping F defined in (1). We demonstrate how to use a construction of a special 2-factor-operator to transform the original problem into a system of linear equations. The construction of the 2-factor-operator combines the mapping F with its first derivative .
In the second part of the paper, we apply a similar approach to the quadratic programming problem (3) in order to derive explicit formulas for the solution , where represents a local minimizer of the QP problem and is the corresponding Lagrange multiplier. Namely, using the special form of the QP problem and the 2-factor-operator, we reduce the system of optimality conditions for the QP problem to a system of linear equations, with the point as its solution. The paper also describes a procedure for identifying the active constraints, which plays a vital role in constructing the linear system.
Although there is literature on solutions of degenerate systems of quadratic equations, the approach presented in this paper is novel and distinct from the methods proposed by other authors. This approach can be applied to various problems and areas of mathematics where the problem involves solving a degenerate equation with a quadratic mapping F. Such nonlinear problems can arise in the numerical solutions and analysis of ordinary differential equations, partial differential equations, optimal control problems, algebraic geometry, and other fields. In the second part of the paper, we specifically focus on using the methods developed in the first sections to solve the QP problem (3). The quadratic programming problems have attracted the attention of many researchers and scientists, so there is an extensive body of literature on the topic. Some publications in this area include [6,7,8,9,10,11,12].
The outline of the paper. The main contribution and novelty of the paper are in the exact formulas for a solution of a nonlinear equation and of the quadratic programming problems, presented in Section 3 and Section 5, respectively.
In Section 2, we recall the main definitions of the p-regularity theory, as presented in [5], including the special case of . Additionally, we introduce the p-factor method for solving singular nonlinear equations of the form and describe various versions of the 2-factor method.
Section 3 presents some of the key results of the paper, focusing on the application of a modified 2-factor method to solve the nonlinear equation with the mapping F defined as where is a matrix, is a vector, and is defined by (2). In this section, we introduce multiple approaches to obtain exact formulas for a solution to the nonlinear equation , demonstrating that the proposed methods converge to a solution of the nonlinear equation in just one iteration.
Section 4 focuses on an auxiliary result used in other parts of the paper. We present a theorem that describes the properties of a special mapping , which enables us to propose a procedure for determining r linearly independent vectors , , at the solution of , without needing to know the exact value of . This procedure relies on information about the system of vectors at some point x within a small neighborhood of .
Section 5 presents other novel results, focusing on deriving exact formulas for a solution of quadratic programming problems. The section is divided into three parts. In Section 5.1, we consider regular quadratic programming problems and propose three approaches to solving the QP problem and obtaining a formula for its solution. These approaches are based on the construction of the 2-factor-operator. Section 5.2 addresses the issue of identifying the active constraints and proposes strategies for numerically determining the set of active constraints . These techniques are then applied in the final part, Section 5.3, to address degenerate QPs. The paper concludes with some closing remarks in Section 6.
Notation. Let denote the rows of the matrix A in problem (3), and let , so that and for .
The active set at any feasible point of problem (3) is the set of indices of the active constraints at , i.e., .
Furthermore, denotes the null-space (kernel) of a given linear operator , and is its image space.
Let be a bilinear symmetric mapping. The 2-form associated with B is the map defined by for . We also use the following notation: and . We denote by and neighborhoods of a point , where is an -neighborhood of , i.e., an open ball of radius centered at .
The notation for the scalar (dot) product of vectors x and y in , used in the paper, is .
We denote by the linear span of the given vectors . We also denote by the distance between a point x and a set S.
2. Elements of the -Regularity Theory
We begin this section with the main definitions of the p-regularity theory, which are given in [5]. The primary focus is on the sufficiently smooth mapping F from to , defined as:
| (4) |
where for . After presenting the general case, we focus on the specific case of . We introduce the definitions of the 2-regular mapping and the 2-factor-operator, which play a central role in the subsequent sections.
2.1. The Main Definitions and Constructions of the p-Regularity Theory
Throughout this section, we consider the nonlinear equation:
| (5) |
where F is defined in Equation (4). Let represent a solution to the nonlinear Equation (5).
The mapping F is called regular at if:
| (6) |
or in other words, if:
where is the Jacobian matrix of the mapping F at . Conversely, the mapping F is called nonregular (irregular, degenerate) if the regularity condition (6) is not satisfied.
Let the space be decomposed into the direct sum:
| (7) |
where is defined as the closure of the image of the first derivative of F evaluated at , and p is chosen as the minimum number for which Equation (7) holds.
The remaining spaces are defined as follows. Let , and let be a closed complementary subspace to . Let be the projection operator onto along . Define as the closed linear span of the image of the quadratic map . More generally, we define inductively for as:
where is a choice of a complementary subspace for with respect to Y, , and is the projection operator onto along with respect to Y, . Finally, we let
Define the following mappings:
| (8) |
where is the projection operator onto along with respect to for . Then F can be represented as or equivalently as
Definition 3.
The linear operator , where , , is defined by:
and is called the p-factor operator.
Consider the nonlinear operator defined by:
Notice that .
Definition 4.
The p-kernel of the operator at the point is the set defined by:
where:
Please note that , where:
is the k-kernel of .
Definition 5.
A mapping F is called p-regular at along h if .
Definition 6.
A mapping F is called p-regular at if it is p-regular along all or .
Now, we will focus on the special case of , which we are using in the paper. We denote the image of the Jacobian matrix by : and the orthogonal complementary subspace of in by . Then:
We also denote an matrix of the orthogonal projection onto in by ,
Similarly to Equation (8), we introduce the mappings:
The p-factor operator plays the central role in the p-regularity theory. We give the following definition of the p-factor-operator for .
Definition 7.
We define a 2–factor-operator of the mapping F at with respect to some vector , , as a linear operator from to , defined by one of the following equations:
(9)
(10)
(11)
Now we are ready to introduce another very important definition of the 2-regularity theory.
Definition 8.
The mapping F is called 2-regular at the point with respect to the element h if the image of a 2–factor-operator, defined by one of the Equations (9)–(11) is equal to .
Definition 9.
The mapping F is called 2-regular at if it is 2-regular at with respect to all the elements h from a set that is defined as:
2.2. The p-Factor-Method for Solving Singular Nonlinear Equations
In this section, we introduce the p-factor-method for solving the singular nonlinear equation . Then we consider the special case of and describe several versions of the 2-factor-method.
Consider Equation (5) in the case when mapping is singular at . In this case, the p-factor method is an iterative procedure defined by:
| (12) |
where , for , and vector h, , is chosen in such a way that the p-factor operator is nonsingular, which implies that the mapping F is p-regular at along h. The following theorem is valid for the p-factor-method (12).
Theorem 2.
Assume that mapping and there exists vector h, , such that the p-factor operator is nonsingular. Given a point , where is sufficiently small and is a neighborhood of , the sequence defined by Equation (12) converges quadratically to the solution of (5):
(13) where is an independent constant.
Now, we are ready to describe several versions of the 2-factor-method.
For solving singular nonlinear Equation (5), the following iterative method, called the 2-factor-method, was proposed in [13]:
| (14) |
where the vector h, , is chosen in such a way that matrix is invertible.
The following theorem states the convergence properties of the 2-factor-method (14).
Theorem 3.
Given a mapping , let be a solution of Equation (5). Assume that there exists a vector such that and F is 2-regular at the point with respect to the vector h with the 2-factor-operator defined by (10).
Then there is a neighborhood of in such that for any , the sequence generated by the 2-factor-method (14) converges to and:
(15) where is some constant.
Proof.
Since is the orthoprojector onto subspace , then for the mapping:
we have . Moreover, because and the mapping F is 2-regular with respect to the vector h, by Definition 8 with defined by (10), we obtain that . Hence, the matrix is invertible.
Therefore, the 2-factor-method given in (14) is an application of Newton’s method to system in a sufficiently small neighborhood of . Then the statement of the theorem follows from the properties of Newton’s method [14] (Proposition 1.4.1). □
Now, we will introduce a modified version of the 2-factor-method (14). Assume that there exists a vector such that and the matrix is invertible. Then for solving Equation (5), we can use the following modified 2-factor-method:
| (16) |
The following theorem states the convergence properties of method (16).
Theorem 4.
Given a mapping , let be a solution of Equation (5). Assume that there exists a vector such that , , and F is 2-regular at the point with respect to the vector h, where the 2-factor-operator is defined by (11).
Then there is a neighborhood of in such that for any , the sequence generated by the 2-factor-method (16) converges to , and relation (15) holds with some .
The proof is similar to one of Theorem 3.
Now we introduce another version of the 2-factor-method.
Assume that the following conditions hold:
| (17) |
Note that if conditions (17) are satisfied, then to solve Equation (5), we can use the following modified 2-factor-method:
| (18) |
For numerical realization of the 2-factor-method in the form (18), we only have to construct vector h satisfying conditions (17). Specifics of some problems allow us to choose vector h without any knowledge of the solution . We discuss the choice of the vector h in the following sections of the paper.
The following theorem states the convergence properties of method (18).
Theorem 5.
Given a mapping , let be a solution of Equation (5). Assume that there exists a vector such that and conditions (17) are satisfied.
Then, there is a neighborhood of in such that for any , the sequence generated by the 2-factor-method (18) converges to and relation (15) holds with some .
The proof is similar to one of Theorem 3.
3. Nonlinear Equations with Quadratic Mappings: the Exact Solution Formula
In this section, we consider the mapping F defined by Equation (1) as follows:
where is a matrix, is a vector, and is the map defined by (2). The mapping B is twice continuously differentiable [15], and its derivatives are given by and for some arbitrary . Let denote a solution of the equation .
We will now illustrate the application of the 2-factor method (18) for solving the nonlinear equation with the mapping F defined by (1). We will present multiple approaches to obtain an exact formula for , with the first approach being a specific case of the second approach. Additionally, we will show that for the mapping F, the method (18) converges to in just one iteration.
First approach to obtain an exact formula for the solution .
For the mapping F defined by (1), the assumptions (17) of Theorem 5 can be simplified to the existence of a vector h that satisfies the following conditions:
| (19) |
Under these assumptions (19), for the mapping F defined by (1) and a given point , the first iteration of the 2-factor-method (18) can be written as:
which is equivalent to:
Using the property , the last equation implies a one-step method for calculating and, consequently, finding the solution :
| (20) |
where the vector h satisfies conditions (19).
The numerical determination of the vector h depends on the specific characteristics of the problem. Alternatively, it can be obtained using the same method as described in the third approach below, which involves transforming the initial system into a system that is completely degenerate at the point .
Second approach to obtain an exact formula for the solution .
Now we present an alternative approach for obtaining a formula for the solution of the equation using the same mapping F defined by (1). This second approach is applicable to a broader variety of problems compared to the first approach.
Let denote the projector onto , and let denote the projector onto , which is the orthogonal complementary subspace of in . We note that and:
Then for the mapping F defined in (1),
Assume that there exists a vector satisfying the conditions:
| (21) |
Given the definition of , it follows that . Substituting this into (1), we obtain . Hence, the point satisfies the following identities:
By adding these equations and assuming (21), we obtain the exact formula for the solution :
| (22) |
Remark 1.
In the case when and, hence, , assumptions (21) become (19), and Equation (22) reduces to (20).
Example 1.
Consider mapping given by:
(23) We can represent the mapping F in the form (1) with:
The equation has a locally unique solution . In this example, and . Hence, by Remark 1, we apply Equation (20) with to obtain:
as claimed.
In a numerical implementation, an additional procedure is required to construct the vector h. Since the exact point is not known in advance, we only assume that a sufficiently small neighborhood of is provided to apply the procedure.
Third approach to obtain an exact formula for the solution .
While the first two approaches rely on knowledge of the element h, which is determined by , the third approach does not require such knowledge. Instead, all we need is for the starting point to belong to a sufficiently small neighborhood of . Specifically, we have , where is sufficiently small.
Suppose that at the point , the first r vectors are linearly independent, where is defined in (4) for . Assume also that the other vectors are linear combinations of the first r vectors. Therefore, there exist coefficients such that:
Let us introduce the subspace defined by:
We denote the orthogonal projection on the subspace as . Then, there exist coefficients such that:
In addition, introduce the notation:
Then:
| (24) |
Notice that is also a solution of the equation , where is defined as:
The definition of implies that is 2-regular at the point . In the case that some of the vectors , are not zero vectors, transformation (24) can be used to reduce those vectors to zero vectors. This ensures that for all . Therefore, without loss of generality, we can assume that the mapping satisfies for . An example of a mapping that satisfies these conditions is:
where , , , , , and
Suppose there exist vectors , , and , and indices , such that the system:
is linearly independent.
Then the mapping defined by:
| (25) |
has as its zero, that is . At the same time, compared to the Jacobian matrix of , the matrix:
is nonsingular. We can, therefore, consider the method:
| (26) |
Theorem 6.
Given a mapping , let be a solution of Equation (5). Assume that there exist vectors , , and , such that mapping defined in (25) is nonsingular at . Let , where is a neighborhood of and is sufficiently small.
Then the sequence , defined by (26) is convergent to the point with the quadratic rate of convergence, that is:
where is an independent constant.
Using definition of mapping F given by Equation (1), mappings introduced in (4) will have the following form:
where is an symmetric matrix, , and ,
Given an initial point , we use the iterative method (26) to obtain:
Because matrix is symmetric for any index i, then for any index j, we have:
Therefore,
| (27) |
Example 1
(Continuation). Consider mapping defined in (23):
where
In this example, is a solution of and:
Therefore, mapping defined in (25) takes the form:
where h is chosen in such a way that the matrix is nonsingular, and vectors are not used. For example, we can take . Then Equation (27) has the form:
(28) which is a solution of in this example.
The approaches described above can be modified to derive other methods for solving the equation . For example, using the equation , where , we obtain the following method:
The sequence converges to under the assumption that is nonsingular. In this modification, unlike the second approach, we can construct an element h without the knowledge of the point , based on the information at an initial point .
Applying the modified method to Example 1, we obtain the same formulas and results as shown in Equation (28) above. To implement this approach, it is necessary to determine the vectors , , which correspond to linearly independent vectors This can be achieved using information at a point , where is sufficiently small. If the assumption of p-regularity is satisfied, the identification of linearly independent vectors is performed using the method described in the next section.
4. Procedure for Identifying Zero Elements
The procedure for identifying zero elements could be used to implement the methods described in the previous sections numerically. Let be defined as:
| (29) |
In this section, we present a theorem that describes the properties of a special mapping , which allows us to propose the method for determining r linear independent vectors , at the solution of . This procedure is based on the information about the system of vectors at some point x in a small neighborhood of . As a result, we can define the mapping with the first r components , corresponding to the linearly independent vectors ,
Let be 2-regular at the point . For some , where is sufficiently small, we define the following mappings:
and:
| (30) |
where denotes the distance between an element x and the set S. Note that if , then .
The mapping is used to determine the maximum number r of linearly independent vectors in the system using a special procedure that relies on the information about the mapping at the point . The properties of the mapping are stated in the following theorem, and the proof can be found in [16].
Theorem 7
(Minorant theorem). Let be 2-regular at the point , and . Then there exist constants , and such that the following inequality holds for any :
where function is defined in (30).
In addition to the properties of the mapping given in Theorem 7, we also need the following lemma (for the proof, see [16]).
Lemma 1.
For the non-negative mappings and , let the following inequalities hold:
where , , and σ are positive constants, with and .
Then, there exists a sufficiently small such that one of the following conditions holds:
- 1.
If for all , then
- 2.
If for all , then
Remark 2.
Based on the assumptions of Lemma 1, there exists a sufficiently small such that if the inequality is satisfied for any , then the inequality is satisfied for all , and hence .
Similarly, if the inequality is satisfied for any , then the inequality is satisfied for all , and hence .
Now we are ready to introduce an iterative method that determines indices corresponding to the linearly independent vectors ,
Method for determining linearly independent gradients at (identifying zero elements).
Using Lemma 1 and Remark 2, for a sufficiently small , , and , consider two possible cases:
Case 1. .
Case 2. .
In Case 1, according to Remark 2, it follows that , whereas in Case 2, we have .
In addition to the properties of the mapping
Let be defined by (29) and be a solution of . Let x be in , where is sufficiently small. Define function using Equation (30).
Step 1. Identify the smallest index in the set such that . According to Case 2 above, this implies that .
- Step 2. Use Step 1 to identify if the set has at least one index j such that . If it does not, the method is finished. Otherwise, identify the next smallest index in the set such that the following condition is satisfied:
According to Case 2 above, it means that the vectors and are linearly independent.
Step k. By this step, we have identified linearly independent vectors , …, , where . Use Step 1 to identify if the set has at least one index j such that . If it does not, the method is finished. Otherwise, identify the next smallest index such that the following condition is satisfied:
The inequality implies that the vectors , are linearly independent.
Repeat Step k until the method is finished.
Without loss of generality, assume that the first r vectors are linearly independent and define mapping as:
| (31) |
where vectors are defined in such a way that:
Namely, let:
be a linear combination of the vectors , …, Coefficients are determined by solving the following system of equations:
In addition, define to be a nonsingular matrix of the form:
Let:
Define the following vectors:
These vectors allow us to transform the mapping to , where and The purpose of this transformation is to simplify the structure of the projection operators.
We present a simple example to illustrate an application of the proposed method.
Example 2.
Let , , where:
Then is a solution of . Take and consider . The Jacobian matrix of F is:
It is easy to see that vectors and are linearly dependent. We can check this by applying the method introduced above.
By using Equation (30), we define function , where:
and α is the angle between vectors and . Note that:
and hence:
Using , we obtain:
We are ready to apply the method described above.
In Step 1, we obtain because:
Hence, vector and .
Then in Step 1 of the method with vector , we also verify whether the following inequality holds:
Using point , we obtain:
Therefore, we conclude that .
Thus, in this example, the mapping defined in (31) has the form , where and .
5. Quadratic Programming Problems
In this section, we consider the quadratic programming (QP) problem (3):
where Q is an symmetric matrix, A is an matrix, , and . The Lagrangian for problem (3) is defined by:
| (32) |
where is the vector of Lagrange multipliers and is the ith row of the matrix A. The Karush-Kuhn-Tucker (KKT) conditions [17] are satisfied at with some if:
| (33) |
The point at which relations (33) are satisfied is called a stationary point or a KKT point. Observe that is a solution of the following system:
| (34) |
We denote by the set of indices of the active constraints at :
The following constraint qualification is used in the paper.
Definition 10
(Linear independence constraint qualification). The linear independence constraint qualification (LICQ) holds at a feasible point if the row-vectors , , corresponding to the active at constraints, are linearly independent.
The modified second-order sufficient conditions (MSOSC) state that there exist a Lagrange multiplier vector and a scalar such that:
| (35) |
for all satisfying:
We divide the presentation in this section into three parts. We start by considering regular QP problems in Section 5.1. Then, in Section 5.2, we discuss the issue of identifying the active constraints and propose numerical strategies for determining the set . We apply these techniques to degenerate QP problems in Section 5.3.
5.1. Regular Quadratic Programming
In this section, we consider regular quadratic programming (QP) problem (3). In other words, we assume that the Linear Independence Constraint Qualification (LICQ) and the Mangasarian-Fromovitz Constraint Qualification (MFCQ) conditions (35) hold. Recall that A is an matrix of coefficients representing the constraints in problem (3). Without loss of generality, assume that the first p constraints are active at , so that:
Then we can rewrite the matrix A in the following form:
| (36) |
where is a matrix of coefficients corresponding to the active constraints at , and is an matrix of coefficients corresponding to the nonactive constraints at . It is important to note that we do not have prior knowledge of the set . We will discuss possible numerical realizations to approximate the set of active constraints in Section 5.2. Additionally, we introduce the following notation associated with the active constraints at the point :
Similarly,
In the following subsections, we will introduce three approaches to solving the QP problem (3) and provide formulas for the solution.
5.1.1. First Approach to Solving the QP Problem
In this subsection, we present an approach to solving the QP problem and obtaining a formula for its solution. This approach is based on the construction of the 2-factor-operator. For our consideration below, we need the following lemma.
Lemma 2.
Let V be an matrix, G be a matrix, such that the columns of are linearly independent, L be an matrix, be a diagonal full rank matrix, and:
(37) Then matrix Γ defined by:
(38) is nonsingular.
Proof.
To prove the lemma, we must prove that the matrix defined by (38) has zero nullspace. Consider the following system that defines the nullspace of in the form of a vector , where , , and :
(39) Since is a full-rank diagonal matrix, the third equation in the system (39) implies that . Then, using the first equation, we obtain:
Consequently, ; otherwise, , which contradicts the assumption (37) of the lemma. Therefore, the first equation in (39) reduces to , and since the columns of are linearly independent, we obtain . Thus, the matrix (38) has a zero nullspace, , and therefore, is nonsingular. This concludes the proof of the lemma. □
Let , and mapping be defined in (34), so that . Introduce mappings and as:
Recall that matrix is defined in (36), and introduce vector such that:
where , ,
Define mapping as:
| (40) |
Recall that is the ith row of the matrix A and . Then:
and mapping defined in (40) can be rewritten as:
Introduce matrix . Then, taking into account the definition of and , we obtain:
Observe that if is a solution of (34), it is also a solution of or, equivalently,
| (41) |
To obtain the formula for the solution , we rewrite the system (41) as:
Assuming that LICQ and MSOSC hold and apply Lemma 2, we obtain that the matrix:
is invertible and obtain the formula for :
| (42) |
5.1.2. Second Approach to Solving the QP Problem
Assume that we can estimate the set , which is in our notation . Taking into account that and that , system (34) can be reduced to the following one:
| (43) |
which can be written as:
| (44) |
Under the assumptions LICQ and MSOSC, the following matrix is invertible,
and system (44) yields the formula for the solution :
| (45) |
Remark 3.
System (41) reduces to system (43) by removing equations , corresponding to the nonactive constraints. Similarly, Equation (42) reduces to (45).
Remark 4.
We note that solutions of QP problems have the following specific property: if is a solution of the QP problem and for the vector , then the points are also solutions of the QP problem.
5.1.3. Examples
In this section, we illustrate the two described approaches with examples. Namely, we consider the construction of system (41) required for the first approach. Then we illustrate using the exact formula (45) derived in the second approach.
Example 2.
Consider the problem:
(46) The matrix A in this example is and The solution to this problem is the point with and . Hence, , , and . Moreover,
By choosing and , the system (41) reduces to the linear system:
Solving the system yields , as claimed.
Now, let us illustrate the second approach. Specifically, using the formula (45) for the solution of problem (46) with , we obtain:
as claimed.
Example 3.
Consider the problem:
(47) The solution to this problem is the point with and . Hence, . Moreover,
By choosing and , the system (41) reduces to the following linear system for problem (47):
Solving yields , as claimed.
To illustrate the second approach, we rewrite the exact formula (45) for the solution of problem (47) in the form:
| (48) |
as claimed.
5.1.4. Third Approach to Solving the QP Problem
In this subsection, we present another approach to solving the QP problem. A formula that we obtain for the solution of the QP problem is also based on the construction of the 2-factor-operator.
First, we replace the inequality constraints in the QP problem with equality constraints of the form:
where . We then define the Lagrangian as follows:
| (49) |
Introduce the notation:
Then the point is a solution of the following system:
| (50) |
The Jacobian matrix of the system (50) is given by:
Then with , we obtain
Assuming that LICQ and MSOSC hold, matrix is singular if and only if the strict complementarity condition does not hold. In other words, the set of indices of the weakly active constraints,
is not empty.
Let be the matrix of the orthoprojector onto , and be the matrix of the orthoprojector onto . Note that will be a projector onto the linear part of the mapping , while will be a projector onto the quadratic part of .
Introduce vector such that . Then,
or
and is defined by:
| (51) |
Observe that , i.e., .
Define H as a diagonal matrix with elements in the rows corresponding to the components of the vector , and K as a diagonal matrix with elements of the vector , so that:
Then:
The 2-factor-operator for the mapping is defined as:
or
We choose a vector according to (51) so that matrix:
is nonsingular. Then can be determined using the following formula:
5.2. Identification of the Active Constraints
In this section, we address the issue of identifying the active constraints and propose strategies for numerically identifying the set of active constraints .
We begin by considering the mapping , where . We can also represent h as an n-vector of functions , …, , such that .
Theorem 8.
Let be 2-regular at the point , and let be a sufficiently small neighborhood of in . Assume that there exists a function such that and for all , we have:
(52) where are independent constants.
Then there exists a sufficiently small δ such that , and for any and any point , the following holds:
Either , which implies that
Or , which implies that
Proof.
The proof is similar to the one in [5]. □
Let:
where denotes the distance between a vector a and a set S. It turns out that if we take , and g is 2-regular at , then inequality (52) holds with .
Theorem 8 can be used for the numerical determination of the set of active constraints in the QP problem. To apply Theorem 8, we need to define a function that satisfies the conditions of the theorem. Recall that for QP problem (3), we denote the Lagrange function defined in (32) by .
Under the assumptions of LICQ and MSOSC, the following holds for and :
where is sufficiently small (see, for example, [18]). Hence, the required function can be defined by:
Then, according to Theorem 8, for every , if:
then it follows that .
Moreover, if we introduce the function:
where: , then satisfies the estimate:
for , where is a sufficiently small number.
Then, for any , if:
then . Here, represents the set of constraints that are weakly active, i.e, for which the associated multipliers are equal to zero, while denotes the set of constraints that are strongly active at the point , i.e., the associated Lagrange multipliers are positive.
5.3. General Case
Consider the Lagrange function in the form:
In this case, if is a solution of problem (3), then there exist multipliers and , not all zeros, such that , , and the point is a solution of the following system:
| (53) |
Introduce the notation:
| (54) |
We are making the following assumption for the rest of the section.
Assumption A1.
Assume that there exists and a sufficiently small such that for any , the following holds:
Remark 5.
It is easy to see that for any ,
where is an independent constant.
As follows from Assumption 1 and Theorem 8, for those indices that satisfy the inequalty:
we make a conclusion that .
We can illustrate Assumption 1 with the following examples, where Assumption 1 holds.
Example 4.
This example illustrates a choice of the function ξ in a more general setting.
Consider mapping F defined by either:
or
In each of the two cases, .
Introduce function defined as:
It follows that for any , the inequality:
holds, where C is an independent constant.
Example 5.
Consider the problem:
(55) The solution to this problem is the point , so . Moreover, the system (53) in this example is given by:
(56)
We also introduce the function , which can be defined using Equation (54), but in this case, we define it as .
Under Assumption 1, we use the function to determine the set . We also take into account the fact that the constraints in the problem are linear and the rank of the matrix
is 1. This implies that the constraints are linearly dependent. Consequently, we can eliminate, for instance, the second constraint from problem (55) and simplify system (56) to the following one:
Now, by introducing:
we construct the modified 2-factor-system:
This system implies that the solution is .
Now we will demonstrate the application of the approach described in Section 5.1.2 to problem (55). By removing the first constraint, we obtain a regular QP problem with . Additionally, in this example,
Then application of Equation (45) derived in Section 5.1.2 yields:
as claimed.
There are various directions in which the approach proposed in this paper can be extended. The next example illustrates a degenerate QP problem, in which MSOSC does not hold at the solution. However, an approach proposed in this paper can be applied to find a solution to this problem. Moreover, the solution set is locally not unique.
There are various directions in which the approach proposed in this paper can be extended. The next example illustrates a degenerate QP problem in which MSOSC does not hold at the solution. However, the approach proposed in this paper can still be applied to find a solution to this problem. It is worth noting that the solution set in this case is locally not unique.
Example 6.
Consider the problem:
(57) The solution to this problem is the set of points . We observe that the objective function in this example is satisfied as an equality for any . Additionally, the system (53) for this example consists of one linear equation and three quadratic equations:
Denote the projection of the point x onto the set by . Also, define the notation . For any point , we have the inequality:
where and is sufficiently small.
Consider, for example, the point .
In problem (57), we replace the inequality with the equation , where . We then introduce the Lagrange function in the form of (49) as follows:
If is a Lagrange multiplier corresponding to the solution , then the point is a solution of the following system:
| (58) |
The Jacobian matrix of this system is given by:
This Jacobian matrix becomes singular at . To overcome this singularity, we can apply the approach described in the paper. Specifically, we notice that for . Moreover, the point is one of the solutions of the system defined in (58), corresponding to a solution of the QP problem (57).
Additionally,
The 2-factor-operator of with respect to the vector , which is defined similarly to the operator in Equation (11), is given by:
Note that the 2-factor-operator is nonsingular and the system:
has the point as its regular solution.
6. Conclusions
The paper focused on applying the p-regularity theory to nonlinear equations with quadratic mappings and quadratic programming (QP) problems. The first part of the paper used the special structure of the nonlinear equation and the construction of a 2-factor operator to derive a formula for the solution of the equation. In the second part, the QP problem was reduced to a system of linear equations using a 2-factor-operator. The solution of the system is a local minimizer of the QP problem with a corresponding Lagrange multiplier. The formula for the solution of the linear system was given. The paper also described a procedure for identifying the active constraints, which was used in constructing the linear system.
The paper primarily focuses on the case where the matrix is degenerate at the solution of the nonlinear equation . However, the matrix does not need to be degenerate. While we do not explicitly address the identification of degeneracy at a solution point, it is possible to determine the degeneracy of the matrix by examining the behavior of the mapping F in a small neighborhood of the solution . Specifically, a function can be defined, such that:
for some natural number p and constants and . Based on the conclusion about the degeneracy of the matrix , an appropriate method can be chosen to solve the system of equations , as stated in the following theorem.
Theorem 9.
Let be such that , and let there exist , where is sufficiently small. Then we have the following two cases:
In the first case, for all , we have:
In this case, , indicating that F is not degenerate at .
In the second case, there exists an index such that:
In this case, , indicating that F is degenerate at .
Certainly, the construction of the function is an important consideration. One approach to constructing such a function is provided in the following lemma, specifically for the case of .
Lemma 3.
Let and assume that either exists, or for any , there exists with . Then, there exists a sufficiently small such that the following inequality holds for all :
where C is a positive constant.
Based on this lemma, one can choose the function .
It is worth noting that the proposed approach also covers the case where the system of equations consists of both linear and quadratic equations. Moreover, the approach can be extended to solve multilinear equations with polynomials of degree p, given by the equation:
where is k-multilinear mapping for . Additionally, polynomial programming problems can be formulated as follows:
where are polynomial mappings.
There are various possible directions for future research work, based on the results obtained in this paper. While the focus of the current work was on obtaining exact formulas for the solutions of nonlinear equations with quadratic mappings and quadratic programming problems, it would be interesting to generalize the proposed approaches to other classes of problems, including systems of equations with both linear and quadratic mappings. Another direction would be an extension of presented methods to polynomial equations and polynomial programming problems. Another direction of future research could be focusing on numerical studies and the implementation of the methods described in the paper.
Acknowledgments
The authors thank the anonymous reviewers for their careful reading of our manuscript and for their insightful comments and suggestions that helped us improve the quality of the paper.
Author Contributions
Conceptualization, A.A.T.; methodology, O.B., A.P. and A.A.T.; validation, O.B., A.P. and A.A.T.; formal analysis, O.B., A.P. and A.A.T.; investigation, O.B., A.P. and A.A.T.; resources, O.B. and A.A.T.; writing—original draft preparation, O.B. and A.P.; supervision, O.B. and A.A.T.; project administration, A.P.; funding acquisition, A.P. and A.A.T. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Data Availability Statement
No new data were created or analyzed in this study. Data sharing is not applicable to this article.
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript.
Funding Statement
This research was funded by the Ministry of Education and Science, grant number 144/23/B.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Barvinok A., Rudelson M. When a system of real quadratic equations has a solution. Adv. Math. 2022;403:108391. doi: 10.1016/j.aim.2022.108391. [DOI] [Google Scholar]
- 2.Bereznev V.A. Theoretical and Applied Problems of Nonlinear Analysis. Russian Academy of Sciences, Computing Center; Moscow, Russia: 2007. [Google Scholar]
- 3.Li N., Zhi L. Improved two-step Newton’s method for computing simple multiple zeros of polynomial systems. Numer. Algorithm. 2022;9:19–50. doi: 10.1007/s11075-022-01253-7. [DOI] [Google Scholar]
- 4.Poirier A., Torres J. Approximating roots by quadratic iteration. Proyecciones J. Math. 2023;42:407–431. doi: 10.22199/issn.0717-6279-5447. [DOI] [Google Scholar]
- 5.Tret’yakov A.A., Marsden J.E. Factor–analysis of nonlinear mappings: p–regularity theory. Commun. Pure Appl. Anal. 2003;2:425–445. [Google Scholar]
- 6.Anitescu M. A superlinearly convergent sequential quadratically constrained quadratic programming algorithm for degenerate nonlinear programming. SIAM J. Optim. 2002;12:949–978. doi: 10.1137/S1052623499365309. [DOI] [Google Scholar]
- 7.Fletcher R. Resolving degeneracy in quadratic programming. Degeneracy in optimization problems. Ann. Oper. Res. 1993;46/47:307–334. doi: 10.1007/BF02023102. [DOI] [Google Scholar]
- 8.De Marchi A. On a primal-dual Newton proximal method for convex quadratic programs. Comput. Optim. Appl. 2022;81:369–395. doi: 10.1007/s10589-021-00342-y. [DOI] [Google Scholar]
- 9.Permenter F. Log-domain interior-point methods for convex quadratic programming. Optim. Lett. 2023;17:1613–1631. [Google Scholar]
- 10.Gould N.I.M., Orban D., Robinson D.P. Trajectory-following methods for large-scale degenerate convex quadratic programming. Math. Program. Comput. 2013;5:113–142. doi: 10.1007/s12532-012-0050-3. [DOI] [Google Scholar]
- 11.Yamakawa Y., Takayuki O. A stabilized sequential quadratic semidefinite programming method for degenerate nonlinear semidefinite programs. Comput. Optim. Appl. 2022;83:1027–1064. doi: 10.1007/s10589-022-00402-x. [DOI] [Google Scholar]
- 12.Dostal Z., Brzobohaty T., Horak D., Kozubek T., Vodstrcil T. On R-linear convergence of semi-monotonic inexact augmented Lagrangians for bound and equality constrained quadratic programming problems with application. Comput. Math. Appl. 2014;67:515–526. doi: 10.1016/j.camwa.2013.11.009. [DOI] [Google Scholar]
- 13.Belash K.N., Tret’yakov A.A. Methods for solving degenerate problems. USSR Comput. Math. Math. Phys. 1988;28:90–94. doi: 10.1016/0041-5553(88)90116-4. [DOI] [Google Scholar]
- 14.Bertsekas D.P. Nonlinear Programming. Athena Scientific; Belmont, MA, USA: 1999. [Google Scholar]
- 15.Alekseev V.M., Tikhomirov V.M., Fomin S.V. Optimal Control. Consultants Bureau; New York, NY, USA: London, UK: 1987. [Google Scholar]
- 16.Szczepanik E., Tret’yakov A.A. p-Regularity Theory and Methods of Solving Nonlinear Optimization Problems. Uniwersytet Przyrodniczo-Humanistyczny w Siedlcach; Siedlce, Poland: 2020. (In Polish) [Google Scholar]
- 17.Nocedal J., Wright S.J. Numerical Optimization. Springer; New York, NY, USA: 1999. [Google Scholar]
- 18.Facchinei F., Fisher A., Kanzow C. On the accurate identification of active constraints. SIAM J. Optim. 1998;9:14–32. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
No new data were created or analyzed in this study. Data sharing is not applicable to this article.
