Abstract
Mixed-Integer Linear Programming (MILP) plays an important role across a range of scientific disciplines and within areas of strategic importance to society. The MILP problems, however, suffer from combinatorial complexity. Because of integer decision variables, as the problem size increases, the number of possible solutions increases super-linearly thereby leading to a drastic increase in the computational effort. To efficiently solve MILP problems, a “price-based” decomposition and coordination approach is developed to exploit 1. the super-linear reduction of complexity upon the decomposition and 2. the geometric convergence potential inherent to Polyak’s stepsizing formula for the fastest coordination possible to obtain near-optimal solutions in a computationally efficient manner. Unlike all previous methods to set stepsizes heuristically by adjusting hyperparameters, the key novel way to obtain stepsizes is purely decision-based: a novel “auxiliary” constraint satisfaction problem is solved, from which the appropriate stepsizes are inferred. Testing results for large-scale Generalized Assignment Problems demonstrate that for the majority of instances, certifiably optimal solutions are obtained. For stochastic job-shop scheduling as well as for pharmaceutical scheduling, computational results demonstrate the two orders of magnitude speedup as compared to Branch-and-Cut. The new method has a major impact on the efficient resolution of complex Mixed-Integer Programming problems arising within a variety of scientific fields.
Subject terms: Health care, Engineering, Mathematics and computing
Introduction
Mixed-Integer Linear Programming (MILP) plays an important role across a range of scientific disciplines such as mathematics, operations research, engineering, and computer science as well as within a range of areas of strategic importance to society such as biology1,2, healthcare3,4, humanitarian applications5–8, manufacturing9–12, pharmacy13–16, power and energy systems17–19, transportation and logistics20,21 and many others.
The associated systems are created by interconnecting I smaller subsystems, each having its own objective and a set of constraints. The subsystem interconnection is modeled through the use of system-wide coupling constraints. Accordingly, the MILP problems are frequently formulated in terms of cost components associated with each subsystem with the corresponding objective functions being additive as such:
| 1 |
Furthermore, coupling constraints are additive in terms of I subsystems:
| 2 |
The primal problem (1), (2) is assumed to be feasible and the feasible region with is assumed to be bounded and finite. The MILP problems modeling the above systems are referred to as separable. Because of the discrete decisions, however, MILP problems are known to be NP-hard and are prone to the curse of combinatorial complexity. As the size of a problem increases, the associated number of combinations of possible solutions (hence the term “combinatorial”) increases super-linearly (e.g., exponentially) thereby making problems of practical sizes difficult to solve to optimality; even near-optimal solutions are frequently difficult to obtain.
A beacon of hope to resolve combinatorial difficulties lies through the exploitation of separability through the dual “price-based” decomposition and coordination Lagrangian Relaxation technique. After the relaxation of coupling constraints (2), the coordination of subproblems amounts to the maximization of a concave non-smooth dual function:
| 3 |
where
| 4 |
Here is the Lagrangian function. The Lagrangian multipliers (“dual” variables) are the decision variables with respect to the dual problem (3), and it is assumed that the set of optimal solutions is not empty. The minimization within (4) with respect to is referred to as the “relaxed problem.”
While the sizes of the primal and the relaxed problems are the same in terms of the number of discrete variables, the main advantage of Lagrangian Relaxation is the exploitation of the reduction of the combinatorial complexity upon decomposition into subproblems. Accordingly, the number of discrete decision variables within the primal problem is , so the worst-case complexity of solving the primal problems is . By the same token, the worst-case complexity required to solve the following subproblem
| 5 |
is . The decomposition “reverses” the combinatorial complexity thereby exponentially reducing the effort. The decomposition, therefore, offers a viable potential to improve the operations of existing systems as well as to scale up the size of the systems to support their efficient operations.
While decomposition efficiently reduces the combinatorial complexity, the coordination aspect of the method to efficiently obtain the optimal “prices” (Lagrangian multipliers) has been the subject of an intense research debate for decades because of the fundamental difficulties of non-smooth optimization. Namely, because of the presence of integer variables x, the dual function (3) is non-smooth comprised of flat convex polygonal facets (each corresponding to a particular solution to the relaxed problem within (4)) intersecting at linear ridges along which the dual function is non-differentiable; in particular, is not differentiable at thereby ruling out the possibility of using necessary and sufficient conditions for the extremum. As a result of the non-differentiability of , subgradient multiplier-updating directions, however, are non-ascending directions thereby leading to a decrease of dual values; subgradient directions may also change drastically thereby resulting in zigzagging of Lagrangian multipliers (see Fig. 1 for illustrations) and slow convergence as a result.
Figure 1.

An example of a dual function demonstrating difficulties faced by subgradient methods. Solid lines denote the level curves, dash-dotted lines denote the ridges of the dual function whereby the usual gradients are not defined (possible subgradient directions at points (A) and (B) are denoted by solid arrows), and the direction from point (B) toward optimal multipliers is denoted by a dashed line.
Traditional methods to maximize rely upon iterative updates of Lagrangian multipliers by taking a series of steps along subgradient directions as:
| 6 |
where is a an optimal solution to the relaxed problem (4) with multipliers equal to Within the Lagrangian Relaxation framework, subgradients are defined as levels of constraint violations Inequality constraints , if present, can be handled by converting into equality constraints by introducing non-negative real-valued slack variables z such that The multipliers are subsequently projected onto the positive orthant delineated by restrictions
Because of the lack of differentiability of , notably, at the optimum the stepsize selection plays an important role to guarantee convergence to the optimum as well as for the success of the overall Lagrangian Relaxation methodology for solving MILP problems.
One of the earlier papers on the optimization of non-smooth convex functions, with being its member, though irrespective of Lagrangian Relaxation, is Polyak’s seminal work22. Intending to achieve the geometric (also referred to as “linear”) rate of convergence so that is monotonically decreasing, Polyak proposed the stepsizing formula, which in terms of the problem under consideration takes the following form:
| 7 |
Within (7) and thereafter in the paper the standard Euclidean norm is used.
Subgradient directions, however, 1. are generally difficult to obtain computationally when the number of subproblems (5) to be solved is large, and 2. change drastically thereby resulting in zigzagging of Lagrangian multipliers and slow convergence. Moreover, 3. stepsizes (7) cannot be set due to the lack of the knowledge about the optimal dual value .
To overcome the first two of the difficulties above, the Surrogate Subgradient method was developed by23 whereby the exact optimality of the relaxed problem (or even subproblems) is not required. As long as the following “surrogate optimality condition” is satisfied:
| 8 |
the multipliers can be updated by using the following version of the Polyak’s formula
| 9 |
and convergence to is guaranteed. Here “tilde” is used to distinguish optimal solutions to the relaxed problem from the solutions that satisfy the “surrogate optimality condition” (8). Unlike that in Polyak’s formula, parameter is less than 1 to guarantee that so that the stepsizing formula (9) is well-defined, as proved by Zhao et al.23. Once are obtained, multipliers are updated by using the same formula as in (6) with stepsizes from (9) and “surrogate subgradient” multiplier-updating directions used in place of subgradient directions . Besides reducing the computational effort owing to (8), the concomitant reduction of multiplier zigzagging has also been observed. The main difficulty is the lack of knowledge about . As a result, the geometric/linear convergence of the method (or any convergence at all) is highly questionable in practice. Nevertheless, the underlying geometric convergence principle behind the formula (8) is promising and will be exploited in “Results” section.
One of the first attempts to overcome the difficulty associated with the unavailability of the optimal [dual] value is the Subgradient-Level method developed by Goffin and Kiwiel 24 by adaptively adjusting a “level” estimate based on the detection of “sufficient descent” of the [dual] function and “oscillation” of [dual] solutions. In a nutshell, a “level” estimate is set as with being the best dual value (“record objective value”) obtained up to an iteration k, and is an adjustable parameter with j denoting the update of Every time oscillations of multipliers are detected, is reduced by half. In doing so, stepsizes appropriately decrease, increases (for maximization of non-smooth functions such as (3)) and the process continues until and
To improve convergence, rather than updating all the multipliers “at once,” within the Incremental Subgradient methods25, multipliers are updated “incrementally.” Convergence results of the Subgradient-Level method24 have been extended for the Incremental Subgradient methods.
Within the Surrogate Lagrangian Relaxation (SLR) method26, the computational effort is reduced along the lines of the Surrogate Subgradient method23 discussed above, that is, by solving one of a few subproblems at a time. To guarantee convergence, within SLR, distances between multipliers at consecutive iterations are required to decrease through a specially-constructed contraction mapping until convergence. As demonstrated by Bragin et al.26, the SLR method converges faster as compared to the above-mentioned Subgradient-Level method24 and the Incremental Subgradient methods25,27 for non-smooth optimization. Unlike the Subgradient-Level and Incremental Subgradient methods25,27, the SLR method does not require obtaining dual values to set stepsizes, which further reduces the effort. Aiming to simultaneously guarantee convergence while ensuring fast reduction of constraint violations and preserving the linearity, the Surrogate Absolute-Value Lagrangian Relaxation (SAVLR) method28 was developed to penalize constraint violations by using “absolute-value” penalty terms. The above methods are reviewed in more detail in Supplementry Information Section.
Because of the presence of the integer variables, there is the so-called the duality gap, which means that even at convergence, is generally less than the optimal cost of the original problem (1), (2). To obtain a feasible solution to (1), (2), the subproblem solutions when put together may not satisfy all the relaxed constraints. Therefore, to solve corresponding MILP problems, heuristics are inevitable and are used to perturb subproblem solutions. The important remark here is that the closer the multipliers are to the optimum, generally, the closer the subproblem solutions are to the global optimum of the original problem, and the easier it is to obtain feasible solutions through heuristics. Therefore, having fast convergence in the dual space to maximize the dual function (3) is of paramount importance for the overall success of the method. Specific heuristics will be discussed at the end of the “Results” section.
Results
Surrogate “Level-Based” Lagrangian Relaxation
In this subsection, a novel Surrogate “Level-Based” Lagrangian Relaxation (SLBLR) method is developed to determine “level” estimates of within the Polyak’s stepsizing formula (9) for fast convergence of multipliers when optimizing the dual function (3). Since the knowledge of is generally unavailable, over-estimates of the optimal dual value, if used in place of within the formula (9), may lead to the oscillation of multipliers and to the divergence. Rather than using heuristic “oscillation detection” of multipliers used to adjust “level” values24, the key of SLBLR is the decision-based “divergence detection” of multipliers based on a novel auxiliary “multiplier-divergence-detection” constraint satisfaction problem.
“Multiplier-Divergence-Detection” problem to obtain the estimate of
The premise behind the multiplier-divergence detection is the rendition of the result due Zhao et al.23:
Theorem 1
Under the stepsizing formula
| 10 |
such that satisfy
| 11 |
the multipliers move closer to optimal multipliers iteration by iteration:
| 12 |
The following Corollary and Theorem 2 are the main key results of this paper.
Corollary 1
If
| 13 |
then
| 14 |
Theorem 2
If the following auxiliary “multiplier-divergence-detection” feasibility problem (with being a continuous decision variable: )
| 15 |
admits no feasible solution with respect to for some and , then such that
| 16 |
Proof
Assume the contrary: the following holds:
| 17 |
By Theorem 1, multipliers approach therefore, the “multiplier-divergence-detection” problem admits at least one feasible solution - Contradiction.
From (16) it follows that such that and the following holds:
| 18 |
The equation (18) can equivalently be rewritten as:
| 19 |
Therefore,
| 20 |
A brief yet important discussion is in order here. The overestimate of the dual value is the sought-for “level” value after the update (the time the problem (15) is infeasible). Unlike previous methods, which require heuristic hyperparameter adjustments to set level values, within SLBLR, level values are obtained by using the decision-based principle per (15) precisely when divergence is detected without any guesswork. In a sense, SLBLR is hyperparameter-adjustment-free. Specifically, neither “multiplier-divergence-detection” problem (15), nor the computations within (18)–(20) requires hyperparameter adjustment. Following Nedić and Bertsekas27, the parameter will be chosen as a fixed value , which is the inverse of the number of subproblems and will not require further adjustments.
Note that (15) simplifies to an LP constraint satisfaction problem. For example, after squaring both sides of the first inequality within (15), after using the binomial expansion, and canceling from both sides, the inequality simplifies to which is linear in terms of
To speed up convergence, a hyperparameter is introduced to reduce stepsizes as follows:
| 21 |
Subsequently after iteration , the problem (15) is sequentially solved again by adding one inequality per multiplier-updating iteration until iteration is reached for some so that (15) is infeasible. Then, stepsize is updated by using per (21) and is used to update multipliers until the next time it is updated to when the “multiplier-divergence-detection” problem is infeasible again, and the process repeats. Per (21), SLBLR requires hyperparameter , yet, it is set before the algorithm is run and subsequently is not adjusted (see “Numerical testing” section for empirical demonstration of the robustness of the method with respect to the choice of hyperparameter ).
To summarize the advantage of SLBLR, hyperparameter adjustment is not needed. The guesswork of when to adjust the level-value, and by how much is obviated — after (15) is infeasible, the level value is formulaically recalculated.
On improvement of convergence
To speed up the acceleration of the multiplier-divergence detection through the “multiplier-divergence-detection” problem, (15) is modified, albeit heuristically, in the following way:
| 22 |
Unlike the problem (15), the problem (22) no longer simplifies to an LP problem. Nevertheless, the system of inequalities delineate the convex region and can still be handled by commercial software.
Discussion of (22)
Equation (22) is developed based on the following principles: 1. Rather than detecting divergence per (15), convergence with a rate slower than is detected. This will lead to a faster adjustment of the level values. While the level value may no longer be guaranteed to be the upper bound to , the merit of the above scheme will be empirically justified in the “Numerical testing” section. 2. While the rate of convergence is unknown, in the “worst-case” scenario is upper bounded by 1 with , thereby reducing (22) to (15). The estimation of is thus much easier than the previously used estimations of (as in Subgradient-Level and Incremental Subgradient approaches). 3. As the stepsize approaches zero, approaches the value of 1 regardless of the value of , once again reducing (22) to (15).
There are three things to note here. 1. Steps in lines 15-16 are optional since other criteria can be used such as the number of iterations or the CPU time; 2. The value of is still needed (line 1) to obtain a valid lower bound. To obtain , all subproblems are solved optimally for a given value of multipliers . The frequency of the search for the value is determined based on criteria as stated in point 1 above; 3. The search for feasible solutions is explained below.
Search for feasible solutions
Due to non-convexities caused by discrete variables, the relaxed constraints are generally not satisfied through coordination, even at convergence. Heuristics are thus inevitable, yet, they are the last step of the feasible-solution search procedure. Throughout all examples considered, following28 (as discussed in Supplementary Information), -absolute-value penalties penalizing constraint violations are considered. After the total constraint violation reaches a small threshold value, a few subproblem solutions obtained by the Lagrangian Relaxation method are perturbed, e.g., see heuristics within accompanying CPLEX codes within28 to automatically select which subproblem solutions are to be adjusted to eliminate the constraint violation to obtain a solution feasible with respect to the overall problem.
Numerical Testing
In this subsection, a series of examples are considered to illustrate different aspects of the SLBLR method. In “Demonstration of convergence of multipliers based on a small example with known optimal multipliers” section, a small example with known corresponding optimal Lagrangian multipliers is considered to test the new method as well as to compare how fast Lagrangian multipliers approach their optimal values as compared to Surrogate Lagrangian Relaxation26 and to Incremental Subgradient25 methods. In “Generalized Assignment Problems” section, large-scale instances of generalized assignment problems (GAPs) of types D and E with 20, 40, and 80 machines and 1600 jobs from the OR-library (https://www-or.amp.i.kyoto-u.ac.jp/members/yagiura/gap/) are considered to demonstrate efficiency, scalability, robustness, and competitiveness of the method with respect to the best results available thus far in the literature. In “Stochastic job-shop scheduling with the consideration of scrap and rework” section, a stochastic version of a job-shop scheduling problem instance with 127 jobs and 19 machines based on Hoitomt et al.29 is tested. In “Multi-stage pharmaceutical scheduling” section, two instances of pharmaceutical scheduling with 30 and 60 product orders, 17 processing units, and 6 stages based on Kopanos et al.13 are tested.
For “Demonstration of convergence of multipliers based on a small example with known optimal multipliers” section and “Generalized Assignment Problems” section, SLBLR is implemented within CPLEX 12.10 by using a Dell Precision laptop Intel(R) Xeon(R) E-2286M CPU @ 2.40GHz with 16 cores and installed memory (RAM) of 32.0 GB. For “Stochastic job-shop scheduling with the consideration of scrap and rework” section and “Multi-stage pharmaceutical scheduling” section, SLBLR is implemented within CPLEX 12.10 by using a server Intel(R) Xeon(R) Gold 6248R CPU @ 3.00GHz with 48 cores and installed memory (RAM) of 192.0 GB.
Demonstration of convergence of multipliers based on a small example with known optimal multipliers
To demonstrate convergence of multipliers, consider the following example (due Bragin et al.30):
| 23 |
| 24 |
| 25 |
As proved by Bragin et al.30, the optimal dual solutions are and Inequality constraints are converted to equality constraints after introducing slack variables. In Fig. 2, the decrease of the corresponding distances from current multipliers to the optimal multipliers () is shown, and the SLBLR method is compared with the Incremental Subgradient method25 and the Surrogate Lagrangian Relaxation method26.
Figure 2.
Results for “Demonstration of convergence of multipliers based on a small example with known optimal multipliers” section: Comparison of SLBLR to 1. Incremental Subgradient method (left) and 2. Surrogate Lagrangian Relaxation method (right).
Within the SLBLR method, the equation (15) is used to detect divergence, and is used to set stepsizes within (21). In essence, only one hyperparameter was required, which has a quite simple explanation - “when the stepsize is ‘too large,’ cut the stepsize in half.” As demonstrated in Fig. 2, the SLBLR method converges fast with decreasing roughly along a straight line on a log-scale graph suggesting that the rate of convergence is likely linear as expected.
As for the Incremental Subgradient method, two hyperparameters are required: R and (corresponding values used are shown in parentheses in the legend of Fig. 2 (left)). A trial-and-error analysis indicated that “acceptable” values are and Increasing or decreasing R to 0.5 and 0.125, respectively, do not lead to improvements. Likewise, increasing or decreasing to 48 and 12, respectively, do not lead to improvements as well. “Plateau” regions in the figure are caused by the fact that as stepsizes get smaller, a larger number of iterations is required for multipliers to travel the predetermined distance R; during these iterations, stepsizes are not updated and multipliers may oscillate around a neighborhood of the optimum without getting closer. While the above difficulty can be alleviated and convergence can be improved by hyperparameters , , and as reviewed in Supplementary Information, however, an even larger number of hyperparameters would be required.
As for the Surrogate Lagrangian Relaxation method, several pairs of hyperparameters (M and r) have been used as well (corresponding values used are shown in parentheses in the legend of Fig. 2 (right)), yet, the performance of Surrogate Lagrangian Relaxaton does not exceed the performance of the SLBLR method.
Herein lies the advantage of the novel SLBLR method: the decision-based principle behind computing the “level” values. This is in contrast to the problem-dependent choice of hyperparameters R and within the Subgradient-Level24 and Incremental Subgradient25 methods, and the choice of M and r within Surrogate Lagrangian Relaxation26,28 (see “Introduction” section and Supplementry Information for more detail).
Even after obtaining “appropriate” values of the aforementioned hyperparameters by using a trial-and-error procedure that entails effort, results obtained by Surrogate Lagrangian Relaxation26 and the Incremental Subgradient method25 do not match or beat those obtained by the SLBLR method. The specific reasons are 1. Heuristic adjustments of the “level” values are required24,25 based on multiplier “oscillation detection” or “significant descent” (for minimization of non-smooth functions). However, these rules do not detect whether multipliers “start diverging.” Moreover, oscillation of multipliers is a natural phenomenon when optimizing non-smooth functions as discussed in “Introduction” section since multipliers may zigzag/oscillate across ridges of the function, so the multiplier “oscillation detection” may not necessarily warrant the adjustment of level values. On the other hand, multiplier “oscillation” is detected by checking whether multipliers traveled a (heuristically) predetermined distance R, hence, the divergence of multipliers can go undetected for a significant number of iterations (hence, the “plateau” regions shown in Fig. 2 (left)), depending on the value of R. To the best of the authors’ knowledge, the subgradient- and surrogate-subgradient-based methods using Polyak’s stepsizes with the intention of achieving the geometric/linear convergence rate either require , which is unavailable, or require multipliers to travel infinite distance to guarantee convergence to the optimum 24. 2. While SLR avoids the need to estimate , the geometric/linear convergence is only possible outside of a neighborhood of 26. Precisely for this reason, the convergence of multipliers within SLR with the corresponding stepsizing parameters and (as shown in Fig. 2 (right)) appears to follow closely convergence within SLBLR up until iteration 50, after which the improvement tapers off.
Generalized assignment problems
To demonstrate the computational capability of the new method as well as to determine appropriate values for key hyperparameters and while using standard benchmark instances, large-scale instances of GAPs are considered (formulation is available in subsection 4.2 of Supplementary Information). We consider 20, 40, and 80 machines with 1600 jobs (https://www-or.amp.i.kyoto-u.ac.jp/members/yagiura/gap/).
To determine values for within (21) and within (22) to be used throughout the examples, several values are tested using GAP instance d201600. In Table 1, with fixed values of and , the best result (both in terms of the cost and the CPU time) is obtained with . With the value of the stepsize decreases “too fast” thereby leading to a larger number of iterations and a much-increased CPU time as a result. Likewise, in Table 2 with fixed values of and , it is demonstrated that the best result (both in terms of the cost and the CPU time) is obtained with . Empirical evidence here suggests that the method is stable for other values of The robustness with respect to initial stepsizes () is tested and the results are demonstrated in Table 3. Multipliers are initialized by using LP dual solutions. The method’s performance is appreciably stable for the given range of initial stepsizes used (Table 3). SLBLR is robust with respect to initial multipliers (Table 4). For this purpose, the multipliers are initialized randomly by using the uniform distribution U[90, 110]. For the testing, the initial stepsize was used. As evidenced from Table 4, the method’s performance is stable, exhibiting only a slight degradation of solution accuracy and an increase of the CPU time as compared to the case with multipliers initialized by using LP dual solutions.
Table 1.
Robustness results for instance d201600 with respect to .
| Feasible cost | Gap (%) | “Auxiliary” time (sec) | Total time (sec) | |
|---|---|---|---|---|
| 1/1.25 | 97827 | 0.0059 | 4.59 | 2904.02 |
| 1/1.5 | 97825 | 0.0037 | 17.10 | 1195.36 |
| 1/2 | 97825 | 0.0048 | 88.59 | 2612.48 |
| 1/4 | 97827 | 0.0059 | 89.01 | 10235.50 |
The best feasible cost values obtained are in bold.
Table 2.
Robustness results for instance d201600 with respect to .
| Feasible cost | Gap (%) | “Auxiliary” time (sec) | Total time (sec) | |
|---|---|---|---|---|
| 0.03125 | 97826 | 0.0048 | 93.79 | 2716.68 |
| 0.125 | 97825 | 0.0037 | 33.62 | 1820.96 |
| 0.5 | 97826 | 0.0048 | 9.61 | 2444.46 |
| 2 | 97825 | 0.0037 | 17.10 | 1195.36 |
The best feasible cost values obtained are in bold.
Table 3.
Robustness results for instance d201600 with respect to initial stepsizes .
| Initial stepsize () | Feasible cost | Gap (%) | “Auxiliary” time (sec) | Total time (sec) |
|---|---|---|---|---|
| 0.0025 | 97825 | 0.0037 | 123.71 | 2427.71 |
| 0.005 | 97825 | 0.0037 | 6.84 | 1226.17 |
| 0.01 | 97826 | 0.0048 | 6.96 | 2143.58 |
| 0.02 | 97825 | 0.0037 | 17.10 | 1195.36 |
| 0.04 | 97826 | 0.0048 | 19.21 | 1941.55 |
The best feasible cost values obtained are in bold.
Table 4.
Robustness results for instance d201600 with respect to initial multipliers .
| Case number | Feasible cost | Total subproblem solving time (sec) |
Feasible solution search time (sec) |
“Auxiliary” time (sec) |
Total time (sec) |
|---|---|---|---|---|---|
| 1 | 97825 | 1098.74 | 375.96 | 22.13 | 1496.84 |
| 2 | 97826 | 1009.42 | 777.16 | 173.48 | 1960.07 |
| 3 | 97826 | 2223.99 | 221.70 | 4.54 | 2450.24 |
| 4 | 97826 | 2333.55 | 402.41 | 4.08 | 2740.04 |
| 5 | 97826 | 1002.77 | 119.91 | 160.73 | 1283.42 |
The best feasible cost values obtained are in bold.
To test the robustness as well as scalability of the method across several large-scale GAP instances, six instances d201600, d401600, d801600, e201600, e401600, and e801600 are considered. SLBLR is compared with Depth-First Lagrangian Branch-and-Bound method (DFLBnB)31, Column Generation32, and Very Large Scale Neighborhood Search (VLSNS)33, which to the best of the authors’ knowledge are the best methods for at least one of the above instances. For completeness, a comparison against Surrogate Absolute-Value Lagrangian Relaxation (SAVLR)28, which is an improved version of Surrogate Lagrangian Relaxation (SLR)26, is also performed. The latter SLR method26 has been previously demonstrated to be advantageous against other non-smooth optimization methods as explained in "Introduction" section. Table 5 presents feasible costs and times (in seconds) for each method. The advantage of SLBLR is the ability to obtain optimal results across a wider range of GAP instances as compared to other methods. Even though the comparison in terms of the CPU time is not entirely fair, feasible-cost-wise, SLBLR decisively beats previous methods. For the d201600 instance, the results obtained by SLBLR and the Column Generation method32 are comparable. For instance d401600, SLBLR obtains a better feasible solution and for instance d801600, the advantage over the existing methods is even more pronounced.
Table 5.
Comparison against the best results currently available.
| Instance | New method (SLBLR) |
Posta31 (DFLBnB) |
Sadykov32 (Column generation) |
Haddadi33 (VLSNS) |
Bragin28 (SAVLR) |
|---|---|---|---|---|---|
| d201600 | 97825 (1195) | 97825 (1026) | 97836 (5364) | 97828 (1371) | |
| d401600 | 97105 (836) | 97106 (919) | 97125 (5364) | 97111 (1183) | |
| d801600 | 97034 (3670) | 97037 (10860) | 97075 (5364) | 97039 (1350) | |
| e201600 | 180645 (85) | 180645 (40) | − | 180645 (749) | − |
| e401600 | 178293 (2478) | 178293 (243) | − | 178293 (749) | − |
| e801600 | 176820 (1762) | 176820 (75) | − | 176821 (749) | − |
The optimality is certified by the LP optimal values, which are 97105 and 97034 for instances d401600 and d801600, respectively.
The optimality is certified through the lower bound results of, i.e., Posta et al.31.
Not solved to optimality within 24 hours and not reported within the original paper by Posta et al.31.
− These instances were not considered within the papers by Sadykov et al.32 and Bragin et al.28.
The best feasible cost values obtained are in bold.
To the best of the authors’ knowledge, no other reported method obtained optimal results for instances d401600 and d801600. SLBLR outperforms SAVLR28 as well, thereby demonstrating that the fast convergence offered by the novel “level-based” stepsizing, with other things being equal, translates into better results as compared to those obtained by SAVLR, which employs the “contraction mapping” stepsizing28. Lastly, the methods developed in31–33 specifically target GAPs, whereas the SLBLR method developed in this paper has broader applicability.
Stochastic job-shop scheduling with the consideration of scrap and rework
To demonstrate the computational capability of the method to solve large-scale stochastic MILP problems, a job-shop scheduling problem is considered. Within a job shop, each job requires a specific sequence of operations and the processing time for each operation. Operations are performed by a set of eligible machines. To avoid late shipments, expected tardiness is minimized. Limited machine capacity brings a layer of difficulty since multiple “individual-job” subproblems are considered together competing for limited resources (machines). Another difficulty arises because of uncertainties, including processing times34–39 and scrap40–42. Re-manufacturing of one part may affect and disrupt the overall schedule within the entire job shop, thereby leading to unexpectedly high delays in production.
In this paper, we modified data from the paper by Hoitomt et al.29 by modifying several jobs by increasing the number of operations (e.g., from 1 to 6) and decreasing the capacities of a few machines; the data are in Tables S1 and S2. The stochastic version of the problem with the consideration of scrap and rework is available within the manuscript by Bragin et al.42. With these changes, the running time of CPLEX spans multiple days as demonstrated in Fig. 3. In contrast, within the new method, a solution of the same quality as that obtained by CPLEX, is obtained within roughly 1 hour of CPU time. The new method is operationalized by relaxing machine capacity constraints42 and coordinating resulting job subproblems; at convergence, the beginning times of several jobs are adjusted by a few time periods to remove remaining machine capacity constraint violations.
Figure 3.
The results for “Stochastic job-shop scheduling with the consideration of scrap and rework” section are illustrated. SLBLR performs more than two orders of magnitude faster than branch-and-cut implemented in CPLEX.
Multi-stage pharmaceutical scheduling
To demonstrate the capability of the method to solve scheduling problems complicated by the presence of sequence-dependent setup times, a multi-stage pharmaceutical scheduling problem proposed by Kopanos et al.13 is considered. Setup times vary based on the sequencing of products on each unit (machine). Scheduling in this context is combinatorial in the number of product orders, units, and stages. The new method is operationalized by relaxing constraints that couple individual processing units, namely assignment, and processing/setup time constraints (constraints (39)-(41) from Supplementary Information). The results obtained by SLBLR and Branch-and-Cut are demonstrated in Fig. 4.
Figure 4.
The results for “Multi-stage pharmaceutical scheduling” section with 30 products orders (left) and 60 product orders (right) are illustrated.
With a relatively small number of product orders, 30, an optimal solution with a feasible cost of 54.97 was found by CPLEX within 1057.78 seconds. The optimality is verified by running CPLEX until the gap is 0%; it took 171993.27 seconds to verify the optimality. SLBLR takes a slightly longer time to obtain the same solution - 1647.35 seconds (Fig. 4 (left)). In contrast, with 60 product orders, CPLEX no longer obtains good solutions in a computationally efficient manner; a solution with a feasible cost of 55.98 is obtained after 1,000,000 seconds. Within SLBLR, a solution with a feasible cost of 55.69 is obtained within 1978.04 seconds. This constitutes more than two orders of magnitude of improvement over CPLEX as demonstrated in Fig. 4 (right; log scale). When doubling the number of products, CPLEX’s performance is drastically deteriorated, while the performance of SLBLR is scalable.
Discussion
This paper develops a novel MILP solution methodology based on the Lagrangian Relaxation method. Salient features of the novel SLBLR method, inherited from the previous versions of Lagrangian Relaxation, are: 1. reduction of the computational effort required to obtain Lagrangian-multiplier-updating directions and 2. alleviation of zigzagging of multipliers. The key novel feature of the method, which the authors believe gives SLBLR the decisive advantage, is the innovative exploitation of the underlying geometric-convergence potential inherent to Polyak’s stepsizing formula without the heuristic adjustment of hyperparameters for the estimate of - the associated “level” values are determined purely through the simple auxiliary “multiplier-divergence-detection” constraint satisfaction problem. Through testing, it is discovered that SLBLR is robust with respect to the choice of initial stepsizes and multipliers, computationally efficient, competitive, and general. Several problems from diverse disciplines are tested and the superiority of SLBLR is demonstrated. While “separable” MILP problems are considered, no particular problem characteristics such as linearity or separability have been used to obtain “level” values, and thus SLBLR has the potential to solve a broad class of MIP problems.
Supplementary Information
Acknowledgements
The work by M.A.B. was supported in part by the US NSF under award ECCS-1810108.
Author contributions
M.A.B. conceptualized the project, developed the novel methodology, and conducted experiments. E.L.T. developed the fourth case study and supported testing and interpretation of results. M.A.B. drafted the initial manuscript with revisions from E.L.T. Both authors provided feedback on the final version. All authors read and approved the final manuscript.
Data availability
Data supporting the results of “Generalized Assignment Problems” section are located at https://www-or.amp.i.kyoto-u.ac.jp/members/yagiura/gap/; for “Stochastic job-shop scheduling with the consideration of scrap and rework” section, data are located in Tables S1 and S2 as well as in Supplementry Information; for “Multi-stage pharmaceutical scheduling” section, data are taken from the paper by Kopanos et al.13.
Competing interests
The authors declare no competing interests.
Footnotes
The original online version of this Article was revised: This Article contained errors in the algorithm in the Results section, under the subheading ‘Algorithm: Pseudocode’. Full information regarding the corrections made can be found in the correction notice for this article.
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Change history
3/1/2023
A Correction to this paper has been published: 10.1038/s41598-023-30358-9
Supplementary Information
The online version contains supplementary material available at 10.1038/s41598-022-26264-1.
References
- 1.Huang PS, Boyken SE, Baker D. The coming of age of de novo protein design. Nature. 2016;537:320–327. doi: 10.1038/nature19946. [DOI] [PubMed] [Google Scholar]
- 2.Yang L, Chen R, Goodison S, Sun Y. An efficient and effective method to identify significantly perturbed subnetworks in cancer. Nat. Comput. Sci. 2021;1:79–88. doi: 10.1038/s43588-020-00009-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Khlif Hachicha H, Zeghal Mansour F. Two-MILP models for scheduling elective surgeries within a private healthcare facility. Health Care Manag. Sci. 2018;21(3):376–392. doi: 10.1007/s10729-016-9390-2. [DOI] [PubMed] [Google Scholar]
- 4.Kayvanfar V, Akbari Jokar MR, Rafiee M, Sheikh S, Iranzad R. A new model for operating room scheduling with elective patient strategy. INFOR Inf. Syst. Oper. Res. 2021;59(2):309–332. doi: 10.1080/03155986.2021.1881359. [DOI] [Google Scholar]
- 5.Smalley HK, Keskinocak P, Swann J, Hinman A. Optimized oral cholera vaccine distribution strategies to minimize disease incidence: A mixed integer programming model and analysis of a Bangladesh scenario. Vaccine. 2015;33(46):6218–6223. doi: 10.1016/j.vaccine.2015.09.088. [DOI] [PubMed] [Google Scholar]
- 6.Hamdan B, Diabat A. Robust design of blood supply chains under risk of disruptions using Lagrangian relaxation. Transp. Res. Part E Logist. Transp. Rev. 2020;134:101764. doi: 10.1016/j.tre.2019.08.005. [DOI] [Google Scholar]
- 7.Ahani N, Andersson T, Martinello A, Teytelboym A, Trapp AC. Placement optimization in refugee resettlement. Oper. Res. 2021;69(5):1468–1486. doi: 10.1287/opre.2020.2093. [DOI] [Google Scholar]
- 8.Kamyabniya A, Noormohammadzadeh Z, Sauré A, Patrick J. A robust integrated logistics model for age-based multi-group platelets in disaster relief operations. Transp. Res. Part E Logist. Transp. Rev. 2021;152:102371. doi: 10.1016/j.tre.2021.102371. [DOI] [Google Scholar]
- 9.Liu Q, Li X, Gao L. Mathematical modeling and a hybrid evolutionary algorithm for process planning. J. Intell. Manuf. 2021;32(2):781–797. doi: 10.1007/s10845-020-01703-w. [DOI] [Google Scholar]
- 10.Hong I-H, Chou C-C, Lee P-K. Admission control in queue-time loop production-mixed integer programming with Lagrangian relaxation (MIPLAR) Comput. Ind. Eng. 2019;129:417–425. doi: 10.1016/j.cie.2019.02.002. [DOI] [Google Scholar]
- 11.Balogh A, Garraffa M, O’Sullivan B, Salassa F. MILP-based local search procedures for minimizing total tardiness in the No-idle Permutation Flowshop problem. Comput. Oper. Res. 2022 doi: 10.1016/j.cor.2022.105862. [DOI] [Google Scholar]
- 12.Öztop H, Tasgetiren MF, Kandiller L, Pan QK. Metaheuristics with restart and learning mechanisms for the no-idle flowshop scheduling problem with makespan criterion. Comput. Oper. Res. 2022;138:105616. doi: 10.1016/j.cor.2021.105616. [DOI] [Google Scholar]
- 13.Kopanos GM, Méndez CA, Puigjaner L. MIP-based decomposition strategies for large-scale scheduling problems in multiproduct multistage batch plants: A benchmark scheduling problem of the pharmaceutical industry. Eur. J. Oper. Res. 2010;207(2):644–655. doi: 10.1016/j.ejor.2010.06.002. [DOI] [Google Scholar]
- 14.Stefansson H, Sigmarsdottir S, Jensson P, Shah N. Discrete and continuous time representations and mathematical models for large production scheduling problems: A case study from the pharmaceutical industry. Eur. J. Oper. Res. 2011;215(2):383–392. doi: 10.1016/j.ejor.2011.06.021. [DOI] [Google Scholar]
- 15.Zhu SX, Ursavas E. Design and analysis of a satellite network with direct delivery in the pharmaceutical industry. Transp. Res. Part E Logist. Transp. Rev. 2018;118:190–207. doi: 10.1016/j.tre.2018.06.005. [DOI] [Google Scholar]
- 16.Ge C, Yuan Z. Production scheduling for the reconfigurable modular pharmaceutical manufacturing processes. Comput. Chem. Eng. 2021;151:107346. doi: 10.1016/j.compchemeng.2021.107346. [DOI] [Google Scholar]
- 17.Schill W-P, Pahle M, Gambardella C. Start-up costs of thermal power plants in markets with increasing shares of variable renewable generation. Nat. Energy. 2017;2(6):1–6. doi: 10.1038/nenergy.2017.50. [DOI] [Google Scholar]
- 18.Chen Y, Pan F, Holzer J, Rothberg E, Ma Y, Veeramany A. A high performance computing based market economics driven neighborhood search and polishing algorithm for security constrained unit commitment. IEEE Trans. Power Syst. 2020;36(1):292–302. doi: 10.1109/TPWRS.2020.3005407. [DOI] [Google Scholar]
- 19.Li X, Zhai Q, Guan X. Robust transmission constrained unit commitment: A column merging method. IET Gener. Transm. Distrib. 2020;14(15):2968–2975. doi: 10.1049/iet-gtd.2018.6314. [DOI] [Google Scholar]
- 20.Archetti C, Peirano L, Speranza MG. Optimization in multimodal freight transportation problems: A survey. Eur. J. Oper. Res. 2022;299(1):1–20. doi: 10.1016/j.ejor.2021.07.031. [DOI] [Google Scholar]
- 21.Reddy KN, Kumar A, Choudhary A, Cheng TCE. Multi-period green reverse logistics network design: An improved Benders-decomposition-based heuristic approach. Eur. J. Oper. Res. 2022 doi: 10.1016/j.ejor.2022.03.014. [DOI] [Google Scholar]
- 22.Polyak BT. Minimization of unsmooth functionals. USSR Comput. Math. Math. Phys. 1969;9(3):14–29. doi: 10.1016/0041-5553(69)90061-5. [DOI] [Google Scholar]
- 23.Zhao X, Luh PB, Wang J. Surrogate gradient algorithm for Lagrangian relaxation. J. Optim. Theory Appl. 1999;100(3):699–712. doi: 10.1023/A:1022646725208. [DOI] [Google Scholar]
- 24.Goffin J-L, Kiwiel KC. Convergence of a simple subgradient level method. Math. Program. 1999;85:207–211. doi: 10.1007/s101070050053. [DOI] [Google Scholar]
- 25.Nedić A, Bertsekas DP. Incremental subgradient methods for nondifferentiable optimization. SIAM J. Optim. 2001;12(1):109–138. doi: 10.1137/S1052623499362111. [DOI] [Google Scholar]
- 26.Bragin MA, Luh PB, Yan JH, Yu N, Stern GA. Convergence of the surrogate Lagrangian relaxation method. J. Optim. Theory Appl. 2015;164(1):173–201. doi: 10.1007/s10957-014-0561-3. [DOI] [Google Scholar]
- 27.Nedić A, Bertsekas D. Convergence rate of incremental subgradient Algorithms. In: Uryasev S, Pardalos PM, editors. Stochastic Optimization: Algorithms and Applications. Boston: Springer; 2001. [Google Scholar]
- 28.Bragin MA, Luh PB, Yan B, Sun X. A scalable solution methodology for mixed-integer linear programming problems arising in automation. IEEE Trans. Autom. Sci. Eng. 2019;16(2):531–541. doi: 10.1109/TASE.2018.2835298. [DOI] [Google Scholar]
- 29.Hoitomt DJ, Luh PB, Pattipati KR. A practical approach to job shop scheduling problems. IEEE Trans. Robot. Autom. 1993;9(1):1–13. doi: 10.1109/70.210791. [DOI] [Google Scholar]
- 30.Bragin MA, Yan B, Luh PB. Distributed and asynchronous coordination of a mixed-integer linear system via surrogate Lagrangian relaxation. IEEE Trans. Autom. Sci. Eng. 2020;18(3):1191–1205. doi: 10.1109/TASE.2020.2998048. [DOI] [Google Scholar]
- 31.Posta M, Ferland JA, Philippe M. An exact method with variable fixing for solving the generalized assignment problem. Comput. Optim. Appl. 2012;52(3):629–644. doi: 10.1007/s10589-011-9432-0. [DOI] [Google Scholar]
- 32.Sadykov, R., Vanderbeck, F., Pessoa, A., & Uchoa, E. Column generation based heuristic for the generalized assignment problem. XLVII Simpósio Brasileiro de Pesquisa Operacional, Porto de Galinhas, Brazil, 3624–3631 (2015).
- 33.Haddadi S. Variable-fixing then subgradient optimization guided very large scale neighborhood search for the generalized assignment problem. 4OR Q. J. Oper. Res. 2019;17(3):261–295. doi: 10.1007/s10288-018-0389-z. [DOI] [Google Scholar]
- 34.Golenko-Ginzburg D, Gonik A. Optimal job-shop scheduling with random operations and cost objectives. Int. J. Prod. Econ. 2002;76(2):147–157. doi: 10.1016/S0925-5273(01)00140-2. [DOI] [Google Scholar]
- 35.Lei D. Simplified multi-objective genetic algorithms for stochastic job shop scheduling. Appl. Soft Comput. 2011;11(8):4991–4996. doi: 10.1016/j.asoc.2011.06.001. [DOI] [Google Scholar]
- 36.Zhang R, Song S, Wu C. A hybrid differential evolution algorithm for job shop scheduling problems with expected total tardiness criterion. Appl. Soft Comput. 2013;13(3):1448–1458. doi: 10.1016/j.asoc.2012.02.024. [DOI] [Google Scholar]
- 37.Shen J, Zhu Y. Chance-constrained model for uncertain job shop scheduling problem. Soft Comput. 2016;20(6):2383–2391. doi: 10.1007/s00500-015-1647-z. [DOI] [Google Scholar]
- 38.Jamili A. Job shop scheduling with consideration of floating breaking times under uncertainty. Eng. Appl. Artif. Intell. 2019;78:28–36. doi: 10.1016/j.engappai.2018.10.007. [DOI] [Google Scholar]
- 39.Horng SC, Lin SS. Apply ordinal optimization to optimize the job-shop scheduling under uncertain processing times. Arab. J. Sci. Eng. 2022;47:9659–9671. doi: 10.1007/s13369-021-06317-9. [DOI] [Google Scholar]
- 40.Wilson JP, Shen Z, Awasthi U, Bollas GM, Gupta S. Multi-objective optimization for cost-efficient and resilient machining under tool wear. J. Adv. Manuf. Process. 2022 doi: 10.1002/amp2.10140. [DOI] [Google Scholar]
- 41.Sun Y, Tu J, Bragin MA, Zhang L. A simulation-based integrated virtual testbed for dynamic optimization in smart manufacturing systems. J. Adv. Manuf. Process. 2022 doi: 10.1002/amp2.10141. [DOI] [Google Scholar]
- 42.Bragin, M. A., Wilhelm, M. E., & Stuber, M. D. Toward agile and robust supply chains: A lesson from stochastic job-shop scheduling. Preprint at arxiv:2206.09326v1 (2022).
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data supporting the results of “Generalized Assignment Problems” section are located at https://www-or.amp.i.kyoto-u.ac.jp/members/yagiura/gap/; for “Stochastic job-shop scheduling with the consideration of scrap and rework” section, data are located in Tables S1 and S2 as well as in Supplementry Information; for “Multi-stage pharmaceutical scheduling” section, data are taken from the paper by Kopanos et al.13.



