Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Aug 26;16(8):e0254239. doi: 10.1371/journal.pone.0254239

An improved Wolf pack algorithm for optimization problems: Design and evaluation

Xuan Chen 1, Feng Cheng 2,*, Cong Liu 3, Long Cheng 4, Yin Mao 5
Editor: Seyedali Mirjalili6
PMCID: PMC8389437  PMID: 34437547

Abstract

Wolf Pack Algorithm (WPA) is a swarm intelligence algorithm that simulates the food searching process of wolves. It is widely used in various engineering optimization problems due to its global convergence and computational robustness. However, the algorithm has some weaknesses such as low convergence speed and easily falling into local optimum. To tackle the problems, we introduce an improved approach called OGL-WPA in this work, based on the employments of Opposition-based learning and Genetic algorithm with Levy’s flight. Specifically, in OGL-WPA, the population of wolves is initialized by opposition-based learning to maintain the diversity of the initial population during global search. Meanwhile, the leader wolf is selected by genetic algorithm to avoid falling into local optimum and the round-up behavior is optimized by Levy’s flight to coordinate the global exploration and local development capabilities. We present the detailed design of our algorithm and compare it with some other nature-inspired metaheuristic algorithms using various classical test functions. The experimental results show that the proposed algorithm has better global and local search capability, especially in the presence of multi-peak and high-dimensional functions.

1 Introduction

With the rapid growth of the complexity of data applications and systems, swarm intelligence has received increasing attentions. Specifically, to solve complex optimization problems in an acceptable time, different algorithms such as Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Ant Colony Optimization (ACO) as well as their variations have been proposed in the last few decades. All the approaches can be broadly decomposed into a population initialization stage followed by an iteration computing process, and they have been widely used on handling numerical and real-world optimization problems in various domains (e.g., engineering and computer systems).

The Wolf Pack Algorithm (WPA) is a typical swarm intelligence algorithm based on the living habits of wolves [1, 2]. Generally, WPA includes three main intelligent behaviors (i.e., wolf’s searching behavior, leader wolf’s calling behavior and fierce wolf’s round-up behavior) and two mechanisms (i.e., the leader wolf’s competition mechanism based on winner is king and the wolf pack’s updating mechanism based on survival of the fittest wolves). Compared to other swarm intelligence, WPA has demonstrated good capability in global convergence and computational robustness [2]. Although the algorithm has been applied in various practical problems, such as for placement and path selection [3], network management [4], UAV trajectory planning [5], data flow control [6], etc., it still has several issues in handling complex problems because of the following issues.

  • Issue 1: The population in WPA lacks effective initialization. Generally, an advanced initialization can ensure the uniform distribution of solutions in searching space to some extent, and thus it can provide better solutions for the iterative process in a swarm intelligence. However, WPA initializes the spatial position of the individual wolf pack in a random way, which reduces the population diversity of the algorithm, and is not conducive to the generation of the optimal solution of the algorithm.

  • Issue 2: The searching wolf in WPA adopts a greedy wandering strategy. However, when the maximum wandering number is set to a large value, the searching wolf will be prone to get into a local optimal solution due to the excessive greed. In this case, it will not be able to continue to expand the exploration space and thus the global search capability of the algorithm of will reduced.

  • Issue 3: The convergence speed of WPA is fast at early stages of computing, but has a significant reduction at the later stages. This is mainly because the effect of fierce wolf’s round-up behavior, which will lead to the local optimization of the algorithm and the poor local search effect and reduce the performance of the algorithm.

  • Issue 4: In WPA, a wolf is randomly selected to be the leader. If its target function value is smaller than that of other wolves, the position of the leader wolf will be replaced. However, updating the leader wolf in such a way will easily result in the whole algorithm falling into a local optimum. Moreover, the lack of a leader wolf’s influence will also reduce the stability of the algorithm.

In fact, to achieve better optimization performance, various works have been proposed to improve the WPA approach, especially for the first three issues described above. For example, the work [7] has introduced different approaches for WPA to improve its population diversity. For the second issue, the study [8] has pointed out that applying chaos strategy in WPA’s searching behavior can effectively avoid falling into local optimum. Moreover, the paper [9] has presented how to use Levy behavior to optimize the round-up behavior for the third problems. In comparison, to the best of our knowledge, the fourth issue has not been studied yet.

To improve the WPA method in a comprehensive way, in this work, we introduce an optimized approach called OGL-WPA, i.e., an improved WPA based on the Opposition-based learning and Genetic algorithm with Levy’s flight. Specifically, we focus on leveraging the existing intelligent techniques (i.e.,opposition-based learning, GA and Levy’s flight) to optimize WPA. We believe that our design and the evaluation conducted in the work as well as the reported results are of value to the community as a basis for understanding the merits of our algorithm. Generally, the contributions of this paper are summarized as follows.

  • To improve the performance of WPA, we introduce an optimized approach called OGL-WPA, with the seamless integration of three intelligent techniques: opposition-based learning, Genetic algorithm and Levy’s flight.

  • We present the details on how to use the three intelligent techniques to optimize WPA as well as the general implementation flow of OGL-WPA.

  • Our evaluation over classical test functions demonstrate that OGL-WPA has better local and global search capability, compared to some popular swarm intelligence algorithms. Specifically, it has significant advantages in solving multi-peak and high-dimensional problems.

The remainder of this paper is organized as follows. The related work of WPA and its principle are introduced in Section 2 and Section 3 respectively. The proposed OGL-WPA is presented in Section 4. The relevant experimental results are reported in Section 5 and the conclusions of this work are summarized in section 6.

2 Related work

In this section, we mainly present the related work of WPA. Since another wolf related swarm intelligence called Grey Wolf Optimizer (GWO) [10] becomes increasingly popular in the current research, we also give the relevant studies of GWO.

2.1 WPA related works

2.1.1 Optimization of WPA

Xiu et al. [9] proposed a tent chaotic map and Levy flight in WPA. Experimental results showed that the optimized algorithm has faster convergence speed and higher precision. Li et al. [8] proposed a Chaos-based WPA, which can effectively improve the local search ability and had good development and balance ability. Simulated results showed that the algorithm has higher effectiveness and robustness.

2.1.2 Hybrid algorithms based on WPA

Chen et al. [7] proposed a hybrid algorithm based on WPA and differential evolution algorithm (DEA). The DEA was introduced into the individual of WPA to increase the individual diversity of the population and improve the local search ability. Simulation experiments showed that the algorithm performance is better than genetic algorithm (GA), DEA, PSO and artificial bee colony (ABC) algorithms. Based on WPA, Chen et al. [11] gave a hybrid improved algorithm WPA-PSO and demonstrated that it can improve the performance effectively.

2.1.3 Applications of WPA

Jiang et al. [12] showed that WPA can improve the efficiency of route planning. Dong et al. [13] proposed a hybrid optimization algorithm based on wolf pack search and local search (WPS-LS) in business travel problems. They found that it had better robustness and planning effect by simulations. Liang et al. [14] proposed a cluster cooperative algorithm based on improved wolf’s behavior, which showed that the model can effectively guarantee the efficiency of solving large-scale complex optimization problems and the operational effectiveness of distributed cluster cooperative attack problems. Moreover, the convergence of the algorithm is proved by Markov asymptotic convergence theory. Zhang et al. [15] proposed a hybrid prediction model of WPA based on fuzzy clustering and least squares vector machine, and applied this model to electric bus, which obtained high prediction accuracy and stability. Han et al. [16] used the WPA for scheduling in the RHFS problem, and the experimental results show that the Wolf pack algorithm can effectively solve the existing scheduling problems in the RHFS problem. Gao et al. [17] proposed a WPA based on quantum coding to solve the 0–1 knapsack problem, it is showed that the algorithm had better global search ability by simulation. Chen et al. [18] proposed the optimized WPA for the route planning of the three-dimensional UAV and the route planning accuracy can be greatly increased by simulation.

2.2 GWO related works

2.2.1 Optimization of GWO

Gupta et al. [19] proposed an improved algorithm RW-GWO based on random walking. Simulated results showed that the performance of the improved GWO improved greatly. Long et al. [20] proposed a GWO based on reverse learning, namely RL-GWO. Simulation results showed that this algorithm is an effective and reliable algorithm for solving function optimization problems.

2.2.2 Hybrid algorithms based on GWO

Arora [21] proposed an algorithm based on the fusion of GWA and Crow algorithm–GWOCSA. Simulated results showed that the algorithm has a good ability to solve complicated problems. Al-Tashi et al. [22] proposed another algorithm which combined binary-based particle swarm and GWA, it called BGWOPSO. The algorithm found the best feature subset through K-nearest neighbor classifier. Simulated results showed that the algorithm was better than GWO, PSO, GA and other algorithms. Singh et al. [23] proposed an improved grey wolf optimization algorithm, which was used to solve the economic power load scheduling problem and achieved good results. Barraza et al. [24] proposed a fusion algorithm based on FWA and GWO. The test of the benchmark function showed that the fusion algorithm had better performance. Gaidhane et al. [25] created GWO-ABC by combining ABC with GWO. Firstly, the algorithm applied the chaotic strategy for population initialization in the WPA. Then, enhance its exploration ability by applying the information sharing strategy of the bees in the ABC. Simulated results showed that the performance of the improved gray wolf algorithm has been improved. Zhi et al. [26] combined the PSO algorithm with the GWO algorithm, which we called PSO-GWO. In the simulated experiments, they used Tent to initialize population, and the best individual idea of PSO is used to update the individual of the wolf group, which can prevent it from falling into local optimum. The presented simulation results showed that PSO-GWO had a better global optimal solution. In addition, Al-Wajih et al. [27] proposed an algorithm called HBGWOHHO, which is based on the combination of GWO algorithm and Harris Hawks Optimization. Moreover, Banaie-Dezfouli et al. [28] proposed a method called R-GWO, which is constructed by a representative based grey wolf optimizer.

2.2.3 Applications of GWO

Maharana et al. [29] applied the GWA and the JAYA algorithm for job scheduling in the workshop. The simulated results showed that the GWA has better performance. Zapotecas-Martnez et al. [30] introduced a multi-objective decomposition GWA, which modeled the multi-objective optimization problem in the form of Pareto’s optimal equilibrium optimal solution set. Simulated results showed that the algorithm has higher quality. In order to achieve lower costs, Kaur et al. [31] used the GWA to optimize the offload scheduling of programs on the mobile device side. Simulated experiments showed that the algorithm has better scheduling effect. Al-Moalmi et al. [32] proposed the use of the GWA for the optimization of virtual machine layout, which can reduce the number and energy of active hosts. Experiments showed that using this algorithm can effectively reduce energy consumption and use CPU and memory resources better. In order to solve large-scale numerical optimization problems, Long et al. [33] proposed an optimized GWO, which we also called ERGWO. The algorithm used a nonlinear adjustment strategy of particle swarms for balancing exploration and development. The simulated results showed that the proposed ERGWO can find high-quality solutions with low computational cost and fast convergence.

From the above related work, we can see that the research on GWO is very active, even more than the WPA. However, as we will demonstrate in this work, an improved WPA can actually perform much better than GWO and also its variants.

3 Basic idea of Wolf pack algorithm

WPA [2] is an intelligent algorithm which simulates the wolf pack predation in the natural universe. Generally, the individuals in the wolf pack are divided into three types: the leader wolf, the detection wolf and the fierce wolf. Different kinds of wolves undertake their own responsibility, meanwhile they work cooperatively to catch the prey.

There is only one leader wolf in a pack which acts as a decision maker, leading the pack to catch food as quickly as possible. But the leader wolf in a pack is not fixed, any individual in the pack can become the leader wolf, the leader wolf is determined by competition, that is, only the wisest and fiercest wolf in the group can become the leader wolf. A wolf pack has multiple detection wolves, mainly responsible for the search for the location of the prey within the scope of activities, and then conveyed the prey information to the leader wolf. The location of a wolf’s prey is obtained mainly by the scent of its prey. If the scent is stronger, the closer the wolf is to the prey. There are several fierce wolves in a wolf pack, the wolf is mainly responsible for rounding up the prey. After the detection wolf to find the trail of its prey and report it to the wolf, the wolf called down to round up the prey. When prey is captured, it is not evenly distributed to each individual in the pack. The prey is given priority to the most powerful wolf, and then assigned to the lesser wolf, and finally not even to the weakest wolf.

The distribution rules ensure that the competent wolves can survive, and the small and weak wolves are gradually eliminated, which is conducive to the overall survival and development of the wolves. Therefore, the whole wolf colony algorithm mainly includes two mechanisms, the leader wolf generation mechanism and the wolf pack update mechanism, and three behaviors, namely, search behavior, call behavior and round-up behavior.

The hunting area of the wolf pack is a European space called N × D, where N means the total number of individuals in the wolf pack and D means the variable dimension of space. The position state of a wolf in a wolf pack is represented by Xi = (xi1,xi2, …, xiD), where xid represented the location of wolf i in the d(d = 1, 2…D) dimension variable space. Note the concentration (target function value) of the scent of prey perceived by each individual in the wolf pack as Yi, then Yi = f(Xi)(i = 1, 2,…N). The goal of the wolf colony algorithm is to find xi in a given range of variable values to maximize the value of f(xi), that is, the maximum value of the target function, which is a function solving optimization problem.

The two mechanisms and three behaviors of the algorithm are given below, and the detailed analysis of the improved algorithm is prepared.

1) Leader wolf’s generating mechanism

The leader wolf changes with the process of rounding up and iteration. During each round up, if the target function value of an individual in a wolf pack is greater than that of the leader wolf, the individual immediately replaces the original position of the leader wolf. During the iteration, the target function value of the optimal wolf after each iteration is compared to the value of the leader wolf in the previous generation, and the position of the leader wolf is updated if better. Throughout the process, we note that the location of the leader wolf is Xlead and its corresponding target function value is Ylead.

2) Wolf pack’s updating mechanism

The food distribution rules of the wolf pack cause part of small and weak wolves to be starved to death, so in the wolf colony algorithm, to remove the target function value of the worst R wolf individual, and then randomly add R wolf individual, where R ∈ [N/(2 × β),N/β] with β for the population update ratio coefficient which is artificially set.

3) Searching behavior

In the process of searching for prey, the wolves do not send out the whole pack, but choose the best other than the leader wolf to search for prey first. Set the number of detection wolf random take a whole number of [Nα+1,Nα], where α is the proportion factor of detection wolf and is artificially set. When a wolf i’s current position of the prey concentration value Yi > Ylead, then Ylead = Yi; When Yi < Ylead, the wolf moves the step length of Stepa in each direction of h, calculates the target function value after each step and then returns to its original position. The location of wolf i in dimension d space after moving in direction p(p = 1, 2, …,h) is:

xidp=xid+sin(2π×p/h)×Stepad. (1)

At this point, Wolf i selects the direction of maximum odor concentration and greater than the current position to go further, update Wolf Status to xi and repeat the search process until the detection Wolf perceived odor concentration value Yi > Ylead or the maximum number of iterations.

4) Calling behavior

When the leader wolf calls on the nearby fierce wolf to set the Stepb step to the leader wolf position, the position of the i wolf in the d dimension variable space at iteration k + 1 is:

xidk+1=xidk+Stepbd×(gdk-xidk)/|gdk-xidk|, (2)

where gdk is the position of the first wolf of the generation K in the d dimension space. When attacking prey, if wolf i smell the prey odor concentration value Yi > Ylead, then Ylead = Yi, the wolf immediately transformed into a leader and launched a call; if Ylead < Yi, wolf i continues to attack until the distance from the leader is less than

dnear=1D×ω×d=1D|maxd-mind|, (3)

where, ω is the distance determinant, and the value range of the D-variable is [mind, maxd].

5) Rounding-up behavior

When the wolf is close to the prey, the wolf will jointly explore the wolf to round up the prey, rounding up the behavior as follow:

xidk+1=xidk+λ×Stepcd×|Gdk-xidk|, (4)

where, λ represents a random number that is uniformly distributed in [−1, 1], Stepcd is the attack step size, and the Gdk represents the position of the prey in d dimension space in iteration K, actually the position of the leader wolf as its location. In addition, the relationship within the three steps in the algorithm are as follows: searching step size Stepa, attacking step size Stepb and attacking step size Stepc).

Stepad=Stepbd/2=2Stepcd=|maxd-mind|/S, (5)

where, S is the step length factor, which indicates the degree of granularity of searching the optimal solution in the solution space.

4 The improved Wolf pack algorithm

Like many other swarm intelligence algorithms, WPA has good global searchinng capability, but it is easy to fall into local optimization and has slow convergence as we have described. To improve the problems, we propose three improvements in our design. Firstly, using opposition-based learning to initialize the population, which can keep the diversity of the population and avoid the algorithm getting into precocity. Secondly, the genetic algorithm is used to improve the selection of the leader wolf. Thirdly, the Levy flight mechanism is used to optimize the round-up behavior.

4.1 Opposition-based learning for population initialization

The initial population construction strategy refers to the distribution of the initial population in the solution space. If the initial solution is close to the optimal solution, the algorithm can converge faster, so the initial population construction will directly affect the performance of the algorithm. The initial solution of WPA is randomly generated and the initialization method is simple and easy to implement, but it will reduce the efficiency of the algorithm because it does not estimate the individual. Fortunately, we can solve this problem by opposition-based learning strategies. Opposition-based learning [34] is an optimization method used in machine learning. In each iteration of the algorithm, all the reverse solutions of the current solution are given, and the solution is selected in the current solution and the reverse solution facilitates the evolution, reducing the blindness of the algorithm. In this paper, the population initialization strategy based on opposition-based learning is adopted, by searching the current solution and the reverse solution simultaneously, and selecting the better solution as the initial solution, thus probability of finding the best initial solution is increased. The steps are as follows:

  • Step 1: Create the initial population randomly. The initial population NP1 = {x1(t), x2(t), …, xN(t)} of the random generation algorithm is calculated according to the formula (6),
    xij=xminj+rand(0,1)(xmaxj-xminj). (6)
    where xij represents the ith individual in dimension j, i is in the range [1, N], j is in the range [1, D], xminj and xmaxj represent the upper and lower bounds of the space, respectively.
  • Step 2: Solve the inverse solution. The inverse population NP1 corresponding to NPop={x˜1(t),x˜2(t),,x˜N(t)} is obtained, and the inverse solution of each individual is given as follows,
    x˜i(t)=xminj+xmaxj-xij. (7)
  • Step 3: Select the optimal individual. First select the individual xbest with the best value of the target function from NP1NPop, and then calculate the average value xmean=(x1+x2++xN+x˜1+x˜2++x˜N)/2N of the solution in NP1NPop, finally compare the corresponding target function values of xbest and xmean respectively. The individual with the largest value of the target function is the best individual of the population, and record it as xopbest, which is obtained by the following formula,
    xopbest={xbest,iff(xbest)>f(xmean)xmean,otherwise. (8)

    The opposition-based learning algorithm can search in a larger search space and guide individuals to evolve toward the optimal value, thus improve the overall convergence speed of the algorithm.

4.2 Genetic algorithm for leader Wolf selection

The goal of WPA is to determine the location of the leader wolf, which is the parameter value of the optimal solution of the algorithm. In WPA, the location of the leader wolf is only determined by the size of the scent concentration of prey, which affects the search ability of the global optimal solution. In this paper, the selection, crossover and mutation of genetic algorithm (GA) are used to select the leader wolf, which makes the new leader wolf have stronger robustness and global optimization, and can accelerate the convergence speed of the algorithm. Set the leader wolf in iteration g to xleadg, the steps are as follows.

4.2.1 Select actions

According to the GA, the fitness function γ is defined as the reciprocal of the objective function. The target function value γ corresponding to the individual xig in the wolves in iteration g is defined as follows:

γ(xig)=1f(xig),i=1,2,,N, (9)

where formula (9)indicates the value of the fitness function corresponding to the individual in the wolf pack. The probability of a wolf individual being selected is set according to the roulette method, as shown in the formula (10):

pig=γ(xig)i=1Nγ(xig),i=1,2,,N, (10)

4.2.2 Cross action

Two individuals, xr1g and xr2g, are randomly selected by probability pr1g and pr2g for cross operation, as shown in formula (11).

{x¯r1g=rxr1g+(1-r)xr2gx¯r2g=(1-r)xr1g+r)xr2g. (11)

The value of cross probability r is 0.95, and x¯r1g and x¯r2g are the values after the two individuals cross.

4.2.3 Mutation operation

The leader wolf individual and the two random individuals are performed a mutation operation after cross operation. The variation operation formula (12) is as follows:

Vig+1={xleadg+λ×|xr1g-xr2g|,rand()<pmxleadg,otherwise, (12)

where Vig+1 represents the individual after xleadg mutation, the mutation probability pm is set to 0.01, rand() is a random number between (0, 1), λ is a random factor between (0, 1), when rand() < pm, Vig+1 is the new individual after xleadg mutation. Otherwise it is still xleadg.

The target function values of individual Vig+1 are compared with xleadg by selecting, crossing, and mutating. If f(Vig+1)>f(xleadg), the new leader wolf Vig+1 is used. Otherwise xleadg continues to be used.

4.3 Levy’s flight for rounding-up behavior

A large number of studies have shown that in the process of hunting in large environment, there are basically Levy’s flight search characteristics, that is, long short distance search trajectory and occasional long distance search trajectory interlace. Edwards et al. [35] studied the activity characteristics of a particular animal and concluded that in a larger space and a limited search, Levy’s flight is the best search strategy in the local area because it can not only satisfy the search in a small range ensure to obtain the target but also satisfy the coarse search in a large range to avoid local optimum. In order to avoid the local optimization of the algorithm, the Levy’s flight mechanism is introduced in the round-up behavior. The characteristics of Levy’s flight are a random transformation process, and there is no fixed function representation. The study shows that the distribution density function of the variation of the Levy’s flight step can be approximated as follows:

Levy(s)|s|-1-β,0<β2, (13)

where s is the random step length of Levy’s flight behavior, and according to the literature [36], s is expressed as follows:

s=μ/|v|1/β. (14)

Parameter μ, v obeys normal distribution, i.e.,

μN(0,σμ2),vN(0,σv2), (15)

with

σμ={Γ(1+β)sin(πβ/2)Γ[(1+β)/2]2β-1/2}1/β,σv=1. (16)

In this paper, the motion steps generated by the Levy’s flight step length are applied to the round-up behavior of the basic wolf algorithm and improved

xidk+1=xidk+λ×s×|Gidk-xidk|, (17)

where λ is a random number between [0, 1] and s is the moving step length, and the Levy’s flight method can make the wolf algorithm jump out of the local optimization in the later search iteration, enhance the exploration ability of the algorithm, and accelerate the convergence speed of the algorithm.

4.4 Algorithm steps

The implementation workflow of our proposed OGL-WPA is shown in Fig 1. It contains main six steps as following.

Fig 1. Flow chart of this algorithm.

Fig 1

  • Step 1: Set the population individual number to be N, maximum number of iterations to be Max, number of searches for wolves to be h, max number of searches for detection wolves to be Tmax, search step size, attack step size, and initialize the population according to opposition-based learning.

  • Step 2: Choose the best individual leader wolf according to the description of genetic algorithm. The location is xlead, the target function value is Ylead and the wolf with the largest target value except the leader wolf is selected as the detection wolf. The formula (1) shall be followed to search the behavior until the Yi is obtained by a wolf greater than Ylead, or the number of searches reaches Tmax.

  • Step 3: Select Mnum fierce wolf randomly from the wolf pack except the leader wolf to attack the prey according to the formula (2). When the wolf smell of prey Yi > Ylead, Yi = Ylead instead of the first wolf to initiate the call behavior. Otherwise, the attack will be continued until the distance is less than dnear in formula (3).

  • Step 4: Perform round-ups in accordance with the Levy’s flight mechanism.

  • Step 5: Update the position of the leader wolf according to the leader mechanism and update the population according to the new mechanism.

  • Step 6: Decide whether the optimization is accurate or the number of iterations is reached to Tmax. If that case, the location of the output leader wolf is the best solution. Otherwise, return to step 2.

5 Experimental simulation

5.1 Experiment setting

To illustrate the superiority of our algorithm, we have compared OGL-WPA with some popular swarm intelligence algorithms such as the ant colony optimization (ACO), particle swarm optimization (PSO), Whale Optimization Algorithm(WOA) [37] and LSHADE [38], which are commonly used for solving complex optimization problems, such as task scheduling [39], data communications optimization [40], resource management [41], etc. Moreover, we also have compared OGL-WPA with the GWO algorithm [10, 30] and β-GWO algorithm [42], which represent another two latest wolf-related swarm intelligence algorithms. We have implemented all the algorithms with MATLAB 2018b, and run all the tests on a commodity laptop with an I7 CPU running at 1.8Ghz and a 8GB memory. For a general case, the parameters used for each algorithms are presented in Table 1.

Table 1. Main parameters of the 7 algorithms.

Algorithms Main parameters
ACO Population size is 100;
pheromone value is 0.005;
volatility coefficient of pheromone is 0.01;
path selection probability is 0.5.
PSO Population size is 100;
inertia weight is 0.5;
two learning factors is 0.5;
random number weight set to be 0.5.
GWO Population size is 100;
maximum iterations is 1000;
alpha = 1000, beta = 2000, GAMMA = 3000.
β-GWO Population size is 100;
maximum iterations is 1000;
scaling factor BETA = 0.8, alpha = 1000,
beta = 2000, gamma = 3000.
WOA Population size is 100;
maximum iterations is 1000;
a decrease linearly from 2 to 0.
LSHADE Population size is 100;
maximum iterations is 1000;
Storage size is 5;
The optimal selection rate is 0.11.
OGL-WPA Population size is 100;
maximum iterations 1000, h is 4, Tmax is 15;
Stepa is 2, Stepb is 1,
step factor is 1, λ is 0.5.

5.2 Test functions

We have used 12 typical test functions (shown in Table 2) to evaluate the performance of our algorithm. Specifically, these test functions are the basic functions for the design of the CEC competitions [43]. They have both high and low dimensions which can be used to illustrate the comprehensive comparison effects between our algorithm and other six algorithms. The experiment selects the average value, the minimum value, the maximum value and the standard deviation as the evaluation index, in which the maximum value and the minimum value reflect the quality of the solution. The average value reflects the accuracy that the algorithm can be achieved under the given number of iterations, and the standard deviation reflects the convergence speed of the algorithm.

Table 2. The used test functions in the evaluation.

No. Function Name Test Function Dimension Search Space Optimal Solution
F1 Sphere f(x)=i=1nxi2 2 [-100,100] 0
5 [-100,100] 0
10 [-100,100] 0
30 [-100,100] 0
F2 Schwefel 2.22 f(x)=i=1nxi+i=1nxi 2 [-100,100] 0
5 [-100,100] 0
10 [-100,100] 0
30 [-10,10] 0
F3 Schwefel 1.2 f(x)=i=1nj=1ixj 2 [-100,100] 0
5 [-100,100] 0
10 [-100,100] 0
30 [-100,100] 0
F4 Schwefel 2.21 f(x)=maxabsxi 2 [-10,10] 0
5 [-10,10] 0
10 [-10,10] 0
30 [-100,100] 0
F5 Rosenbrock f(x)=i=1n1100xi+1xi22+xi12 2 [-30,30] 0
5 [-30,30] 0
10 [-30,30] 0
30 [-30,30] 0
F6 Step f(x)=i=1nxi+0.52 2 [-100,100] 0
5 [-100,100] 0
10 [-100,100] 0
30 [-100,100] 0
F7 Rastrigin f(x)=i=1nxi210cos2πxi+10 2 [-5.12,5.12] 0
5 [-5.12,5.12] 0
10 [-5.12,5.12] 0
30 [-5.12,5.12] 0
F8 Ackley f(x)=20+e20exp151ni=1nxi2exp1ni=1ncos2πxi 2 [-32,32] 0
5 [-32,32] 0
10 [-32,32] 0
30 [-32,32] 0
F9 Griewank f(x)=11000i=1nxi2i1ncosxii+1 2 [-600,600] 0
5 [-600,600] 0
10 [-600,600] 0
30 [-600,600] 0
F10 Penalgy 1 f(x)=πn10sinπy1+i=1n1y1121+10sin2πyi+1+yn12+i=1nuxi,10,100,4,yi=1+xi+14,uxi,a,k,m=kxiam,xi>a0a<xi<akxiam,xi<a 2 [-50,50] 0
5 [-50,50] 0
10 [-50,50] 0
30 [-50,50] 0
F11 Bent Cigar f(x)=x12+106i=2nxi2 2 [-50,50] 0
5 [-50,50] 0
10 [-50,50] 0
30 [-50,50] 0
F12 Sumsquares-2 fx=i=1nixi2 2 [-50,50] 0
5 [-50,50] 0
10 [-50,50] 0
30 [-50,50] 0

5.3 Experimental results

For the 12 test functions, we just report the comparisons of fitness values of the five algorithms in Figs 213. Specifically, the detailed comparisons of the four indicators (i.e., minimum value, maximum value, average value and variance) achieved in the dimensions of 2, 5, 10, and 30 as well as the elapsed time have been given online as an S1 Appendix, the link of which has been given in the support information at the end of this work.

Fig 2. F1 iteration curve.

Fig 2

Fig 13. F12 iteration curve.

Fig 13

Fig 2 shows that the convergence of OGL-WPA happens when the number of iteration reaches 180. In comparison, the WOA happens when the number of iteration is 20, LSHADE is 260, PSO happens when the number of iteration is around 150, β-GWO is about 950, and the convergence of GWO happens when the number of iteration is around 900. From Fig 2, it can be found that the fitness of ACO decreases at the first half of iterations, and in oscillation process at the latter half, which means it can not converge. Moreover, we can see that our proposed algorithms can achieve the smallest fitness value, compared to other algorithms.

Fig 3 demonstrates that the convergence of OGL-WPA happens when the number of iterations is 800, WOA happens when the number of iteration is 20, LSHADE is 230, the convergence of PSO happens when the number of iteration is around 200. We can also see that fitness values of β-GWO and GWO decrease with increasing number of iteration. Although the whole processes of them are both smooth, they can not reach convergence at all. Moreover, for the ACO algorithm, we can see that its fitness value decreases and increases with increasing the number of iterations, demonstrating its poor performance in terms of convergence.

Fig 3. F2 iteration curve.

Fig 3

Fig 4 shows that the convergence of OGL-WPA happens when the number of iterations is about 50, WOA is 10, LSHADE is about 80, the convergence of PSO happens when the number of iteration is about 180, the convergence of β-GWO happens when the number of iteration is about 900, the convergence of GWO happens when the number of iteration is 760. Similar to previous results, the fitness value of ACO decreases and increases with increasing the number of iterations, which means that it cannot converge in a well way. Compared to the fitness values of all the algorithms, it can be seen that the proposed OGL-WPA can actually achieve the best performance for the optimization problem.

Fig 4. F3 iteration curve.

Fig 4

Fig 5 shows that the convergence of OGL-WPA happens when the number of iterations is about 350. Moreover, the convergence of WOA happens when the number of iteration is 10, LSHADE is 210, PSO is about 150. It also shows that the descending trend of β-GWO and GWO increases with increasing the numbers of iterations, and finally their fitness values can reach a fix point. Similar to the results above, ACO cannot converge well, and our proposed method can get a smallest fitness value for the test function, compared to all other algorithms.

Fig 5. F4 iteration curve.

Fig 5

Fig 6 shows that the convergence of OGL-WPA happens when the number of iterations is about 100. Moreover, WOA convergences when the number of iteration is 10, LSHADE is 220, the convergence of PSO achieves at the beginning of iteration. The results also show the overall descending trend of β-GWO and GWO appear with the number of iterations increase, then the whole processes of them are smooth.

Fig 6. F5 iteration curve.

Fig 6

Fig 7 shows that the convergence of OGL-WPA happens when the number of iterations is 780. The convergence of WOA is around 300, LSHADE is 230 and the convergence of PSO happens when the number of iteration is around 200. It also shows that the overall descending trend of β-GWO and GWO first appears with the increased number of iterations, and then the whole process goes to a smooth status. From Fig 7 it can be also found that the convergence of ACO decreases before the number of iterations is 100, and then the whole process is in oscillation with the increased number of iterations, which means it cannot converge well.

Fig 7. F6 iteration curve.

Fig 7

Fig 8 shows that the convergence of OGL-WPA happens when the number of iterations is around 400, WOA is around 20, LSHADE is 230, the convergence of PSO happens when the number of iteration is around 190. It also shows the β-GWO and GWO decrease with the number of iterations increase. The whole processes of them are smooth, but they can not reach convergence at all. The ACO alongs with the increase of the number of iterations, but the whole process is in oscillation causing convergence can not be reached.

Fig 8. F7 iteration curve.

Fig 8

Fig 9 shows that the convergence of OGL-WPA happens when the number of iterations is around 100, WOA is 10, LSHADE is 220, the convergence of PSO happens when the number of iteration is around 200. It also shows the β-GWO and GWO decrease with the increased number of iterations. The whole processes of them are smooth, but they can not reach convergence at all. It can be found that the ACO achieves convergence directly.

Fig 9. F8 iteration curve.

Fig 9

Fig 10 shows that the convergence of OGL-WPA happens when the number of iterations is around 450. In comparison, the convergence of WOA happens when the number of iteration is 10, LSHADE is 10, the convergence of PSO happens almost at the beginning. It also shows that the β-GWO and GWO decrease with the number of iterations increase. The whole processes of them are smooth, but they cannot reach convergence at all. Moreover, from the results, we can see that the ACO is in the status of oscillation with increasing the number of iterations, which means it cannot converge neither. In contrast, our algorithm can achieve the smallest fitness value in a quick way, demonstrating its strong capability in solving the optimization problems similar to the test function 9.

Fig 10. F9 iteration curve.

Fig 10

Fig 11 shows that that the convergence of OGL-WPA happens when the number of iterations is around 50, WOA is 50, LSHADE is 210, and the convergence of PSO happens when the number of iteration is around 200. It also shows the β-GWO and GWO decrease with the increased number of iterations, but they cannot reach convergence in a good way. Moreover, the ACO appears the decreasing trend when the number of iterations is about 200, and then it goes into the oscillation status, which demonstrates its poor performance in converge again.

Fig 11. F10 iteration curve.

Fig 11

Fig 12 shows that the convergence of OGL-WPA happens when the number of iterations is around 900, WOA is 10, LSHADE is 250, and PSO happens when the number of iteration is 200. Moreover, the convergence of β-GWO happens when the number of iteration is about 650. The ACO is in the status of oscillation in the whole process, which means it cannot converge. In addition, we can see that our algorithm can get the best fitness value again, compared to other approaches.

Fig 12. F11 iteration curve.

Fig 12

Fig 13 shows that the convergence of OGL-WPA happens when the number of iterations is around 900, WOA is 10, LSHADE is 220, PSO is 180. Moreover, the convergence of β-GWO happens when the number of iteration is about 900. Again, the ACO is in the oscillation status with increasing the number of iterations, demonstrating a poor convergence performance. Although the LSHADE can achieve the best performance in terms of fitness values, our method can get the second best result.

Generally, from the results reported in Figs 211 above, we can see that PSO is better than OGL-WPA on the convergence speed in functions F1, F2, F4, F5, F6, F7, F9, F10, F11 and F12. However, the fitness value of PSO is obviously lower than OGL-WPA. Moreover, GWO and β-GWO have done improvements on WPA, but they always perform worse than OGL-WPA in all the functions. In the meantime, WOA and LSHADE have done improvements on WPA, but they always perform worse than OGL-WPA in all the functions except F2, F6, F7, F12. Specifically, our proposed method can always converge well and can also achieve a good result for all the 12 test functions, demonstrating its advantages in processing different optimization problems.

Tables 3 and 4 report the running time of each algorithm in the presence of the 12 test functions. From the results there, we can see that the computational time of the proposed algorithm in this paper is generally longer than the simple algoritihm, such as ACO, PSO, GWO. However, it can converge well and always get a better fitness value. On the other hand, compared to the advanced algorithsm such as BETAGWO, WOA and LSHADE, the computational overhead of our algorithm is generally samller. In addition, we also have performed statistical analysis of our results and reported the results in Table 5. There, we have used the Wilcoxon test, and the R+ and R- [44] represent the maximum and minimum sum ranks, respectively. The parameters of Wilcoxon test are set to 0.01 and 0.05. From the results presented there, we can see that in all cases, the R+ value provided by OGL-WPA is higher than R-. Therefore, we can say that OGL-WPA can indeed achieve a better performance in generally, compared to other methods.

Table 3. Computational time of each algorithm—Part I.

Algo. Dim. F1 F2 F3 F4 F5 F6
OGL-WPA 2 1.424 2.036 2.39 1.287 1.507 1.204
5 2.647 4.038 8.249 2.742 6.084 3.137
10 4.863 7.81 26.695 5.396 13.011 4.96
30 15.95 27.58 236.85 18.159 75.258 17.927
ACO 2 0.289 0.367 0.611 0.284 0.325 0.228
5 0.646 0.717 1.382 0.629 1.119 0.595
10 1.239 1.387 3.858 1.27 1.921 1.672
30 3.361 3.715 22.7 3.994 8.2 3.702
PSO 2 0.192 0.172 0.279 0.161 0.21 0.148
5 0.359 0.346 0.807 0.338 0.527 0.397
10 0.995 0.99 2.484 0.683 1.176 0.732
30 2.247 2.711 16.536 1.931 5.198 1.923
GWO 2 0.255 0.206 0.302 0.191 0.34 0.223
5 0.521 0.545 1.005 0.536 0.694 0.531
10 1.098 1.239 2.908 1.198 1.619 1.174
30 4.475 4.778 19.29 4.782 7.936 4.793
BETAGWO 2 0.734 1.083 1.156 0.783 0.808 0.924
5 4.153 4.139 5.847 4.004 4.605 4.198
10 12.582 12.176 14.755 13.709 15.659 14.997
30 103.31 104.15 126.333 102.097 113.266 145.464
WOA 2 1.050 1.518 1.564 0.954 1.308 1.143
5 5.103 4.826 6.885 5.090 5.877 5.560
10 16.386 15.248 18.926 17.426 22.179 17.885
30 134.963 150.652 188.610 147.939 135.533 161.436
LSHADE 2 2.304 2.693 3.02 1.975 2.64 2.337
5 13.057 10.409 17.162 11.368 13.144 14.673
10 32.532 34.434 40.815 40.94 37.705 39.829
30 294.629 339.91 397.47 320.735 296.749 442.459

Table 4. Computational time of each algorithm—Part II.

Algo. Dim. F7 F8 F9 F10 F11 F12
OGL-WPA 2 2.148 1.689 1.042 2.667 1.245 3.241
5 4.913 3.52 2.558 9.843 4.544 11.542
10 9.577 7.297 6.31 26.326 12.883 33.248
30 38.945 31.309 33.09 173.54 72.544 157.541
ACO 2 0.29 0.435 0.466 1.071 0.978 1.921
5 0.762 1.046 0.954 3.532 2.722 4.581
10 1.345 1.83 1.589 9.227 7.821 12.441
30 4.721 6.104 5.674 74.516 45.548 73.954
PSO 2 0.258 0.252 0.23 0.648 0.712 0.821
5 0.487 0.526 0.464 2.012 2.124 2.459
10 0.791 1.127 0.976 6.464 5.887 7.241
30 2.481 3.484 3.059 52.312 40.814 55.631
GWO 2 0.214 0.289 0.229 0.67 0.692 0.721
5 0.589 0.771 0.64 2.259 2.315 2.812
10 1.3 1.645 1.379 6.668 6.887 7.354
30 5.224 6.05 5.605 50.534 51.527 58.972
BETAGWO 2 0.997 0.995 0.954 1.397 1.548 1.821
5 4.118 4.115 3.976 5.468 5.955 7.0581
10 14.669 14.315 14.208 19.591 21.763 24.542
30 139.55 123.03 106.73 154.03 167.821 182.972
WOA 2 1.382 1.506 1.326 1.745 2.125 2.265
5 5.655 5.494 4.766 8.533 8.359 8.853
10 18.894 20.534 18.855 26.028 29.083 29.670
30 195.772 193.093 113.920 240.921 214.840 210.727
LSHADE 2 2.767 2.415 3.083 3.635 4.476 5.527
5 12.69 12.214 11.58 16.552 14.888 25.139
10 38.338 39.793 41.004 60.072 54.519 64.823
30 339.547 341.349 350.002 425.281 479.914 527.511

Table 5. Analysis of Wilcoxon statistical test results.

Algorithm R+ R- P-value
OGL-WPA vs ACO 335 56 0.0012
OGL-WPA vs PSO 204 75 0.0505
OGL-WPA vs GWO 178 98.5 0.2022
OGL-WPA vs BETAGWO 133 125 0.8596
OGL-WPA vs WOA 144 115 0.4391
OGL-WPA vs LSHADE 107 105 0.9773

6 Conclusion

In this paper, we have proposed an improved WPA approach called OGL-WPA. Specifically, we focus on leveraging the existing intelligent techniques, i.e., Opposition-based learning, Genetic algorithm and Levy’s flight, to handle the issues in WPA (e.g., popular initialization, local optimum and convergence speed). To the best of our knowledge, this is the first work on how to improve the WPA in a comprehensive way. We have given the detailed design and implementation of OGL-WPA. Moreover, we also have compared our approach with some other swarm intelligent algorithms over different test functions with extensive experiments. Our experimental results have shown that the proposed OGL-WPA has better global search and local search capability, especially in the cases for multi-peak and high-dimensional functions.

Supporting information

S1 Appendix. Additional experimental results.

The results about the comparison of different dimensions are available at 10.5281/zenodo.5109519.

(PDF)

Data Availability

We have uploaded the source code we have used in the paper to Github and created a DOI: 10.5281/zenodo.5171844. Readers can use the source code to replicate the reported results in the paper.

Funding Statement

F. Cheng was supported by the Undergraduate Education Research and Reform Project of Southwest Jiaotong University in 2020 (NO. 20201035-07). C. Liu was supported by the National Science Foundation of China (61902222), the Taishan Scholar Youth Program of Shandong Province (tsqn201909109).

References

  • 1.Yang C. G., Tu X. Y. and Chen J., “Algorithm of mamiage in honey bees optimization based on the wolf pack search,” in Proc Intelligent Pervasive Computing, 2007, pp. 462–467. [Google Scholar]
  • 2.Wu H. S., Zhang F. M., “Wolf pack algorithm for unconstrained global optimization.” Mathematical Problems in Engineering, 2014. [Google Scholar]
  • 3.Chen Y., Mei Y., Yu J., Su X. and Xu N., “Three-dimensional unmanned aerial vehicle path planning using modified wolf pack search algorithm”. Neurocomputing, vol. 266, pp. 445–457, 2017. doi: 10.1016/j.neucom.2017.05.059 [DOI] [Google Scholar]
  • 4.Liu Q., Cheng L., Jia A. L., Liu C., “Deep reinforcement learning for communication flow control in wireless mesh networks”. IEEE Network, vol. 35, no. 2, pp. 112–119, 2021. doi: 10.1109/MNET.011.2000303 [DOI] [Google Scholar]
  • 5.Wu H., Li H., Xiao R., Liu J., “Modeling and simulation of dynamic ant colony’s labor division for task allocation of UAV swarm”. Physica A: Statistical Mechanics and its Applications, vol. 491, pp. 127–141, 2018. doi: 10.1016/j.physa.2017.08.094 [DOI] [Google Scholar]
  • 6.Liu Q., Cheng L., Alves R., Ozcelebi T., Kuipers F., Xu G., et al., “Cluster-based flow control in hybrid software-defined wireless sensor networks”. Computer Networks, vol. 187, no. 107788, 2021. [Google Scholar]
  • 7.Chen X. Y., Tang C. J., Wang J. et al., “A novel hybrid based on wolf pack algorithm and differential evolution algorithm,” in Proc Symposium on Computational Intelligence and Design, 2018, pp. 69–74. [Google Scholar]
  • 8.Li H. and Wu H., “N oppositional wolf pack algorithm for parameter identification of the chaotic systems,” Optik, vol. 127, no. 20, pp. 9853–9864, Oct. 2016. doi: 10.1016/j.ijleo.2016.07.056 [DOI] [Google Scholar]
  • 9.Xiu Z. and Hua W. Z., “Improved wolf pack algorithm based on tent chaotic mapping and Levy flight,” in Proc Robots and Intelligent System, 2017, pp. 165–169. [Google Scholar]
  • 10.Mirjalili S., Mohammad S., Lewis A. “Grey wolf optimizer”, Advances in engineering software, vol. 69, pp. 46–61, 2014. doi: 10.1016/j.advengsoft.2013.12.007 [DOI] [Google Scholar]
  • 11.Chen H. R., Cui L., Guo Q. et al., “Improved Particle Swarm Optimization Using Wolf Pack Search,” Journal of Physics: Conference Series, vol. 1176, no. 5, Mar. 2018. [Google Scholar]
  • 12.Jiang Y., Zhang L. and Liu J., “The Path Planning of Mobile Sink Based on Wolf Pack Algorithm,” in Proc Intelligent Transportation, Big Data and Smart City, 2019, pp. 147–150. [Google Scholar]
  • 13.Dong R., Wang S., Wang G. et al., “Hybrid Optimization Algorithm Based on Wolf Pack Search and Local Search for Solving Traveling Salesman Problem,” Journal of Shanghai Jiaotong University (Science), vol. 24, no. 1, pp.41–47, Feb. 2019. doi: 10.1007/s12204-019-2039-9 [DOI] [Google Scholar]
  • 14.Liang W. H., He J. H., Wang S. X. et al., “Improved cluster collaboration algorithm based on wolf pack behavior,” Cluster Computing, pp.1–16, Feb. 2018. [Google Scholar]
  • 15.Zhang X., “Short-term load forecasting for electric bus charging stations based on fuzzy clustering and least squares support vector machine optimized by wolf pack algorithm,” Energies, vol. 11, no. 6, pp. 1449, Jun. 2018. doi: 10.3390/en11061449 [DOI] [Google Scholar]
  • 16.Han Z. H., Tian X. T., Ma X. F. et al., “Scheduling for Re-entrant Hybrid Flowshop Based on Wolf Pack Algorithm,” IOP Conference Series: Materials Science and Engineering, vol. 382, no. 3, pp.1–6, Jul. 2018. [Google Scholar]
  • 17.Gao Y. J., Zhang F. M., Zhao Y. et al., “Quantum-Inspired Wolf Pack Algorithm to Solve the 0-1 Knapsack Problem,” Mathematical Problems in Engineering, vol. 2018, pp. 1–10, Jun. 2018. doi: 10.1155/2018/4791590 [DOI] [Google Scholar]
  • 18.Chen Y. B., Mei Y. S., Yu J. Q. et al., “Three-dimensional unmanned aerial vehicle path planning using modified wolf pack search algorithm,” Neurocomputing, vol. 266, pp. 445–457, Nov. 2017. doi: 10.1016/j.neucom.2017.05.059 [DOI] [Google Scholar]
  • 19.Gupta S. and Deep K., “A novel random walk grey wolf optimizer,” Swarm and evolutionary computation, vol. 44, pp. 101–112, Feb. 2019. doi: 10.1016/j.swevo.2018.01.001 [DOI] [Google Scholar]
  • 20.Long W., Wu T., Cai S. et al., “A Issuevel Grey Wolf Optimizer Algorithm with Refraction Learning,” IEEE Access, vol.7, pp. 57805–57819, Apr. 2019. doi: 10.1109/ACCESS.2019.2910813 [DOI] [Google Scholar]
  • 21.Arora S., Singh H., sharma M. et al., “A New Hybrid Algorithm Based on Grey Wolf Optimization and Crow Search Algorithm for Unconstrained Function Optimization and Feature Selection,” IEEE Access, vol. 7, pp. 26343–26361, Feb. 2019. doi: 10.1109/ACCESS.2019.2897325 [DOI] [Google Scholar]
  • 22.Al-Tashi Q., Kadir S., Rais H., et al., “Binary Optimization Using Hybrid Grey Wolf Optimization for Feature Selection,” IEEE Access, vol. 7, pp. 39496–39508, Mar. 2019. doi: 10.1109/ACCESS.2019.2906757 [DOI] [Google Scholar]
  • 23.Singh D. and Dhillon J. S., “Ameliorated grey wolf optimization for ecoIssuemic load dispatch problem,” Energy, vol. 169, pp. 398–419, Feb. 2019. doi: 10.1016/j.energy.2018.11.034 [DOI] [Google Scholar]
  • 24.Barraza J., Rodríguez L., Castillo O. et al., “A new hybridization approach between the fireworks algorithm and grey wolf optimizer algorithm,” Journal of Optimization, vol. 2018, pp. 1–18, May. 2018. doi: 10.1155/2018/6495362 [DOI] [Google Scholar]
  • 25.Gaidhane P. J. and Nigam M. J., “A hybrid grey wolf optimizer and artificial bee colony algorithm for enhancing the performance of complex systems,” Journal of computational science, vol. 27, pp. 284–302, Jul. 2018. doi: 10.1016/j.jocs.2018.06.008 [DOI] [Google Scholar]
  • 26.Zhi J. T., Jin L. L. and Guo L. W., “An improved hybrid grey wolf optimization algorithm,” Soft Computing, vol. 23, no. 15, pp. 1–15, Aug. 2019. [Google Scholar]
  • 27.Al-Wajih R., Abdulkadir S. J., Aziz N., et al., “Hybrid Binary Grey Wolf With Harris Hawks Optimizer for Feature Selection,” IEEE Access, vol. 9, pp.31662–31677, Feb, 2021. doi: 10.1109/ACCESS.2021.3060096 [DOI] [Google Scholar]
  • 28.Banaie-Dezfouli M., Nadimi-Shahraki M. and Beheshti Z., “R-GWO: Representative-based grey wolf optimizer for solving engineering problems,” Applied Soft Computing, vol. 106, pp. 107328, Jul, 2021. doi: 10.1016/j.asoc.2021.107328 [DOI] [Google Scholar]
  • 29.Maharana D. and Kotecha P., “Optimization of Job Shop Scheduling Problem with Grey Wolf Optimizer and JAYA Algorithm,” in Proc Smart InIssuevations in Communication and Computational Sciences, pp. 47–58, 2018. [Google Scholar]
  • 30.Zapotecas-Martínez S., García-Nájera A. and López-Jaimes A., “Multi-objective grey wolf optimizer based on decomposition,” Expert Systems with Applications, vol. 120, pp. 357–371, Apr. 2019. doi: 10.1016/j.eswa.2018.12.003 [DOI] [Google Scholar]
  • 31.Kaur P. and Mehta S., “Efficient computation offloading using grey wolf optimization algorithm,” AIP Conference Proceedings, vol. 2061, no. 1, Jan. 2019. [Google Scholar]
  • 32.Al-Moalmi A., Luo J., Salah A. et al., “Optimal Virtual Machine Placement Based on Grey Wolf Optimization,” Electronics, vol. 8, no. 3, pp. 1–22, Mar. 2019. doi: 10.3390/electronics8030283 [DOI] [Google Scholar]
  • 33.Long W., Cai S., Tang M. et al., “An efficient and robust grey wolf optimizer algorithm for large-scale numerical optimization,” Soft Computing, pp. 1–30, Mar. 2019. [Google Scholar]
  • 34.H. R. Tizhoosh, “Opposition-based learning:a new sheme for machine intelligence,” in Proc Computational Intelligence for Modelling, Control and Automation and Intelligent Agents, Web Technologies and InternetCommerce, pp. 695–701, 2006.
  • 35.Edwards A. M., Phillips R. A., Watkins N. W. et al., “Revisiting Levy flight search patterns of wandering albatrosses, bumblebees and deer,” Nature, vol. 449, pp. 1044–1048, Oct. 2007. doi: 10.1038/nature06199 [DOI] [PubMed] [Google Scholar]
  • 36.Mantegna R. N. and Stanley H. E., “Stochastic process withe ultraslow convergence to a Gaussian:the truncated Levy flight,” Physical Review Letters, vol. 73, no. 22, pp. 2946–2949, Nov. 1994. doi: 10.1103/PhysRevLett.73.2946 [DOI] [PubMed] [Google Scholar]
  • 37.Mirjalili S. and Lewis A., “The whale optimization algorithm,” Advan ces in Engineering Software, vol. 95, pp. 51–67, 2016. doi: 10.1016/j.advengsoft.2016.01.008 [DOI] [Google Scholar]
  • 38.R. Tanabe and A. S. Fukunaga, “Improving the search performance of SHADE using linear population size reduction,” 2014 IEEE congress on evolutionary computation (CEC), pp. 1658–1665, 2014
  • 39.Chen X., Cheng L., Liu C., Liu Q., Liu J., Mao Y., et al., “A woa-based optimization approach for task scheduling in cloud computing systems,” IEEE Systems Journal, vol. 14, no. 3, pp. 3117–3128, 2021. doi: 10.1109/JSYST.2019.2960088 [DOI] [Google Scholar]
  • 40.Cheng L., Wang Y., Liu Q., Epema D., Liu C., Mao Y., et al., “Network-aware locality scheduling for distributed data operators in data centers,” IEEE Transactions on Parallel and Distributed Systems, vol. 32, no. 6, pp. 1494–1510, 2021. doi: 10.1109/TPDS.2021.3053241 [DOI] [Google Scholar]
  • 41.W. Zheng, M. Tynes, H. Gorelick, Y. Mao, L. Cheng, Y. Hou, “FlowCon: Elastic flow configuration for containerized deep learning applications,” Proceedings of the 48th International Conference on Parallel Processing, pp. 1–10, 2019
  • 42.Saxena A., Kumar R. and Das S., “β-Chaotic map enabled Grey Wolf Optimizer,” Applied Soft Computing, vol. 75, 2019. doi: 10.1016/j.asoc.2018.10.044 [DOI] [Google Scholar]
  • 43.G. Wu, R. Mallipeddi, P. Suganthan, “Problem definitions and evaluation criteria for the CEC 2017 competition on constrained real-parameter optimization,” Technical Report, 2017
  • 44.Mohamed A. W. and Mohamed A. K., “Adaptive guided differential evolution algorithm with novel mutation for numerical optimization,” International Journal of Machine Learning and Cybernetics, vol. 10, No. 2, pp. 253–277, 2019. doi: 10.1007/s13042-017-0711-7 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 Appendix. Additional experimental results.

The results about the comparison of different dimensions are available at 10.5281/zenodo.5109519.

(PDF)

Data Availability Statement

We have uploaded the source code we have used in the paper to Github and created a DOI: 10.5281/zenodo.5171844. Readers can use the source code to replicate the reported results in the paper.


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES