Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2015 Dec 24;2016:8085953. doi: 10.1155/2016/8085953

An Enhanced Artificial Bee Colony Algorithm with Solution Acceptance Rule and Probabilistic Multisearch

Alkın Yurtkuran 1,*, Erdal Emel 1
PMCID: PMC4706902  PMID: 26819591

Abstract

The artificial bee colony (ABC) algorithm is a popular swarm based technique, which is inspired from the intelligent foraging behavior of honeybee swarms. This paper proposes a new variant of ABC algorithm, namely, enhanced ABC with solution acceptance rule and probabilistic multisearch (ABC-SA) to address global optimization problems. A new solution acceptance rule is proposed where, instead of greedy selection between old solution and new candidate solution, worse candidate solutions have a probability to be accepted. Additionally, the acceptance probability of worse candidates is nonlinearly decreased throughout the search process adaptively. Moreover, in order to improve the performance of the ABC and balance the intensification and diversification, a probabilistic multisearch strategy is presented. Three different search equations with distinctive characters are employed using predetermined search probabilities. By implementing a new solution acceptance rule and a probabilistic multisearch approach, the intensification and diversification performance of the ABC algorithm is improved. The proposed algorithm has been tested on well-known benchmark functions of varying dimensions by comparing against novel ABC variants, as well as several recent state-of-the-art algorithms. Computational results show that the proposed ABC-SA outperforms other ABC variants and is superior to state-of-the-art algorithms proposed in the literature.

1. Introduction

Optimization techniques play an important role in the field of science and engineering. Over the last five decades, numerous algorithms have been developed to solve complex optimization algorithms. Since more and more present-day problems turn out to be nonlinear, multimodal, discontinuous, or dynamic in nature, derivative-free, nonexact solution methods attract ever-increasing attention. Evolutionary biology or swarm behaviors inspired most of these methods. There have been several classes of algorithms proposed in this evolutionary or swarm intelligence framework including genetic algorithms [1, 2], memetic algorithms [3], differential evolution (DE) [4], ant colony optimization (ACO) [5], particle swarm optimization (PSO) [6], artificial bee colony algorithm (ABC) [7], cuckoo search [8], and firefly algorithm [9].

The ABC is a biologically inspired population-based metaheuristic algorithm that mimics the foraging behavior of honeybee swarms [7]. Due to its simplicity and ease of application, the ABC has been widely used to solve both continuous and discrete optimization problems since its introduction [10]. It has been shown that ABC tends to suffer poor intensification performance on complex problems [1113]. To improve the intensification performance of ABC, many researchers have focused on the search rules as they control the tradeoff between diversification and intensification. Diversification means the ability of an algorithm to search for unvisited points in the search region, whereas intensification is the process of refining those points within the neighborhood of previously visited locations to improve solution quality. Various new search strategies, mostly inspired from PSO and DE, have been proposed in the literature. Zhu and Kwong [14] proposed a global best guided ABC, which utilizes the global best individual's information within the search equation similar to PSO. Gao et al. [15] introduced another variant of global best ABC. Inspired by DE, Gao and Liu [13] introduced a modified version of the ABC in which ABC/Best/1 and ABC/Rand/1 were employed as local search equations. Kang et al. [16] described the Rosenbrock ABC, which combines Rosenbrock's rotational method with the original ABC. To improve diversification, Alatas [11] employed chaotic maps for initialization and chaotic searches within a search strategy. Akay and Karaboga [17] introduced a modified version of the ABC in which frequency of perturbation is controlled adaptively and the ratio of variance operator was introduced. Liao et al. [18] proposed a detailed experimental analysis and comparison of an ABC variant with different search equations. Gao et al. [19] introduced two new search equations for onlooker and employed bee phases and a new robust comparison technique for candidate solutions. ABC. Qiu et al. [20] were inspired from DE/current-to-best/1 strategy in DE algorithm and proposed a modified ABC. Banitalebi et al. [21] proposed an enhanced compact ABC, which did not store the actual population of candidate solution; instead their approach employed probabilistic representation. Wang et al. [22] presented multistrategy ABC, in which a pool of different search strategies was constructed and various search strategies were used during the search process. Gao et al. [23] introduced a bare bones ABC with parameter adaptation and fitness-based neighborhood to improve the intensification performance of standard ABC. Ma et al. [24] reduced the redundant search moves and maintained the diversity of the swarm by introducing hybrid ABC with life cycle and social learning. Furthermore, ABC has been successfully applied to solve various types of optimization problems, such as production scheduling [25, 26], vehicle routing [27], location-allocation problem [28], image segmentation [29], wireless sensor network routing [30], leaf-constrained minimum spanning tree problem [31], clustering problem [32], fuel management optimization [33], and many others [3436]. Readers can refer to Karaboga et al. [10] for an extensive literature review of the ABC and its applications.

This study presents an enhanced ABC with solution acceptance rule and probabilistic multisearch (ABC-SA) in order to solve global optimization problems efficiently. In ABC-SA, three search mechanisms with different diversification and intensification characteristics are employed. Moreover, search mechanism selection probabilities p s1, p s2, and p s3 are introduced to control the balance between diversification and intensification. In our proposed approach, a search mechanism is established using the selection probabilities to generate a new neighbor solution from the current one. Additionally, a solution acceptance rule is implemented, in which not only better solutions but also worse solutions may be accepted by using a probability function. A nonlinearly decreasing acceptance probability function is employed, thus allowing worse solutions to be more likely accepted in the early phases of the search. Therefore, ABC-SA algorithm explores the search space more widespread, especially in the early phases of the search process. By using solution acceptance rule and implementing different search mechanisms of contrasting nature, ABC-SA balances the trade-off between diversification and intensification efficiently. The proposed approach is tested on six different benchmark functions with varying dimensions and compared to novel ABC, PSO, and DE variants. Computational results reveal that ABC-SA outperforms competitor algorithms in terms of solution quality.

The main contributions of the proposed study are as follows:

  1. Three different search mechanisms are employed with varying diversification and intensification abilities. Probabilistic multisearch with predetermined probability values are employed to determine the search mechanism to be used to generate candidate solutions. Therefore, ABC-SA explores and exploits the search space efficiently.

  2. Instead of a greedy selection, a new candidate solution acceptance rule is integrated, where a worse solution may have a chance to be accepted as new solution. By the help of this new acceptance rule, ABC-SA achieves better diversification performance, specifically in the early phases of the search.

The remainder of this paper is structured as follows: Section 2 presents the traditional ABC; Section 3 introduces the proposed framework; the instances, parameter settings, and computational results are presented in Section 4 and finally Section 5 concludes the paper.

2. Artificial Bee Colony Algorithm

The ABC has inspired from the organizational nature and foraging behavior of honeybee swarms. In the ABC algorithm, the bee colony comprises three kinds of bees: employed bees, onlooker bees, and scout bees. Each bee has a specialized task in the colony to maximize the nectar amount that is stored in the hive. In ABC, each food source is placed in the D-dimensional search space and represents a potential solution to the optimization problem. The amount of nectar in the food source is assumed to be the fitness value of a food source. Generally, the number of employed and onlooker bees is the same and equal to the number of food sources.

Each employed bee belongs to a food source and is responsible for mining the corresponding food source. Then, employed bees pass the nectar information to onlooker bees in the “dance area.” Onlooker bees wait in the hive and select a food source to mine based on the information coming from the employed bees. Here, more beneficial food sources will have higher selection probabilities to be selected by onlooker bees. In ABC, in order to decide if a food source is abandoned or not, trial counters and a predetermined limit parameter are used. If a solution represented by a food source does not improve during a number of trials (limit), the food source is abandoned. When the food source is abandoned, the corresponding employed bee will become a scout bee and randomly generate a new food source and replace it with the abandoned one.

The ABC algorithm consists of four main steps: initialization, employed bee phase, onlooker bee phase, and scout bee phase. After the initialization step, the other three main steps of the algorithm are carried out repeatedly in a loop until the termination condition is met. The main steps of the ABC algorithm are as follows.

Step 1 (initialization). —

In the initialization step, the ABC generates a randomly distributed population of SN solutions (food sources), where SN also denotes the number of employed or onlooker bees. Let x i = {x i,1, x i,2,…, x i,D} represent the ith food source, where D is the problem size. Each food source is generated within the limited range of jth index by

xi,j=xjmin+φi,jxjmaxxjmin, (1)

where i = 1, 2,…, SN, j = 1, 2,…, D,   φ i,j is a uniformly distributed random real number in [0,1], and x j min and x j max are the lower and upper bounds for the dimension j, respectively. Moreover, a trial counter for each food source is initialized.

Step 2 (employed bee phase). —

In the employed bee phase, each employed bee visits a food source and generates a neighboring food source in the vicinity of the selected food source. Employed bees search a new solution, v i, by performing a local search around each food source i = 1, 2,…, SN as follows:

vi,j=xi,j+ϕxi,jxr1,j, (2)

where j is a randomly selected index j ∈ {1, 2,…, D} and r1 ∈ {1,2,…, SN}  is a randomly chosen food source that is not equal to i; that is, (r1 ≠ i). ϕ is a random number within the range [−1, 1] generated specifically for each i and j combination. A greedy selection is applied between x i and v i by selecting the better one.

Step 3 (onlooker bee phase). —

Unlike the employed bees, onlooker bees select a food source depending on the probability value p, which is determined by nectar amount associated with that food source. The value of p i is calculated for ith food source as follows:

pi=fitij=1SNfitj, (3)
fiti=11+fi,fi01+absfi,fi<0, (4)

where fiti is the fitness value of solution i and calculated as in (4) for minimization problems. Different fitness functions are employed for maximization problems. By using this type of roulette wheel based probabilistic selection, better food sources will more likely be visited by onlooker bees. Therefore, onlooker bees try to find new candidate food sources around good solutions. Once the onlooker bee chooses the food source, it generates a new solution using (2). Similar to the employed bee phase, a greedy selection is carried out between x i and v i.

Step 4 (scout bee phase). —

A trial counter is associated with each food source, which depicts the number of tries that the food source cannot be improved. If a food source cannot be improved for a predetermined number of tries (limit) during the onlooker and employed bee phases, then the employed bee associated with that food source becomes a scout bee. Then, the scout bee finds a new food source using (1). By implementing the scout bee phase, the ABC algorithm easily escapes from minimums and improves its diversification performance.

It should be noted that, in the employed bee phase, a local search is applied to each food source, whereas in the onlooker bee phase better food sources will more likely be updated. Therefore, in ABC algorithm, the employed bee phase is responsible for diversification whereas the onlooker bee phase is responsible of intensification. The flow chart of the ABC is given in Figure 1.

Figure 1.

Figure 1

Flowchart of ABC.

3. Proposed Framework

In this section, the proposed algorithm is described in detail. First, a solution acceptance rule is presented. Second, a novel probabilistic multisearch mechanism is proposed. Finally, the complete ABC-SA mechanism is given.

3.1. Solution Acceptance Rule

In order to strengthen the diversification ability of ABC-SA mechanism, a solution acceptance rule is proposed. Instead of greedy selection in both employed and onlooker bee phases, an acceptance probability is given to worse solutions. The main idea behind this acceptance probability is not to restrict the search moves to only better solutions. By accepting a worse solution, the procedure may escape from a local optimum and explore the search space effectively. In ABC-SA algorithm, if a worse solution is generated, it is accepted if the following condition holds:

r<pa (5)
pa=po1+cositer/Max.iterπ2, (6)

where r is a random real number within [0,1], p a is the acceptance probability, p o denotes the initial probability, and iter and Max.iter represent the current iteration number and the maximum iteration number, respectively. According to (6), the acceptance probability p a is nonlinearly decreased from p 0 to zero during the search process. As can be seen from (6), p a = 0 when iter = Max.iter and the range of p a is [0, p 0]. A typical p a graph is given in Figure 2 and Algorithm 1 presents the implementation of the solution acceptance rule. At this point, it is important to note that the trial counter is incremented, whether a worse candidate solution is accepted or not.

Figure 2.

Figure 2

Acceptance probability curve.

Algorithm 1.

Algorithm 1

Solution acceptance rule.

3.2. Probabilistic Multisearch Strategy

In standard ABC, a candidate solution is generated using the information of the parent food source with the guidance of the term ϕ i,j(x i,jx r1,j) in (3). However, there is no guarantee that a better individual influences the candidate solution; therefore, it is possible to have a poor convergence speed and intensification performance. In fact, studying search equations is a trending topic to improve the ABC's performance. Recently, numerous search equations have been proposed, such as [1316, 19, 20, 37, 38]. It is well known that the balance between diversification and intensification is the most critical part of any metaheuristic algorithm.

In ABC-SA approach, instead of employing a single search mechanism throughout the search process, a probabilistic multisearch mechanism with three different search rules is used. A probabilistic selection is employed using predefined probability parameters to select the search rule within both employed and onlooker bee phases. The three search rules which were proposed by [7, 13, 14], respectively, are presented as follows:

vi,j=xi,j+ϕxi,jxr1,j, (7)
vi,j=xi,j+ϕxi,jxr1,j+ψxgbest,jxi,j, (8)
vi,j=xlbest,j+ϕxi,jxr1,j, (9)

where i is a food source, j is a randomly selected index for all i = 1, 2,…, SN, and j ∈ {1, 2,…, D}, respectively. r1 is a randomly chosen food source where r1 ≠ i. gbest stands for the global best solution, whereas lbest is the best solution in the current population. ϕ represents a real random number within the range of [−1, 1] [7]. Finally, ψ is a real random number within the range of [0, C] where C is a predetermined number [13].

Equation (7) is the original search rule, which was discussed in previous sections. Equation (8) is presented to improve the intensification capability of ABC. Equation (8) uses the information provided by the global best solution which is similar to PSO. In (9), lbest guides the search with the random effect of the term ϕ(x i,jx r1,j). Equation (7) has an explorative character, whereas (9) favors intensification. On the other hand, (8) explores the search space using the second term and exploits effectively by the third term. Therefore, (8) balances diversification and intensification performance. In summary, the proposed ABC-SA uses three different search rules to achieve a trade-off between diversification and intensification. In ABC-SA, search probabilities p s1, p s2, and p s3 are introduced such that ∑k p sk = 1 and k = 1,2, 3 to select a search rule to be used in the employed and the onlooker bee phases. A roulette wheel method is employed with three cumulative ranges 0 ≤ p s1P s1, 0 ≤ p s2 < P s2P s1, and 0 ≤ p s3P s3P s2 assigned to (7), (8), and (9), respectively, where P s3 = 1. Algorithm 2 shows the mechanism of probabilistic multisearch.

Algorithm 2.

Algorithm 2

Probabilistic multisearch.

3.3. Proposed Approach

Algorithm 3 summarizes the ABC-SA framework. The novel parts of the ABC-SA mechanism are the probabilistic multisearch (Lines 9 and 19) and the solution acceptance rule (Lines 10 and 20) sections.

Algorithm 3.

Algorithm 3

ABC-SA framework.

4. Computational Results

4.1. Test Instances

In literature, many test functions with different characters were used to test algorithms [11, 13, 15, 17, 1921, 34, 3740]. Unimodal functions have one local minimum as the global optimum. These functions are generally used to test the intensification ability of algorithms. Multimodal functions have one or more local optimums which may be the global optimum. Therefore, diversification behavior of algorithms is analyzed on multimodal instances. Separable functions can be written as sum of n functions with one variable, whereas nonseparable functions can not be reformulated as subfunctions. In this study, to analyze the performance of the proposed ABC-SA algorithm, 13 scalable benchmark functions with dimensions D = 50,  D = 100, and D = 200 are used and listed in Table 1. They are Rosenbrock, Ackley, Rastrigin, Weierstrass, Schwefel 2.26, Shifted Sphere, Shifted Schwefel 1.2, Shifted Rosenbrock, Shifted Rastrigin, Step, Penalized 2, and Alpine. In Table 1, function label, name, formulation, type (UN: unimodal and nonseparable, MS: multimodal and separable, and MN: multimodal and nonseparable), range, and optimal values (f(x )) are given.

Table 1.

Test functions used in experiments.

Label Name Formulation Type Range f(x )
F1 Rosenbrock f1X=i=1D-1100xi+1-xi22+xi-12 UN [−2.048, 2.048]D 0
F2 Ackley f2X=-20exp-0.21Di=1Dxi2-exp1Di=1Dcos2πxi+20+e MS [−32.768, 32.768]D 0
F3 Rastrigin f3X=i=1Dxi2-10cos2πxi+10 MS [−5.12, 5.12]D 0
F4 Griewank f4X=14000i=1Dxi2-i=1Dcosxii+1 MN [−600, 600]D 0
F5 Weierstrass f5X=i=1Dk=0kmaxakcos2πbkxi+0.5-Dk=0kmaxakcos2πbk0.5,a=0.5,b=3,kmax=20 MS [−0.5, 0.5]D 0
F6 Schwefel 2.26 f6X=418.9829×D-i=1D-xisinxi MS [−500, 500]D −418.98 × D
F7 Shifted Sphere f7X=i=1Dzi2-fbias,z=x-o,fbias=-450 US [−100, 100]D f bias
F8 Shifted Schwefel 1.2 f8X=i=1Dj=1izj+fbias,z=x-o,fbias=-450 UN [−100, 100]D f bias
F9 Shifted Rosenbrock f9X=i=1D-1100zi2-zi+12+zi-12+fbias,z=x-o+1,fbias=390 MN [−100, 100]D f bias
F10 Shifted Rastrigin f10X=i=1Dzi2-10cos2πzi+10+fbias,z=x-o,fbias=-330 MS [−5, 5]D f bias
F11 Step f11X=i=1Dxi+0.52 US [−100, 100]D 0
F12 Penalized 2 f12X=110sin2πx1+i=1D-1xi-121+sin23πxi+1+xn-12
[1 + sin⁡2(2πx i+1)] + ∑i=1 D u(x i, 5,100,4)
MN [−50, 50]D 0
F13 Alpine f13X=i=1Dxi·sinxi+0.1·xi MS [−10, 10]D 0

4.2. Parameters Settings

Parameter settings may have a great influence on the computational results. The ABC-SA mechanism has seven control parameters such as maximum iteration number (Max.iter), limit, population size (SN), p s1, p s2, ψ, and p 0. Maximum iteration number is the termination condition, and p 0 is the initial acceptance probability. First, Max.iter is set to 4,000, limit = 0.2 × D × SN, where D is the dimension of the problem [21], SN is taken to be 40 for 50D and 100D problems and 50 for 200D problems [40], and ψ is set to be a random real number within (0, 1.5) [14]. Then, preliminary experiments were conducted with appropriate combinations of the following parameter values to determine the best settings:

  • p 0 = 0.25, 0.20, 0.15, 0.10, and 0.05,

  • p s1 = 0.2, 0.4, and 0.6,

  • p s2 = 0.2, 0.4, and 0.6,

  • p s3 = 0.2, 0.4, and 0.6.

From the results of the pilot studies, p 0 = 0.10, p s1 = 0.20, p s2 = 0.60, and p s3 = 0.20 settings achieved the best results. Therefore, these parameter settings are used for further experiments.

4.3. Comparison with ABC Variants

In this section, aforementioned ABC-SA is implemented and evaluated by benchmarking with other well-known ABC variants including the original ABC [7], GABC [14], and IABC [13] on problems F1–F13.

The parameters of test algorithms are set to their original values given in their corresponding papers, except for the maximum number of function evaluations, population size, and limit, which are set to the same values for all ABC variants. ABC-SA, ABC, and GABC implement random initialization mechanisms whereas IABC employs a chaotic initialization as described in [13]. All algorithms have been simulated in MATLAB environment and executed on the same computer with Intel Xeon CPU (2.67 GHz) and 16 GB of memory.

The computational results are presented in Table 2 for 50D problems, Table 3 for 100D problems, and Table 4 for 200D problems. In Tables 24, results are given in terms of mean and standard deviation of the objective values due to the repetitive runs for the global best solutions. All algorithms were run 30 times with random seeds and the stopping criteria set to 4,000 iteration, which means that 320,000 functions evaluations for 50D and 100D problems and 400,000 function evaluations for 200D problems approximately. For a precise and pairwise comparison, statistical significances of the differences between the means of two algorithms are analyzed using t-tests where significance level is set to 0.05. In Tables 24, “+” in the columns next to competing algorithms shows that ABC-SA outperforms the competitor algorithm, “=” indicates that the difference between the ABC-SA and the compared algorithm is not statistically significant, and “−” depicts that the competitor algorithm is better than ABC-SA at a level of 0.05 significance.

Table 2.

Comparisons of ABC-SA and ABC variants on 50D problems.

Func. ABC-SA ABC GABC IABC
Mean Std. Dev. Mean Std. Dev. Sign Mean Std. Dev. Sign Mean Std. Dev. Sign
F1 3.10E + 01 1.18E + 01 3.84E + 01 1.07E + 01 = 3.32E + 01 6.71E + 00 = 3.25E + 01 1.75E + 01 =
F2 5.30E − 14 4.10E − 15 1.17E − 13 1.62E − 14 + 7.96E − 14 1.18E − 15 + 7.44E − 14 9.63E − 15 +
F3 0.00E + 00 0.00E + 00 2.02E − 11 7.11E − 12 + 8.53E − 14 4.16E − 14 + 8.53E − 14 3.58E − 14 +
F4 1.11E − 16 2.17E − 16 4.78E − 12 2.61E − 13 + 6.11E − 16 3.62E − 16 + 1.26E − 13 6.80E − 14 +
F5 0.00E + 00 0.00E + 00 3.84E − 14 1.90E − 14 + 8.53E − 15 9.94E − 15 + 1.71E − 14 8.99E − 15 +
F6 −2.09E + 04 2.51E − 15 −2.09E + 04 6.09E + 00 + −2.09E + 04 4.05E − 11 = −2.09E + 04 5.57E − 14 =
F7 −4.50E + 02 0.00E + 00 −4.50E + 02 6.36E − 14 = −4.50E + 02 2.84E − 14 = −4.50E + 02 4.92E − 14 =
F8 1.92E + 04 5.19E + 03 3.22E + 04 1.79E + 03 + 3.61E + 04 5.89E + 03 + 2.92E + 04 5.21E + 03 +
F9 3.98E + 02 3.11E + 00 5.03E + 02 3.96E + 00 + 4.13E + 02 2.45E + 01 + 4.23E + 02 2.48E + 01 +
F10 −3.30E + 02 0.00E + 00 −3.30E + 02 3.14E − 14 = −3.30E + 02 0.00E + 00 = −3.30E + 02 1.02E − 14 =
F11 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00 = 0.00E + 00 0.00E + 00 = 0.00E + 00 0.00E + 00 =
F12 4.69E − 15 1.90E − 16 5.90E − 14 9.82E − 15 + 1.93E − 14 4.77E − 15 + 8.77E − 15 7.43E − 15 +
F13 3.69E − 24 9.34E − 26 2.95E − 23 3.01E − 23 + 7.71E − 24 3.90E − 25 + 6.53E − 24 1.66E − 25 +

Table 3.

Comparisons of ABC-SA and ABC variants on 100D problems.

Func. ABC-SA ABC GABC IABC
Mean Std. Dev. Mean Std. Dev. Sign Mean Std. Dev. Sign Mean Std. Dev. Sign
F1 7.65E + 01 4.76E + 01 1.26E + 02 3.55E + 01 + 1.35E + 02 2.95E + 01 + 1.34E + 02 2.27E + 01 +
F2 6.16E − 13 9.63E − 14 6.97E − 07 3.25E − 07 + 4.56E − 12 1.18E − 12 + 7.06E − 11 1.31E − 14 +
F3 2.27E − 13 5.41E − 13 3.97E − 01 5.81E − 01 + 2.37E − 11 5.97E − 11 + 1.21E − 12 6.69E − 13 +
F4 1.58E − 14 2.64E − 14 1.30E − 04 7.13E − 04 = 7.73E − 10 3.90E − 09 = 3.54E − 12 9.27E − 12 +
F5 1.48E − 13 4.40E − 14 1.64E − 04 4.98E − 05 + 7.62E − 13 3.24E − 13 + 1.20E − 10 1.25E − 11 +
F6 −4.19E + 04 6.84E + 01 −4.06E + 04 2.61E + 02 + −4.16E + 04 1.53E + 02 + −4.19E + 04 6.24E + 01 =
F7 −4.50E + 02 1.90E − 14 −4.50E + 02 1.81E − 13 = −4.50E + 02 8.53E − 14 = −4.50E + 02 4.90E − 14 =
F8 8.85E + 04 1.00E + 03 1.61E + 05 2.43E + 04 + 1.60E + 05 1.15E + 04 + 1.95E + 05 7.92E + 03 +
F9 3.89E + 02 9.55E + 00 3.95E + 02 1.04E + 01 + 4.14E + 02 4.12E + 01 + 4.10E + 02 1.76E + 01 +
F10 −3.30E + 02 0.00E + 00 −3.22E + 02 9.41E − 04 + −3.30E + 02 7.20E − 07 = −3.30E + 02 5.68E − 14 =
F11 1.17E + 01 5.24E + 00 4.32E + 01 6.43E + 00 + 9.10E + 01 6.92E + 00 + 2.08E + 01 9.52E + 00 +
F12 9.12E − 13 1.04E − 15 5.54E − 11 7.27E − 13 + 8.71E − 13 6.90E − 14 1.15E − 12 3.91E − 13 +
F13 1.52E − 20 5.48E − 19 8.14E − 19 9.17E − 20 + 7.03E − 20 4.45E − 19 = 8.44E − 20 1.06E − 19 =

Table 4.

Comparisons of ABC-SA and ABC variants on 200D problems.

Func. ABC-SA ABC GABC IABC
Mean Std. Dev. Mean Std. Dev. Sign Mean Std. Dev. Sign Mean Std. Dev. Sign
F1 4.06E + 02 3.34E + 01 4.32E + 02 4.21E + 01 = 4.47E + 02 5.89E + 01 + 4.58E + 02 6.77E + 01 +
F2 1.85E − 05 3.30E − 05 4.44E − 02 2.13E − 02 + 6.68E − 05 1.65E − 05 + 1.47E − 04 3.69E − 05 +
F3 8.27E − 06 2.53E − 05 3.17E + 01 5.89E + 00 + 6.51E + 00 1.74E + 00 + 4.02E + 00 1.03E + 00 +
F4 6.05E − 10 2.26E − 09 5.75E − 04 2.58E − 03 = 1.79E − 08 5.29E − 08 = 6.05E − 08 4.68E − 08 +
F5 6.42E − 03 4.93E − 04 1.24E − 01 1.35E − 02 + 9.76E − 03 1.21E − 03 + 1.83E − 02 1.40E − 03 +
F6 −8.19E + 04 1.00E + 02 −7.63E + 04 7.61E + 02 + −7.89E + 04 4.20E + 02 + −8.37E + 04 3.68E + 02
F7 8.56E + 05 1.81E + 04 8.60E + 05 1.46E + 04 = 8.73E + 05 2.49E + 04 + 8.62E + 05 1.75E + 04 =
F8 2.89E + 06 9.10E + 04 2.91E + 06 2.07E + 05 = 2.93E + 06 3.13E + 05 = 2.98E + 06 3.90E + 05 =
F9 7.63E + 11 4.37E + 08 8.02E + 11 5.05E + 09 + 7.87E + 11 4.72E + 08 + 7.61E + 11 4.93E + 08
F10 3.74E + 03 4.19E + 01 3.78E + 03 5.94E + 01 + 3.76E + 03 5.55E + 01 + 3.76E + 03 4.60E + 01 =
F11 7.34E + 02 4.40E + 02 8.82E + 03 9.22E + 02 + 5.39E + 03 9.03E + 02 + 1.04E + 03 6.01E + 02 +
F12 3.79E − 11 4.90E − 13 1.91E − 09 9.41E − 14 + 4.88E − 11 2.25E − 12 + 6.22E − 11 3.89E − 12 +
F13 9.55E − 18 1.04E − 18 5.02E − 15 1.55E − 15 + 4.44E − 16 5.03E − 17 + 1.92E − 17 7.33E − 18 +

Tables 24 show that, according to pairwise t tests, ABC-SA obtains statistically better results on 25, 28, and 28 cases out of 39 comparisons for each of the 50D, 100D, and 200D problem types, respectively. Specifically, ABC-SA is inferior to IABC on F6 and F9 with 200D and GABC on F12 with 100D. There is no significant difference on the results that are obtained by ABC and ABC-SA on F1 (50D and 100D), F4 (100D and 200D), F7 (all dimensions), F8 200D, F10 50D, and F11 50D. Moreover, on F1 50D, F6 50D, F4 (100D and 200D), F7 (50D and 100D), F8 200D, F10 (50D and 100D), F11 50D, and F13 100D, GABC and ABC-SA perform statistically similar. Further, ABC-SA and IABC perform equally well, namely, on F1 50D, F6 (50D and 100D), F7 (all dimensions), F8 200D, F10 (all dimensions), F11 50D, and F13 100D. Standard deviation of the results also indicates that ABC-SA has a stable performance. According to the results of Tables 24, one can safely conclude that ABC-SA significantly surpasses ABC, GABC, and IABC on 50D, 100D, and 200D problems.

To vividly describe the effectiveness of ABC-SA framework, the convergence curves of some benchmark problems are given in Figure 3. According to the figure, ABC-SA shows better convergence behavior on the majority of test cases when compared to ABC, GABC, and IABC.

Figure 3.

Figure 3

Convergence curves for ABC-SA and ABC variants.

Furthermore, mean acceptance rate curves for solution acceptance rule in ABC-SA framework are given in Figure 4 and the acceptance rate is determined as follows:

acceptancerate=numberofacceptedworsesolutionstotalnumberofworsesolutions. (10)

The curves in Figure 4 clearly coincide with the acceptance probability curve given in Figure 2. Figure 4 also shows the nonlinear decreasing of acceptance rate throughout the search process.

Figure 4.

Figure 4

Mean acceptance rates.

4.4. Comparison with PSO and DE Variants

The performance of ABC-SA is also tested against novel and powerful variants of DE and PSO. The competitor algorithms are self-adapting DE (jDE) [41], adaptive DE with optional external archive (JADE) [42], self-adaptive DE (SaDE) [43], comprehensive learning PSO (CLPSO) [44], self-organizing hierarchical PSO with time-varying acceleration coefficients (HPSO-TVAC) [45], and fully informed particle swarm (FIPS) [46]. The results of these algorithms are taken directly from corresponding studies. The experimental results for 30D problems are shown in Tables 5 and 6 for DE and PSO variants, respectively. Some of the benchmarks problems are not included in the comparisons, since results on these problems were not reported in competitor studies. The previous parameter setting for ABC-SA is used, but this time maximum function evaluation number (Max.FE) is employed as the stopping criteria. Since the results of competitor algorithms are taken directly from corresponding studies, statistical significance tests could not be applied. Therefore, in this part of the analysis, mean and standard deviations of results are compared directly. In Tables 5 and 6, the winner algorithms are indicated in bold character according to the mean results of 30 independent runs. As can be seen from Tables 5 and 6, ABC-SA outperforms other algorithms on all cases, except in the case of F2, F4, and F12. SaDE performs better than the ABC-SA on F4 and HPSO-TVAC outperforms ABC-SA on only F2 and F12. ABC achieves better results on the majority of the instances in terms of robustness according to the standard deviations of the results. These results also indicate the effectiveness of ABC-SA when compared to other novel swarm based and evolutionary algorithms.

Table 5.

The comparisons of ABC-SA and DE variants on 30D problems.

Func. Max.FE ABC-SA jDE JADE SaDE
Mean Std. Dev. Mean Std. Dev. Mean Std. Dev. Mean Std. Dev.
F1 300,000 1.19E − 01 7.91E − 02 1.30E + 01 1.40E + 01 3.20E − 01 1.10E + 00 2.10E + 01 7.70E + 01
F2 50,000 3.01E − 09 7.44E − 10 2.37E − 04 7.10E − 05 3.35E − 09 2.84E − 09 3.81E − 06 8.26E − 07
F3 100,000 2.77E − 09 6.09E − 10 2.37E − 04 7.10E − 05 3.35E − 09 2.84E − 09 3.81E − 06 8.26E − 07
F4 50,000 7.23E − 09 4.11E − 09 7.29E − 06 1.05E − 05 1.57E − 08 1.09E − 07 2.52E − 09 1.24E − 09
F6 100,000 8.63E − 11 4.94E − 11 1.70E − 10 2.62E − 10 2.62E − 04 3.59E − 04 1.13E − 08 1.08E − 08
F11 10,000 4.41E + 00 3.90E + 00 6.13E + 02 1.72E + 02 5.62E + 00 1.87E + 00 5.07E + 01 1.34E + 01
F12 50,000 1.41E − 10 7.23E − 11 1.80E − 05 1.42E − 05 1.87E − 10 1.09E − 09 1.93E − 09 1.53E − 09
F13 300,000 4.99E − 11 2.49E − 10 6.08E − 10 8.36E − 10 2.78E − 05 8.43E − 06 2.94E − 06 3.47E − 06

Table 6.

The comparisons of ABC-SA and PSO variants on 30D problems.

Func. Max.FE ABC-SA FIPS HPSO-TVAC CLPSO
Mean Std. Dev. Mean Std. Dev. Mean Std. Dev. Mean Std. Dev.
F1 200,000 7.61E + 00 2.18E − 01 2.51E + 01 5.10E − 01 2.39E + 01 2.65E + 01 1.13E + 01 9.85E + 00
F2 200,000 4.01E − 13 1.05E − 13 2.33E − 07 7.19E − 08 7.29E − 14 3.00E − 14 3.66E − 07 7.57E − 08
F3 200,000 6.91E − 10 9.44E − 11 6.51E + 01 1.33E + 01 9.43E + 00 3.48E + 00 9.05E − 05 1.25E − 04
F4 200,000 8.89E − 12 6.56E − 13 9.01E − 12 1.84E − 11 9.75E − 03 8.33E − 03 9.02E − 09 8.57E − 09
F6 200,000 1.79E − 12 8.89E − 11 9.93E + 02 5.09E + 02 1.59E + 03 3.26E + 02 3.82E − 04 1.28E − 05
F11 200,000 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00 0.00E + 00
F12 200,000 5.03E − 23 1.44E − 24 2.70E − 14 1.57E − 14 2.79E − 28 2.18E − 28 1.25E − 12 9.45E − 12

5. Conclusion and Future Work

This paper presents a modified ABC algorithm, namely, the ABC-SA, enhanced with a solution acceptance rule and a probabilistic multisearch strategy. In ABC-SA, instead of a greedy selection, a new acceptance rule is presented, where a worse candidate solution has a probability to be accepted. Furthermore, to balance the diversification and intensification tendency of the algorithm, a probabilistic multisearch mechanism is employed. In the probabilistic multisearch, a search rule is selected among three alternatives according to their predetermined probabilities. The proposed algorithm is very effective as compared to other novel ABC variants and state-of-the-art algorithms. Several experimental studies are conducted and results show that ABC-SA outperforms all other competitor algorithms on the majority of the test cases. Future research will be along the line of implementing the ABC-SA to solve complex engineering problems.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  • 1.Golberg D. E. Genetic Algorithms in Search, Optimization, and Machine Learning. Vol. 1989. Addison-Wesley; 1989. [Google Scholar]
  • 2.Holland J. H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. MIT Press; 1992. [Google Scholar]
  • 3.Moscato P. C3P Report. 826. Caltech Concurrent Computation Program; 1989. On evolution, search, optimization, genetic algorithms and martial arts: towards memetic algorithms. [Google Scholar]
  • 4.Storn R., Price K. Differential evolution—a simple and efficient heuristic for global optimization over continuous spaces. Journal of Global Optimization. 1997;11(4):341–359. doi: 10.1023/a:1008202821328. [DOI] [Google Scholar]
  • 5.Colorni A., Dorigo M., Maniezzo V. Distributed optimization by ant colonies. Proceedings of the 1st European Conference on Artificial Life; 1991; Paris, France. pp. 134–142. [Google Scholar]
  • 6.Eberhart R. C., Kennedy J. New optimizer using particle swarm theory. Proceedings of the 6th International Symposium on Micro Machine and Human Science; October 1995; New York, NY, USA. pp. 39–43. [Google Scholar]
  • 7.Karaboga D., Basturk B. A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm. Journal of Global Optimization. 2007;39(3):459–471. doi: 10.1007/s10898-007-9149-x. [DOI] [Google Scholar]
  • 8.Yang X.-S., Deb S. Engineering optimisation by cuckoo search. International Journal of Mathematical Modelling and Numerical Optimisation. 2010;1(4):330–343. doi: 10.1504/ijmmno.2010.035430. [DOI] [Google Scholar]
  • 9.Fister I., Jr., Yang X.-S., Brest J. A comprehensive review of firefly algorithms. Swarm and Evolutionary Computation. 2013;13:34–46. doi: 10.1016/j.swevo.2013.06.001. [DOI] [Google Scholar]
  • 10.Karaboga D., Gorkemli B., Ozturk C., Karaboga N. A comprehensive survey: artificial bee colony (ABC) algorithm and applications. Artificial Intelligence Review. 2014;42(1):21–57. doi: 10.1007/s10462-012-9328-0. [DOI] [Google Scholar]
  • 11.Alatas B. Chaotic bee colony algorithms for global numerical optimization. Expert Systems with Applications. 2010;37(8):5682–5687. doi: 10.1016/j.eswa.2010.02.042. [DOI] [Google Scholar]
  • 12.Banharnsakun A., Sirinaovakul B., Achalakul T. Job shop scheduling with the best-so-far ABC. Engineering Applications of Artificial Intelligence. 2012;25(3):583–593. doi: 10.1016/j.engappai.2011.08.003. [DOI] [Google Scholar]
  • 13.Gao W., Liu S. Improved artificial bee colony algorithm for global optimization. Information Processing Letters. 2011;111(17):871–882. doi: 10.1016/j.ipl.2011.06.002. [DOI] [Google Scholar]
  • 14.Zhu G., Kwong S. Gbest-guided artificial bee colony algorithm for numerical function optimization. Applied Mathematics and Computation. 2010;217(7):3166–3173. doi: 10.1016/j.amc.2010.08.049. [DOI] [Google Scholar]
  • 15.Gao W., Liu S., Huang L. A global best artificial bee colony algorithm for global optimization. Journal of Computational and Applied Mathematics. 2012;236(11):2741–2753. doi: 10.1016/j.cam.2012.01.013. [DOI] [Google Scholar]
  • 16.Kang F., Li J., Ma Z. Rosenbrock artificial bee colony algorithm for accurate global optimization of numerical functions. Information Sciences. 2011;181(16):3508–3531. doi: 10.1016/j.ins.2011.04.024. [DOI] [Google Scholar]
  • 17.Akay B., Karaboga D. A modified artificial bee colony algorithm for real-parameter optimization. Information Sciences. 2012;192(1):120–142. doi: 10.1016/j.ins.2010.07.015. [DOI] [Google Scholar]
  • 18.Liao T., Aydin D., Stützle T. Artificial bee colonies for continuous optimization: experimental analysis and improvements. Swarm Intelligence. 2013;7(4):327–356. doi: 10.1007/s11721-013-0088-5. [DOI] [Google Scholar]
  • 19.Gao W.-F., Liu S.-Y., Huang L.-L. Enhancing artificial bee colony algorithm using more information-based search equations. Information Sciences. 2014;270(1):112–133. doi: 10.1016/j.ins.2014.02.104. [DOI] [Google Scholar]
  • 20.Qiu J., Wang J., Yang D., Xie J. An artificial bee colony algorithm with modified search strategies for global numerical optimization. Journal of Theoretical & Applied Information Technology. 2013;48(1):293–302. [Google Scholar]
  • 21.Banitalebi A., Aziz M. I. A., Bahar A., Aziz Z. A. Enhanced compact artificial bee colony. Information Sciences. 2015;298:491–511. doi: 10.1016/j.ins.2014.12.015. [DOI] [Google Scholar]
  • 22.Wang H., Wu Z., Rahnamayan S., Sun H., Liu Y., Pan J.-S. Multi-strategy ensemble artificial bee colony algorithm. Information Sciences. 2014;279:587–603. doi: 10.1016/j.ins.2014.04.013. [DOI] [Google Scholar]
  • 23.Gao W., Chan F. T., Huang L., Liu S. Bare bones artificial bee colony algorithm with parameter adaptation and fitness-based neighborhood. Information Sciences. 2015;316:180–200. doi: 10.1016/j.ins.2015.04.006. [DOI] [Google Scholar]
  • 24.Ma L., Hu K., Zhu Y., Chen H. A hybrid artificial bee colony optimizer by combining with life-cycle, Powell's search and crossover. Applied Mathematics and Computation. 2015;252:133–154. doi: 10.1016/j.amc.2014.11.104. [DOI] [Google Scholar]
  • 25.Pan Q.-K., Tasgetiren M. F., Suganthan P. N., Chua T. J. A discrete artificial bee colony algorithm for the lot-streaming flow shop scheduling problem. Information Sciences. 2011;181(12):2455–2468. doi: 10.1016/j.ins.2009.12.025. [DOI] [Google Scholar]
  • 26.Li J.-Q., Pan Q.-K., Gao K.-Z. Pareto-based discrete artificial bee colony algorithm for multi-objective flexible job shop scheduling problems. The International Journal of Advanced Manufacturing Technology. 2011;55(9–12):1159–1169. doi: 10.1007/s00170-010-3140-2. [DOI] [Google Scholar]
  • 27.Szeto W. Y., Wu Y., Ho S. C. An artificial bee colony algorithm for the capacitated vehicle routing problem. European Journal of Operational Research. 2011;215(1):126–135. doi: 10.1016/j.ejor.2011.06.006. [DOI] [Google Scholar]
  • 28.Yurtkuran A., Emel E. A modified artificial bee colony algorithm for p-center problems. The Scientific World Journal. 2014;2014:9. doi: 10.1155/2014/824196.824196 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Ma M., Liang J., Guo M., Fan Y., Yin Y. SAR image segmentation based on artificial bee colony algorithm. Applied Soft Computing. 2011;11(8):5205–5214. doi: 10.1016/j.asoc.2011.05.039. [DOI] [Google Scholar]
  • 30.Karaboga D., Okdem S., Ozturk C. Cluster based wireless sensor network routing using artificial bee colony algorithm. Wireless Networks. 2012;18(7):847–860. doi: 10.1007/s11276-012-0438-z. [DOI] [Google Scholar]
  • 31.Singh A. An artificial bee colony algorithm for the leaf-constrained minimum spanning tree problem. Applied Soft Computing. 2009;9(2):625–631. doi: 10.1016/j.asoc.2008.09.001. [DOI] [Google Scholar]
  • 32.Karaboga D., Ozturk C. A novel clustering approach: Artificial Bee Colony (ABC) algorithm. Applied Soft Computing Journal. 2011;11(1):652–657. doi: 10.1016/j.asoc.2009.12.025. [DOI] [Google Scholar]
  • 33.De Oliveira I. M. S., Schirru R. Swarm intelligence of artificial bees applied to in-core fuel management optimization. Annals of Nuclear Energy. 2011;38(5):1039–1045. doi: 10.1016/j.anucene.2011.01.009. [DOI] [Google Scholar]
  • 34.Gao W.-F., Liu S.-Y., Jiang F. An improved artificial bee colony algorithm for directing orbits of chaotic systems. Applied Mathematics and Computation. 2011;218(7):3868–3879. doi: 10.1016/j.amc.2011.09.034. [DOI] [Google Scholar]
  • 35.Hong W.-C. Electric load forecasting by seasonal recurrent SVR (support vector regression) with chaotic artificial bee colony algorithm. Energy. 2011;36(9):5568–5578. doi: 10.1016/j.energy.2011.07.015. [DOI] [Google Scholar]
  • 36.Kumar S. K., Tiwari M. K., Babiceanu R. F. Minimisation of supply chain cost with embedded risk using computational intelligence approaches. International Journal of Production Research. 2010;48(13):3717–3739. doi: 10.1080/00207540902893425. [DOI] [Google Scholar]
  • 37.Gao W.-F., Liu S.-Y. A modified artificial bee colony algorithm. Computers & Operations Research. 2012;39(3):687–697. doi: 10.1016/j.cor.2011.06.007. [DOI] [Google Scholar]
  • 38.Gao W.-F., Liu S.-Y., Huang L.-L. A novel artificial bee colony algorithm with Powell's method. Applied Soft Computing Journal. 2013;13(9):3763–3775. doi: 10.1016/j.asoc.2013.05.012. [DOI] [Google Scholar]
  • 39.Karaboga D., Akay B. A comparative study of artificial bee colony algorithm. Applied Mathematics and Computation. 2009;214(1):108–132. doi: 10.1016/j.amc.2009.03.090. [DOI] [Google Scholar]
  • 40.Karaboga D., Basturk B. On the performance of artificial bee colony (ABC) algorithm. Applied Soft Computing. 2008;8(1):687–697. doi: 10.1016/j.asoc.2007.05.007. [DOI] [Google Scholar]
  • 41.Brest J., Greiner S., Bošković B., Mernik M., Zumer V. Self-adapting control parameters in differential evolution: a comparative study on numerical benchmark problems. IEEE Transactions on Evolutionary Computation. 2006;10(6):646–657. doi: 10.1109/tevc.2006.872133. [DOI] [Google Scholar]
  • 42.Zhang J., Sanderson A. C. JADE: adaptive differential evolution with optional external archive. IEEE Transactions on Evolutionary Computation. 2009;13(5):945–958. doi: 10.1109/TEVC.2009.2014613. [DOI] [Google Scholar]
  • 43.Qin A. K., Huang V. L., Suganthan P. N. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Transactions on Evolutionary Computation. 2009;13(2):398–417. doi: 10.1109/tevc.2008.927706. [DOI] [Google Scholar]
  • 44.Liang J. J., Qin A. K., Suganthan P. N., Baskar S. Comprehensive learning particle swarm optimizer for global optimization of multimodal functions. IEEE Transactions on Evolutionary Computation. 2006;10(3):281–295. doi: 10.1109/TEVC.2005.857610. [DOI] [Google Scholar]
  • 45.Ratnaweera A., Halgamuge S. K., Watson H. C. Self-organizing hierarchical particle swarm optimizer with time-varying acceleration coefficients. IEEE Transactions on Evolutionary Computation. 2004;8(3):240–255. doi: 10.1109/tevc.2004.826071. [DOI] [Google Scholar]
  • 46.Mendes R., Kennedy J., Neves J. The fully informed particle swarm: simpler, maybe better. IEEE Transactions on Evolutionary Computation. 2004;8(3):204–210. doi: 10.1109/tevc.2004.826074. [DOI] [Google Scholar]

Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES