Skip to main content
Biomimetics logoLink to Biomimetics
. 2024 Mar 19;9(3):186. doi: 10.3390/biomimetics9030186

A Novel Artificial Electric Field Algorithm for Solving Global Optimization and Real-World Engineering Problems

Abdelazim G Hussien 1,2,*, Adrian Pop 1, Sumit Kumar 3, Fatma A Hashim 4,5, Gang Hu 6
Editors: Yongquan Zhou, Shuai Li
PMCID: PMC10968472  PMID: 38534871

Abstract

The Artificial Electric Field Algorithm (AEFA) stands out as a physics-inspired metaheuristic, drawing inspiration from Coulomb’s law and electrostatic force; however, while AEFA has demonstrated efficacy, it can face challenges such as convergence issues and suboptimal solutions, especially in high-dimensional problems. To overcome these challenges, this paper introduces a modified version of AEFA, named mAEFA, which leverages the capabilities of Lévy flights, simulated annealing, and the Adaptive s-best Mutation and Natural Survivor Method (NSM) mechanisms. While Lévy flights enhance exploration potential and simulated annealing improves search exploitation, the Adaptive s-best Mutation and Natural Survivor Method (NSM) mechanisms are employed to add more diversity. The integration of these mechanisms in AEFA aims to expand its search space, enhance exploration potential, avoid local optima, and achieve improved performance, robustness, and a more equitable equilibrium between local intensification and global diversification. In this study, a comprehensive assessment of mAEFA is carried out, employing a combination of quantitative and qualitative measures, on a diverse range of 29 intricate CEC’17 constraint benchmarks that exhibit different characteristics. The practical compatibility of the proposed mAEFA is evaluated on five engineering benchmark problems derived from the civil, mechanical, and industrial engineering domains. Results from the mAEFA algorithm are compared with those from seven recently introduced metaheuristic algorithms using widely adopted statistical metrics. The mAEFA algorithm outperforms the LCA algorithm in all 29 CEC’17 test functions with 100% superiority and shows better results than SAO, GOA, CHIO, PSO, GSA, and AEFA in 96.6%, 96.6%, 93.1%, 86.2%, 82.8%, and 58.6% of test cases, respectively. In three out of five engineering design problems, mAEFA outperforms all the compared algorithms, securing second place in the remaining two problems. Results across all optimization problems highlight the effectiveness and robustness of mAEFA compared to baseline metaheuristics. The suggested enhancements in AEFA have proven effective, establishing competitiveness in diverse optimization problems.

Keywords: artificial electric field algorithm, AEFA, escaping local operator, global optimization

1. Introduction

Metaheuristic optimization algorithms have recently garnered significant attention and become a focal point for many computer science and engineering researchers. They are increasingly prevalent across various domains, including numerical optimization [1,2], cloud computing [3], neural networks [4], feature selection [5,6], classification [7], clustering [8], predicting chemical activities [9], text document clustering [10], and face detection & recognition [11].

Examples of metaheuristic algorithms include Genetic Algorithms (GA) [12], CMAES) [13], Cuckoo Search Optimization (CS) [14], the Harmony Search (HS) Algorithm [15], Artificial Bee Colony (ABC) [16], the Krill Herd Algorithm (KHA) [17], the Ant Lion Optimizer (ALO) [18], Moth-flame Optimization (MFO) [19], the Crow search Algorithm (CSA) [20], the Whale Optimizer Algorithm (WOA) [21], the Lightning Search Algorithm (LSA) [22], the Salp Swarm Algorithm (SSA) [23], Harris Hawks Optimization (HHO) [24], Virus Colony Search (VCS) [25], and the Snake Optimizer (SO) [26].

Metaheuristics (MHs) typically adopt strategies from swarm intelligence (SI), evolutionary algorithms (EA), physics-based principles, and human-based concepts. EAs are inspired by biological behaviors, while SI draws on food foraging, mutation, territorial fights, and mating behaviors. Physics laws and human behavior also influence optimizer development [27]. Consequently, these refined algorithms are increasingly applied to effectively address challenges in engineering design optimization, yielding more promising solutions.

Despite the wide range of problem-solving capabilities offered by metaheuristics, the “No free Lunch” hypothesis suggests that no single MH can provide the best solution for every complex problem [28]. Each optimization problem requires a specific strategy to be effectively addressed [29].

While solving complex engineering problems, metaheuristics may have drawbacks such as slow convergence and being trapped in local search domains, resulting in higher computational costs [30]. To address these limitations, researchers have devised hybridized, modified, and enhanced MHs that incorporate more beneficial attributes. A few examples include the Hybrid Grey Wolf and Crow Search [31], Hybrid Heat Transfer and Passing Vehicle Search [32], Hybrid Artificial Hummingbird-Simulated Annealing [33], Modified Symbiotic Organisms Search [34], Modified Marine Predator Algorithm [35], Improved Ant Colony Optimization [36], and Improved Salp Swarm Algorithm [37]. To create effective MHs, a balance between global diversification and local intensification is crucial. Both exploration and exploitation phases are important in finding superior solutions and achieving results in the least amount of time. Despite numerous hybrid MHs being implemented in engineering design optimization over the last few decades, the quest for even more potent methods is ongoing. This field continues to evolve and presents new challenges for researchers to address.

The Artificial Electric Field Algorithm (AEFA) is a recent population-based algorithm proposed by Anita and Yadav [38] that is inspired by the electrostatic force of Coulomb’s law. AEFA has been successfully employed in many applications such as photovoltaics [39], fuel cell estimation [40], the economic load dispatch problem [41], non-linear system modeling [42], the quadratic assignment problem [43], soil shear [44], and wind turbine allocation [45].

In addition, many researchers have introduced different variants of AEFA algorithms. For example, Izci et al. [46] hybridized AEFA with the Nelder–Mead simplex. Likewise, Petwal and Rani [47] developed a multiobjective version of AEFA. They strengthened AEFA by adding polynomial mutation and bounded exponential crossover operators. A novel set of velocity and position bounds has been proposed by Anita et al. [48] to address engineering problems. Anita and Yadav [49] developed a discrete AEFA for high-order graph matching. Furthermore, a detailed study of stability and exploratory abilities was presented in [50].

Kahraman et al. [51] introduce a novel approach called the Natural Survivor Method (NSM), devised as a model for population management mirroring natural processes, taking into account environmental variables and analytical relationships. Within the NSM framework, scores reflecting individuals’ adaptability to their natural surroundings are computed to identify survivors. Additionally, the update mechanism in this proposed method is constructed based on NSM scores rather than traditional fitness values.

Although AEFA has demonstrated effectiveness in addressing intricate engineering problems, it exhibits a notable susceptibility to converging toward suboptimal solutions. The iterative process of population size selection poses challenges, as optimal sizes vary across different problems. Like any other algorithm, AEFA demonstrates a moderate convergence rate, potentially requiring more time than alternative algorithms to reach an optimal solution. Moreover, the efficacy of AEFA is notably contingent on parameter choices, including the number of particles, electric charge, and inertia. AEFA performance may be limited on large-scale problems due to high computational complexity. Moreover, AEFA may face premature convergence issues, particularly in multimodal optimization problems. Furthermore, if the algorithm focuses too much on exploration, it may sacrifice exploitation, and vice versa, thus striking a balance between global diversification and local intensification is crucial for optimizing performance. Hybridizing AEFA with other algorithms or the inclusion of performance improvement strategies can result in faster convergence, higher precision, and greater robustness [52]. As AEFA is a relatively new optimizer, there is great potential for uncovering new improvements that can further enhance its efficiency and performance.

The objective of this research was to enhance the performance of the existing AEFA algorithm by addressing its limitations, with a focus on achieving faster convergence, increased accuracy, and improved robustness. To achieve these goals, a modified version of AEFA named modified AEFA (mAEFA) is introduced based on two techniques: Lévy flights and Simulated Annealing (SA). The first operator is used to increase randomness and the second is used to increase exploitation.

This study offers the following significant contributions:

  • The Lévy flight distribution mechanism is incorporated to increase the search space of AEFA and improve the exploration potential of the algorithm. This integration aims to prevent the algorithm from getting trapped in local optima, thereby contributing to an overall improvement in performance;

  • A simulated annealing mechanism is integrated with AEFA to improve search exploitation by allowing the algorithm to accept solutions that are worse than the current best solution with a certain probability, which helps to avoid getting stuck in local optima and explore other areas of the search space. This can lead to finding better solutions that may have been missed otherwise. Therefore, hybridizing simulated annealing into an algorithm can lead to improved performance and robustness;

  • A thorough evaluation of mAEFA performance, utilizing both quantitative and qualitative methods, is carried out on a variety of complex CEC 2017 constraint benchmarks with varying characteristics;

  • The behavior of mAEFA is evaluated and compared to 12 prevalent MH approaches on five engineering benchmark problems drawn from diverse fields.

The rest of this paper is structured as follows: Section 2 discusses recent work on metaheuristics and Section 3 introduces the original AEFA. Section 4 presents the suggested modified version of AEFA. Section 5 details the results of the conducted statistical tests and respective discussion, and Section 6 concludes the paper, including outlining future directions.

2. Related Work

The Lévy flight (LF) mechanism is a popular choice for researchers aiming to bolster the optimization efficacy of algorithms, spanning diverse research domains such as control system design [53], wind speed forecasting [54], and high-dimensional optimization problems [55]. Characterized by random steps, the application of LF enables optimization algorithms to navigate the vicinity of existing solutions, while intermittent significant leaps mitigate the risk of entrapment within local minima. For instance, Zhang et al. [56] recently demonstrated that LF enhances particle diversity within PSO, thereby refining the accuracy of lithium-ion battery State-of-Health prediction. Similarly, Hussien et al. [57] noted the efficacy of LF-based Transient Search Optimization in augmenting the transient response of terminal voltages within islanded microgrids. Furthermore, Barua et al. [58] amalgamated the LF strategy with the Arithmetic Optimization Algorithm to address various engineering optimization challenges. Their findings underscore the hybrid algorithm’s superior optimization capabilities, requiring fewer evaluations and surpassing several established algorithms in performance benchmarks.

Simulated annealing (SA) is an MH extensively employed to enhance the search capability of optimization algorithms [59]. Distinguished by its probabilistic jump feature, inspired by the physical process of annealing solids, SA effectively mitigates the risk of stagnation in local optima, facilitating the attainment of global optima. For example, Xu et al. [60] endeavored to enhance the Whale Optimization Algorithm by integrating it with SA, yielding superior optimization performance and stability across multiple dimensional problems compared to alternative algorithms such as WOA, GWO, and PSO. Similarly, Fontes et al. [61] employed a hybrid of SA and PSO for tackling the job shop scheduling problem, achieving high-quality solutions within reasonable computation times. In another study, Sajjad et al. [62] noted that the incorporation of SA notably enhanced the convergence of the multiobjective Grasshopper Algorithm when applied to problem solving in IoT applications.

The performance of population-based metaheuristic algorithms is significantly influenced by the size, nature, and diversity of the initial population, as well as the number of iterations. Agushaka et al. [63] observed that BA performed better with larger population sizes, while GWO, WOA, BOA, MS, and LSHADE-cnEpSin benefited from more iterations. Conversely, MFO, LSHADE, EHO, and HHO showed optimal performance with balanced population sizes and iterations. Kazimipour et al. [64] categorize population initialization techniques based on randomness, compositionality, and generality aspects. While random number generators are commonly used for initialization, their limitations have prompted researchers to explore alternative distributions, such as Latin Hypercube Sampling (LHS), to enhance efficiency [65]. Quasirandom sequences like Halton, Sobol, and Faure have demonstrated effectiveness in uniformly covering the search space, as seen in PSO and Global Best PSO. Chaotic sequences have also shown promise in initializing populations in CS and BFO, despite concerns about computational complexity. Hybridization with other metaheuristic algorithms, like the Greedy Randomized Adaptive Search Procedure (GRASP) and Metropolis–Hastings (MH), has the potential for improving initialization schemes, though scalability and time complexity are challenges. Leveraging ad hoc knowledge of the problem domain, as seen in the Bat Algorithm (BA), and innovative approaches such as patch environments and quasi-opposition-based learning in CS, aim to improve algorithm efficacy. Each method presents unique advantages and challenges, highlighting the importance of selecting suitable initialization strategies tailored to the problem domain and available computational resources [63,64,65].

The MH process involves two main steps: selection of solution candidates from the population and determination of the search direction. Selection methods are vital in MHs and are categorized into three types: non-deterministic, deterministic, and probabilistic [66]. Non-deterministic methods randomly select solution candidates to enhance search diversity. Deterministic methods consider fitness values, selecting the best candidates to guide the search towards successful positions. Probabilistic methods combine characteristics of both elitist and random methods, with examples including the roulette wheel and tournament methods. New selection methods, such as Fitness–Distance balance (FDB) [66], Fitness–Distance-Constraint (FDC) [67], Adaptive FDB [68], and Dynamic FDB [69], address specific challenges in MH search.

3. Artificial Electric Field Algorithm

The Artificial Electric Field Algorithm (AEFA) is a recent population-based metaheuristic algorithm that mimics Coulomb’s electrostatic force law [38]. Coulomb’s electrostatic law states that the force between any two charged particles is inversely proportional to the distance squared between the particles and directly proportional to their charge product.

Particles can migrate/move in the search space. With the help of the electrostatic force, particles are able to interact with each other (either repulsion or attraction).

In AEFA, one only considers the attraction force, which means that particle with the highest charge will absorb other particles with lower charges. The objective function in AEFA is the candidate solution (particle) charges and the agent fitness value. To mathematically model the electric force, the following steps are used.

The position of i-th agent (Xi) can be given as Xi=(xi1,xi2,...,xid), where xid is the position of i-th agent at d-dimension.

The position of the best agent (solution) obtained by any particle, i, at any time, t, denoted as Pi(t), is given by the following Equation (1)

Pid(t+1)=Pid(t)f(Pi(t))<f(Xi(t+1))Xid(t+1)iff(Xi(t+1))<f(Pi(t)) (1)

where Pi is the position of the personal best particle. Xbest is the position of the best charged particle determined based on fitness. The best objective value via all particles is given as Pbest=Xbest.

To calculate the total force that is acting on agent i at time t, the following equation is used

Fijd=j=1,jiNrand()Fijd (2)

where rand() is a random number between 0 and 1, N refers to the number of search agents, and Fi refers to resultant force that affect on the i-th agent.

The acceleration of the i-th agent at d dimension can be calculated using Newton’s second law of motion as below

aid(t)=QiEid(t)Mi(t) (3)

Qi and Ei represent the charges and electric field of the ith particle, and Mi(t) refers to ith particle mass. The velocity and position of particles can be updated using the two following equations:

Vid(t+1)=rand()×Vid(t)+aid(t) (4)
Xid(t+1)=Xid(t)+Vid(t+1) (5)

where rand() is a random number belongs to the interval [0,1]. The particles’ charge can be obtained using fitness functions and assuming that each particle has the same charge.

4. Proposed Algorithm

Despite the success and the power of AEFA, it has many drawbacks, as it may become stuck in local optima or have slow convergence, especially in high-dimensional and complex problems. This paper suggests an improved version of AEFA called modified AEFA (mAEFA) that tries to overcome the limitation of the original algorithm using two techniques, Lévy flights and Simulated Annealing, in each updating process. The flow chart is given in Figure 1.

Figure 1.

Figure 1

mAEFA flow chart.

4.1. Lévy Flights

Randomization has a great emphasis and effect in algorithms as it prevents them from falling in local optimal regions/areas. The Lévy flight mechanism has been successfully implemented and employed in many algorithms. In this version, we used Lévy flight to generate a random vector instead of the vector given in Equation (2) and use it to calculate the force as follows:

ran=2×randLévy(D) (6)
Fi=j=1,jiNranFijd (7)

where rand is a random number between 0 and 1 and D is the problem dimension. This will provide the search agent with a high opportunity to cover the entire search space.

4.2. Simulated Annealing

Simulated Annealing (SA) is a single-solution metaheuristic algorithm developed by Kirkpatrick et al. [70] that can be considered as an extension of the hill climbing technique. SA use a certain probability in accepting a worse solution to be able to avoid local optima. SA use Boltzmann probability, P=eθ/T, where θ is the difference between solutions and T refers to a temperature parameter.

SA is employed here to enhance the exploitation abilities of the AEFA. In this paper, we used SA to generate a new solution and update the current agent if the new solution is better than the current one. Here, T is equal to 2×|N|, where N is the number of agents.

4.3. Adaptive S-Best Mutation

The equilibrium between exploration and exploitation stands out as a crucial aspect determining an algorithm’s effectiveness in search capability. In the original AEFA algorithm, the absence of a gradual transition from exploration to exploitation results in new solutions being generated from the current solution without any discernible bias. To address this limitation, we introduce the s-best mutation approach, where S represents the top-ranked solution based on its fitness. This mutation strategy randomly selects one solution from the top S candidates to generate new solutions. The range of potential solutions is controlled by variables S1,S2,,SN, and S0,1. Through this approach, each new solution, indexed as i, is derived by either mutating the current solution or randomly mutating any solution selected from the top S candidates. The parameter S decreases non-linearly over time, enabling a broad exploration of the search space initially, gradually shifting towards a focus on solutions closer to the global optimum. This straightforward approach enhances diversity at the onset of generations, facilitating effective exploration, while transitioning towards exploitation as generations progress. During the exploitation phase, the algorithm concentrates on a narrower range of solutions surrounding the global best. The s-best mutation scheme operates as follows:

y=Xsbest+F×(Xr1Xr2) (8)

where Xbest signifies the agent derived from the top S×N agents, differing from the deterministic best solution. Additionally, the individuals Xr1 and Xr2 represent two distinct agents randomly selected from the entire generation, meeting the criteria that r1r2, and neither is equal to the current agent. F denotes a constant and real-value factor typically chosen to be between 0 and 1, and r is linearly decreased as specified as follows:

St=1(11N)×t1T1 (9)

where T signifies the maximum iteration number, t represents the current time, and N denotes the size of the population. Therefore, larger values of S are associated with earlier solutions, facilitating effective exploration. Conversely, as generations progress, S diminishes to foster exploitation enhancements.

4.4. NSM Operator

The survivor-selection process based on NSM scores operates according to a model that analytically connects the three described criteria. According to this model, the solution candidate that contributes more to the diversity of the mating pool and population compared to its competitors, while also achieving a better fitness value for the objective function, will survive. Within the NSM framework, the survivor is determined as the one that obtains the best score according to these three criteria.

5. Experimental Results and Discussion

To evaluate the proposed mAEFA method, this study employs the CEC’17 test suite, consisting of 30 functions commonly used to assess MHs [71]. Five engineering problems are also used for evaluation. The objective is to assess the search capability and convergence behavior of the proposed method. Given the stochastic nature of MHs, the experiments are conducted 30 times to account for randomness and observe result variations across runs.

Several newly developed algorithms are employed for comparison with the suggested mAEFA. These algorithms consist of the Coronavirus Herd Immunity Optimizer (CHIO) [72], Gravitational Search Algorithm (GSA) [73], Smell Agent Optimization (SAO) [74], Grasshopper Optimisation Algorithm (GOA) [75], Particle Swarm Optimization (PSO) [76], Liver Cancer Algorithm (LCA) [77], and basic AEFA [38]. For fairness in comparing algorithms, all were executed on the same hardware system to solve the CEC’17 test suite. The experiments were standardized with a maximum of 500 iterations. MATLAB 2021 served as the programming language for all algorithms, operating on a 64-bit Windows 8.1 system with an i7 Core and 8 GB RAM.

5.1. Experimental Series 1: CEC’17

5.1.1. Comparing mAEFA with State-of-Art Algorithms

CEC’17 includes functions representing diverse, complex, and dynamic optimization problems, commonly used for evaluating algorithm effectiveness. This study assesses the mAEFA algorithm using these functions, gaining insights into its effectiveness. Furthermore, it’s worth noting that each experiment is constrained by a maximum evaluation limit of 50,000 iterations. The setup entails employing 30 agents within a 30-dimensional space, as delineated in Table 1. Throughout the entirety of the experimentation, MATLAB2021 is utilized as the programming language, operating on a 64-bit Windows 8.1 platform. Table 2 presents the 30 functions, classified into four sets: F1 to F3 represents the unimodal set, F4 to F10 represents the multimodal set, F11 to F20 represents the hybrid set, and F21 to F30 represents the composition set. F2 is excluded from the evaluation process, leaving a total of 29 functions utilized to evaluate the mAEFA algorithm and other algorithms. The search range for all test functions, as indicated in Table 2, spans from −100 to 100, with a dimensionality of 30. In order to test and validate the enhanced version of mAEFA, we conducted a comparison with the original AEFA as well as six other distinct algorithms: CHIO, GSA, SAO, GOA, PSO, and LCA. The parameter seetinbg of each algorithm is given in Table 3.

Table 1.

Experiments parameters settings.

No. Parameter Name Value
1 Population Size 30
2 Dim 30
3 Max number of evaluation 50000
Table 2.

Benchmark functions.

No. Types Name Opt
F1 Unimodal Shifted and Rotated Bent Cugar Function 100
F2 Shifted and Rotated Sum of Different Power Function 200
F3 Shifted and Rotated Zakharov Function 300
F4 Multimodal Shifted and Rotated Rosenbrock’s Function 400
F5 Shifted and Rotated Rastrigin’s Function 500
F6 Shifted and Rotated Expanded Scaffer’s F6 Fuction 600
F7 Shifted and Rotated Lunacek Bi-Rastrigin Function 700
F8 Shifted and Rotated Non-Continuous Rastrigin’s Function 800
F9 Shifted and Rotated Lécy Function 900
F10 Shifted and Rotated Schwefel’s Function 1000
F11 Hybrid Hybrid function 1 (N = 3) 1100
F12 Hybrid function 2 (N = 3) 1200
F13 Hybrid function 3 (N = 3) 1300
F14 Hybrid function 4 (N = 4) 1400
F15 Hybrid function 5 (N = 4) 1500
F16 Hybrid function 6 (N = 4) 1600
F17 Hybrid function 6 (N = 5) 1700
F18 Hybrid function 6 (N = 5) 1800
F19 Hybrid function 6 (N = 5) 1900
F20 Hybrid function 6 (N = 6) 2000
F21 Composition Composition Function 1 (N = 3) 2100
F22 Composition Function 2 (N = 3) 2200
F23 Composition Function 3 (N = 4) 2300
F24 Composition function 4 (N = 4) 2400
F25 Composition function 5 (N = 5) 2500
F26 Composition function 6 (N = 5) 2600
F27 Composition function 7 (N = 6) 2700
F28 Composition function 8 (N = 6) 2800
F29 Composition function 9 (N = 3) 2900
F30 Composition function 10 (N = 3) 3000

Search range: [−100, 100] Dimension: 30.

Table 3.

Metaheuristic algorithms parameters settings.

Alg. Parameter Value
CHIO BRr 0.07
MaxAge 100
GSA G0 100
α 20
k ∈ [n,2]
SAO A [2, 0]
fc 2
GOA cmax 1
CMAES αmu 2
AEFA K(t) Initialized at first
LCA - -

The evaluation was based on the calculation of the average and standard deviation.

The comparison results are presented in Table 4. It is evident from the table that the developed algorithm (mAEFA) ranks first in 20 out of the 29 functions. Additionally, mAEFA obtains the second-best results in four functions and the third-best result in three functions. In contrast, GSA ranks first in only five functions, while AEFA achieves the best results in only four functions. Among the majority of the CEC’17 29 test functions, the algorithms that exhibit the lowest performance are SAO, GOA, LCA, and CHIO.

Table 4.

Statistical outcomes comparing mAEFA against various other metaheuristics on CEC’17 benchmark functions.

Fun mAEFA AEFA LCA CHIO GOA GSA CMA-ES SAO
F1 min 1.51×102 2.80 ×102 4.89 ×1010 3.97 ×106 4.48 ×1010 7.63 ×102 6.01 ×1010 4.94 ×1010
max 1.04 ×104 5.61 ×103 1.63 ×1011 2.40 ×108 9.75 ×1010 3.88 ×108 1.00 ×1011 7.72 ×1010
mean 2.73 ×103 2.12 ×103 9.03 ×1010 9.94 ×107 6.97 ×1010 4.56 ×107 8.25 ×1010 6.75 ×1010
std 2.65 ×103 1.32 ×103 3.07 ×1010 7.16 ×107 1.43 ×1010 8.60 ×107 1.06 ×1010 7.72 ×109
rank 2 1 8 4 6 3 7 5
F3 min 7.37 ×104 7.30 ×104 1.16 ×105 9.91 ×104 1.04 ×105 7.57 ×104 1.15 ×105 8.48 ×104
max 8.87 ×104 1.13 ×105 5.69 ×109 1.59 ×105 6.43 ×105 1.18 ×105 2.77 ×105 1.76 ×105
mean 8.30 ×104 8.91 ×104 2.27 ×108 1.32 ×105 3.86 ×105 9.47 ×104 2.07 ×105 1.20 ×105
std 4.25 ×103 9.26 ×103 1.03 ×109 1.81 ×104 1.10 ×105 9.40 ×103 3.18 ×104 2.37 ×104
rank 1 2 8 5 7 3 6 4
F4 min 4.91 ×102 5.09 ×102 1.44 ×104 5.08 ×102 6.06 ×103 5.45 ×102 9.50 ×103 7.33 ×103
max 5.86 ×102 5.81 ×102 7.34 ×104 5.93 ×102 3.28 ×104 9.26 ×102 2.97 ×104 2.56 ×104
mean 5.27 ×102 5.34 ×102 3.60 ×104 5.50 ×102 2.09 ×104 6.54 ×102 1.98 ×104 1.85 ×104
std 1.76 ×101 1.59 ×101 1.68 ×104 2.35 ×101 6.73 ×103 9.02 ×101 5.01 ×103 3.55 ×103
rank 1 2 8 3 7 4 6 5
F5 min 5.17 ×102 5.52 ×102 9.37 ×102 6.26 ×102 8.98 ×102 6.80 ×102 9.82 ×102 8.88 ×102
max 5.41 ×102 6.15 ×102 1.27 ×103 6.87 ×102 1.14 ×103 7.94 ×102 1.10 ×103 1.05 ×103
mean 5.28 ×102 5.85 ×102 1.07 ×103 6.56 ×102 1.02 ×103 7.42 ×102 1.03 ×103 9.78 ×102
std 6.34 ×100 1.51 ×101 9.93 ×101 1.54 ×101 6.40 ×101 2.75 ×101 3.28 ×101 3.53 ×101
rank 1 2 8 3 6 4 7 5
F6 min 6.00 ×102 6.08 ×102 6.88 ×102 6.04 ×102 6.88 ×102 6.50 ×102 6.84 ×102 6.74 ×102
max 6.01 ×102 6.38 ×102 7.53 ×102 6.12 ×102 7.50 ×102 6.65 ×102 7.21 ×102 7.08 ×102
mean 6.00 ×102 6.23 ×102 7.23 ×102 6.08 ×102 7.18 ×102 6.57 ×102 7.08 ×102 6.96 ×102
std 4.06 ×101 7.22 ×100 2.02 ×101 1.83 ×100 1.57 ×101 3.62 ×100 1.03 ×101 7.88 ×100
rank 1 3 8 2 7 4 6 5
F7 min 7.39 ×102 7.43 ×102 1.40 ×103 8.74 ×102 1.53 ×103 8.91 ×102 2.30 ×103 1.45 ×103
max 7.57 ×102 7.76 ×102 3.81 ×103 9.43 ×102 3.05 ×103 1.03 ×103 3.09 ×103 1.74 ×103
mean 7.49 ×102 7.54 ×102 2.51 ×103 9.13 ×102 1.79 ×103 9.49 ×102 2.77 ×103 1.55 ×103
std 4.58 ×100 8.25 ×100 8.83 ×102 1.83 ×101 2.98 ×102 3.29 ×101 1.89 ×102 6.34 ×101
rank 1 2 7 3 6 4 8 5
F8 min 8.14 ×102 8.51 ×102 1.16 ×103 9.18 ×102 1.13 ×103 9.28 ×102 1.22 ×103 1.11 ×103
max 8.36 ×102 8.93 ×102 1.46 ×103 9.79 ×102 1.44 ×103 1.00 ×103 1.34 ×103 1.23 ×103
mean 8.25 ×102 8.67 ×102 1.30 ×103 9.55 ×102 1.24 ×103 9.58 ×102 1.29 ×103 1.19 ×103
std 5.63 ×100 1.20 ×101 9.98 ×101 1.63 ×101 7.13 ×101 1.82 ×101 2.69 ×101 2.79 ×101
rank 1 2 8 3 6 4 7 5
F9 min 9.00 ×102 9.45 ×102 1.17 ×104 2.65 ×103 1.31 ×104 3.50 ×103 1.68 ×104 1.10 ×104
max 9.77 ×102 1.60 ×103 4.70 ×104 7.51 ×103 3.35 ×104 5.21 ×103 3.05 ×104 1.90 ×104
mean 9.15 ×102 1.24 ×103 2.75 ×104 5.18 ×103 2.13 ×104 4.24 ×103 2.49 ×104 1.43 ×104
std 2.13 ×101 1.92 ×102 1.20 ×104 1.20 ×103 5.16 ×103 4.28 ×102 2.63 ×103 1.72 ×103
rank 1 2 8 4 6 3 7 5
F10 min 4.45 ×103 3.87 ×103 9.35 ×103 4.30 ×103 8.08 ×103 4.25 ×103 8.19 ×103 8.44 ×103
max 5.75 ×103 5.56 ×103 1.16 ×104 5.41 ×103 1.04 ×104 6.21 ×103 9.24 ×103 1.01 ×104
mean 5.04 ×103 4.79 ×103 1.07 ×104 5.02 ×103 9.51 ×103 5.09 ×103 8.82 ×103 9.29 ×103
std 2.96 ×102 4.21 ×102 5.74 ×102 2.33 ×102 5.40 ×102 5.35 ×102 2.51 ×102 4.40 ×102
rank 3 1 8 2 7 4 5 6
F11 min 1.37 ×103 3.29 ×103 1.07 ×104 1.53 ×103 1.10 ×104 2.27 ×103 8.64 ×103 6.72 ×103
max 5.60 ×103 9.98 ×103 2.61 ×105 4.80 ×103 5.75 ×104 7.76 ×103 2.04 ×104 1.56 ×104
mean 2.79 ×103 6.87 ×103 4.19 ×104 2.38 ×103 3.29 ×104 4.42 ×103 1.44 ×104 1.17 ×104
std 1.04 ×103 1.67 ×103 5.08 ×104 6.88 ×102 1.21 ×104 1.15 ×103 3.30 ×103 2.14 ×103
rank 2 4 8 1 7 3 6 5
F12 min 3.91 ×105 2.25 ×106 7.45 ×109 4.62 ×106 4.76 ×109 2.95 ×106 6.18 ×109 8.81 ×109
max 1.13 ×107 2.60 ×107 3.66 ×1010 2.69 ×107 2.12 ×1010 2.51 ×108 1.68 ×1010 2.27 ×1010
mean 3.47 ×106 1.05 ×107 2.24 ×1010 1.25 ×107 1.15 ×1010 8.39 ×107 1.10 ×1010 1.70 ×1010
std 2.70 ×106 6.10 ×106 7.68 ×109 5.32 ×106 4.05 ×109 7.17 ×107 2.19 ×109 3.23 ×109
rank 1 2 8 3 6 4 5 7
F13 min 4.04 ×103 1.48 ×104 5.99 ×109 9.32 ×105 2.51 ×109 2.22 ×104 2.97 ×109 2.61 ×109
max 6.38 ×104 6.22 ×104 4.36 ×1010 1.69 ×107 2.00 ×1010 4.84 ×104 9.81 ×109 1.82 ×1010
mean 2.91 ×104 3.30 ×104 2.04 ×1010 6.41 ×106 7.77 ×109 3.25 ×104 5.85 ×109 9.74 ×109
std 1.61 ×104 1.16 ×104 1.06 ×1010 3.73 ×106 3.67 ×109 6.89 ×103 1.88 ×109 4.29 ×109
rank 1 3 8 4 6 2 5 7
F14 min 2.50 ×105 3.72 ×105 3.06 ×106 1.52 ×104 9.88 ×105 5.58 ×105 4.97 ×105 1.55 ×105
max 2.38 ×106 2.17 ×106 2.33 ×108 1.19 ×106 3.21 ×107 1.91 ×106 4.51 ×106 4.25 ×107
mean 1.22 ×106 1.23 ×106 5.39 ×107 4.23 ×105 7.73 ×106 1.11 ×106 1.87 ×106 6.26 ×106
std 5.41 ×105 4.59 ×105 5.83 ×107 2.83 ×105 7.57 ×106 3.79 ×105 1.26 ×106 8.83 ×106
rank 3 4 8 1 7 2 5 6
F15 min 2.20 ×103 8.99 ×103 7.26 ×108 5.15 ×104 1.94 ×108 9.06 ×103 2.11 ×108 3.51 ×107
max 1.73 ×104 3.53 ×104 1.19 ×1010 3.42 ×106 4.82 ×109 2.47 ×104 1.10 ×109 3.89 ×109
mean 6.74 ×103 1.61 ×104 5.16 ×109 8.77 ×105 1.60 ×109 1.53 ×104 6.69 ×108 1.62 ×109
std 4.02 ×103 5.63 ×103 3.24 ×109 7.52 ×105 1.14 ×109 3.41 ×103 2.02 ×108 1.09 ×109
rank 1 3 8 4 6 2 5 7
F16 min 1.99 ×103 2.48 ×103 6.00 ×103 2.15 ×103 4.52 ×103 2.86 ×103 4.84 ×103 4.91 ×103
max 3.03 ×103 3.55 ×103 1.68 ×104 2.97 ×103 7.17 ×103 4.16 ×103 5.75 ×103 9.79 ×103
mean 2.53 ×103 2.93 ×103 8.98 ×103 2.64 ×103 5.15 ×103 3.51 ×103 5.33 ×103 7.03 ×103
std 3.02 ×102 2.71 ×102 2.23 ×103 1.81 ×102 6.19 ×102 3.34 ×102 2.90 ×102 1.42 ×103
rank 1 3 8 2 5 4 6 7
F17 min 1.79 ×103 2.15 ×103 3.40 ×103 1.90 ×103 3.24 ×103 2.50 ×103 3.16 ×103 3.18 ×103
max 2.59 ×103 3.40 ×103 3.39 ×105 2.27 ×103 7.46 ×103 3.38 ×103 4.18 ×103 2.40 ×104
mean 2.21 ×103 2.61 ×103 3.85 ×104 2.11 ×103 4.10 ×103 2.93 ×103 3.63 ×103 6.71 ×103
std 2.02 ×102 2.28 ×102 7.27 ×104 1.07 ×102 8.46 ×102 2.37 ×102 2.67 ×102 4.57 ×103
rank 2 3 8 1 6 4 5 7
F18 min 5.48 ×104 1.40 ×105 5.68 ×107 2.47 ×105 6.98 ×106 9.68 ×104 6.99 ×106 5.36 ×106
max 3.48 ×106 5.55 ×106 1.96 ×109 2.93 ×106 2.46 ×108 8.86 ×105 6.05 ×107 3.68 ×108
mean 7.18 ×105 1.76 ×106 5.90 ×108 9.86 ×105 6.51 ×107 4.39 ×105 3.22 ×107 1.06 ×108
std 7.67 ×105 1.43 ×106 4.61 ×108 6.19 ×105 5.47 ×107 2.41 ×105 1.25 ×107 9.89 ×107
rank 2 4 8 3 6 1 5 7
F19 min 2.20 ×103 4.93 ×104 3.51 ×108 3.26 ×104 3.66 ×108 5.79 ×104 2.75 ×108 4.20 ×107
max 1.48 ×104 4.14 ×106 1.62 ×1010 1.32 ×106 7.46 ×109 4.73 ×105 1.73 ×109 2.90 ×109
mean 6.60 ×103 9.02 ×105 6.00 ×109 5.97 ×105 2.06 ×109 1.99 ×105 9.98 ×108 1.24 ×109
std 3.01 ×103 8.85 ×105 4.41 ×109 4.22 ×105 1.50 ×109 9.72 ×104 3.79 ×108 7.84 ×108
rank 1 4 8 3 7 2 5 6
F20 min 2.65 ×103 2.30 ×103 3.24 ×103 2.23 ×103 2.89 ×103 2.75 ×103 2.76 ×103 2.84 ×103
max 3.10 ×103 3.36 ×103 4.24 ×103 2.64 ×103 3.74 ×103 3.44 ×103 3.24 ×103 3.48 ×103
mean 2.87 ×103 2.83 ×103 3.72 ×103 2.41 ×103 3.41 ×103 3.08 ×103 3.05 ×103 3.19 ×103
std 1.41 ×102 2.49 ×102 2.55 ×102 9.71 ×101 1.67 ×102 1.76 ×102 1.08 ×102 1.62 ×102
rank 3 2 8 1 7 5 4 6
F21 min 2.31 ×103 2.36 ×103 2.76 ×103 2.24 ×103 2.64 ×103 2.58 ×103 2.69 ×103 2.69 ×103
max 2.35 ×103 2.46 ×103 3.06 ×103 2.49 ×103 2.88 ×103 2.71 ×103 2.83 ×103 2.90 ×103
mean 2.33 ×103 2.40 ×103 2.90 ×103 2.44 ×103 2.78 ×103 2.64 ×103 2.77 ×103 2.81 ×103
std 8.00 ×100 1.89 ×101 8.88 ×101 7.77 ×101 5.60 ×101 3.95 ×101 3.70 ×101 5.31 ×101
rank 1 2 8 3 6 4 5 7
F22 min 2.30 ×103 2.30 ×103 8.68 ×103 2.43 ×103 9.10 ×103 6.37 ×103 9.67 ×103 9.37 ×103
max 2.30 ×103 2.32 ×103 1.33 ×104 6.85 ×103 1.16 ×104 8.30 ×103 1.06 ×104 1.12 ×104
mean 2.30 ×103 2.30 ×103 1.19 ×104 2.95 ×103 1.07 ×104 7.31 ×103 1.01 ×104 1.04 ×104
std 7.55 ×101 4.21 ×100 1.19 ×103 1.22 ×103 5.89 ×102 4.72 ×102 2.15 ×102 4.09 ×102
rank 1 2 8 3 7 4 5 6
F23 min 2.67 ×103 2.77 ×103 3.49 ×103 2.70 ×103 3.08 ×103 3.47 ×103 3.13 ×103 3.44 ×103
max 2.72 ×103 3.00 ×103 4.35 ×103 2.85 ×103 3.50 ×103 4.18 ×103 3.40 ×103 4.19 ×103
mean 2.69 ×103 2.87 ×103 3.84 ×103 2.79 ×103 3.26 ×103 3.83 ×103 3.31 ×103 3.73 ×103
std 1.14 ×101 4.51 ×101 2.26 ×102 4.49 ×101 9.72 ×101 1.63 ×102 5.06 ×101 1.65 ×102
rank 1 3 8 2 4 7 5 6
F24 min 2.84 ×103 2.87 ×103 3.81 ×103 2.98 ×103 3.13 ×103 3.37 ×103 3.34 ×103 3.50 ×103
max 2.88 ×103 3.00 ×103 4.88 ×103 3.11 ×103 3.70 ×103 3.67 ×103 3.54 ×103 4.08 ×103
mean 2.86 ×103 2.94 ×103 4.26 ×103 3.05 ×103 3.35 ×103 3.50 ×103 3.45 ×103 3.81 ×103
std 1.17 ×101 3.02 ×101 2.70 ×102 3.34 ×101 1.46 ×102 6.55 ×101 5.72 ×101 1.48 ×102
rank 1 2 8 3 4 6 5 7
F25 min 2.90 ×103 2.93 ×103 4.92 ×103 2.91 ×103 5.12 ×103 2.93 ×103 7.37 ×103 4.72 ×103
max 2.97 ×103 3.02 ×103 2.40 ×104 2.97 ×103 1.41 ×104 3.05 ×103 1.41 ×104 6.94 ×103
mean 2.93 ×103 2.98 ×103 1.23 ×104 2.94 ×103 8.03 ×103 2.99 ×103 1.08 ×104 5.79 ×103
std 2.14 ×101 2.40 ×101 6.06 ×103 1.66 ×101 2.14 ×103 2.01 ×101 1.53 ×103 5.14 ×102
rank 1 3 8 2 6 4 7 5
F26 min 2.93 ×103 2.80 ×103 1.09 ×104 3.27 ×103 8.78 ×103 6.32 ×103 1.04 ×104 1.08 ×104
max 6.16 ×103 7.00 ×103 2.06 ×104 5.43 ×103 1.39 ×104 9.01 ×103 1.21 ×104 1.45 ×104
mean 4.36 ×103 4.50 ×103 1.47 ×104 3.93 ×103 1.11 ×104 7.79 ×103 1.11 ×104 1.32 ×104
std 6.60 ×102 1.34 ×103 2.84 ×103 5.49 ×102 1.29 ×103 5.70 ×102 4.56 ×102 8.17 ×102
rank 2 3 8 1 5 4 6 7
F27 min 3.22 ×103 3.31 ×103 4.18 ×103 3.22 ×103 3.20 ×103 4.43 ×103 3.42 ×103 3.86 ×103
max 3.27 ×103 3.55 ×103 6.35 ×103 3.25 ×103 3.80 ×103 5.65 ×103 3.80 ×103 5.80 ×103
mean 3.24 ×103 3.41 ×103 5.31 ×103 3.23 ×103 3.22 ×103 5.03 ×103 3.66 ×103 4.80 ×103
std 1.46 ×101 5.33 ×101 5.13 ×102 6.12 ×100 1.10 ×102 3.23 ×102 9.67 ×101 5.53 ×102
rank 3 4 8 2 1 7 5 6
F28 min 3.24 ×103 3.30 ×103 6.87 ×103 3.30 ×103 3.30 ×103 3.40 ×103 6.11 ×103 3.31 ×103
max 3.37 ×103 3.41 ×103 1.52 ×104 3.39 ×103 3.30 ×103 4.11 ×103 9.24 ×103 9.42 ×103
mean 3.31 ×103 3.35 ×103 1.13 ×104 3.34 ×103 3.30 ×103 3.59 ×103 8.12 ×103 6.46 ×103
std 2.80 ×101 2.90 ×101 2.37 ×103 1.94 ×101 2.33 ×103 1.87 ×102 8.01 ×102 2.28 ×103
rank 2 4 8 3 1 5 7 6
F29 min 3.50 ×103 4.46 ×103 6.11 ×103 3.52 ×103 5.18 ×103 4.94 ×103 5.13 ×103 5.98 ×103
max 4.41 ×103 5.53 ×103 1.55 ×105 3.99 ×103 1.13 ×104 6.16 ×103 6.97 ×103 1.87 ×104
mean 3.94 ×103 4.86 ×103 3.74 ×104 3.77 ×103 7.29 ×103 5.42 ×103 6.19 ×103 9.73 ×103
std 2.28 ×102 2.41 ×102 3.90 ×104 1.02 ×102 1.56 ×103 2.88 ×102 4.29 ×102 3.34 ×103
rank 2 3 8 1 6 4 5 7
F30 min 2.60 ×104 2.44 ×106 1.35 ×109 1.29 ×105 5.02 ×108 1.35 ×106 2.74 ×108 4.59 ×108
max 1.49 ×106 1.18 ×107 8.48 ×109 2.80 ×106 3.74 ×109 5.28 ×106 1.14 ×109 7.19 ×109
mean 3.23 ×105 5.72 ×106 3.98 ×109 6.71 ×105 1.35 ×109 3.17 ×106 6.92 ×108 2.38 ×109
std 3.59 ×105 2.22 ×106 1.83 ×109 5.39 ×105 7.31 ×108 9.97 ×105 2.31 ×108 1.61 ×109
rank 1 4 8 2 6 3 5 7

To validate the effectiveness of mAEFA, the statistical results of the nonparametric Wilcoxon rank-sum (WRS) test are presented in Table 5. This test helps determine the significance of the differences between mAEFA and other algorithms measured with a p-value of ≤0.05. Based on the results, the mAEFA algorithm demonstrates superior performance compared to the LCA algorithm across all 29 test functions. It also outperforms SAO and GOA in 28 functions, CHIO in 27 functions, PSO in 25 functions, GSA in 24 functions, and AEFA in 17 functions, showcasing better results in each case.

Table 5.

Wilcoxon rank-sum test results for the comparative algorithms against the proposed mAEFA using CEC’17 benchmark functions, where a = 0.05 and Dim = 30.

mAEFA Vs AEFA LCA CHIO GOA GSA CMA-ES SAO
F1 9.59 ×1001 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.18 ×1004 3.02 ×1011 3.02 ×1011
F3 1.52 ×1003 3.02 ×1011 3.02 ×1011 3.02 ×1011 2.78 ×1007 3.02 ×1011 9.92 ×1011
F4 2.92 ×1002 3.02 ×1011 1.17 ×1004 3.02 ×1011 6.70 ×1011 3.02 ×1011 3.02 ×1011
F5 3.00 ×1011 3.00 ×1011 3.00 ×1011 3.00 ×1011 3.00 ×1011 3.00 ×1011 3.00 ×1011
F6 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F7 1.19 ×1001 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F8 3.00 ×1011 3.01 ×1011 3.01 ×1011 3.01 ×1011 3.01 ×1011 3.01 ×1011 3.01 ×1011
F9 4.50 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F10 2.81 ×1002 3.02 ×1011 4.92 ×1001 3.02 ×1011 7.96 ×1001 3.02 ×1011 3.02 ×1011
F11 1.46 ×1010 3.02 ×1011 1.22 ×1001 3.02 ×1011 1.29 ×1006 3.02 ×1011 3.02 ×1011
F12 3.01 ×1007 3.02 ×1011 2.03 ×1009 3.02 ×1011 8.89 ×1010 3.02 ×1011 3.02 ×1011
F13 1.58 ×1001 3.02 ×1011 3.02 ×1011 3.02 ×1011 9.63 ×1002 3.02 ×1011 3.02 ×1011
F14 8.88 ×1001 3.02 ×1011 4.31 ×1008 4.69 ×1008 5.01 ×1001 1.26 ×1001 6.36 ×1005
F15 7.12 ×1009 3.02 ×1011 3.02 ×1011 3.02 ×1011 1.10 ×1008 3.02 ×1011 3.02 ×1011
F16 2.13 ×1005 3.02 ×1011 6.79 ×1002 3.02 ×1011 8.99 ×1011 3.02 ×1011 3.02 ×1011
F17 1.56 ×1008 3.02 ×1011 3.03 ×1002 3.02 ×1011 5.49 ×1011 3.02 ×1011 3.02 ×1011
F18 4.22 ×1004 3.02 ×1011 1.50 ×1002 3.02 ×1011 5.40 ×1001 3.02 ×1011 3.02 ×1011
F19 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F20 2.58 ×1001 3.02 ×1011 3.02 ×1011 8.15 ×1011 1.53 ×1005 6.74 ×1006 2.39 ×1008
F21 3.02 ×1011 3.02 ×1011 8.88 ×1006 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F22 6.59 ×1006 2.88 ×1011 2.88 ×1011 2.88 ×1011 2.88 ×1011 2.88 ×1011 2.88 ×1011
F23 3.02 ×1011 3.02 ×1011 1.09 ×1010 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F24 5.49 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F25 1.01 ×1008 3.02 ×1011 8.24 ×1002 3.02 ×1011 2.15 ×1010 3.02 ×1011 3.02 ×1011
F26 7.96 ×1001 3.02 ×1011 1.41 ×1004 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F27 3.02 ×1011 3.02 ×1011 2.68 ×1006 5.57 ×1010 3.02 ×1011 3.02 ×1011 3.02 ×1011
F28 2.57 ×1007 3.02 ×1011 3.59 ×1005 2.71 ×1002 3.02 ×1011 3.02 ×1011 1.55 ×1009
F29 3.02 ×1011 3.02 ×1011 7.30 ×1004 3.02 ×1011 3.02 ×1011 3.02 ×1011 3.02 ×1011
F30 3.02 ×1011 3.02 ×1011 4.71 ×1004 3.02 ×1011 3.34 ×1011 3.02 ×1011 3.02 ×1011

The convergence behavior of the proposed mAEFA algorithm, along with the other compared algorithms, applied to the CEC’17 functions is depicted in Figure 2, Figure 3 and Figure 4. It can be observed that mAEFA demonstrates rapid convergence for F3–F6, F8, F11–F14, F16, F17, F19, F21, F23, F24, and F27–30 compared to the classical AEFA algorithm. Additionally, mAEFA on average depicted better convergence behavior relative to the other algorithms in most of the CEC test suite benchmarks. The mAEFA algorithm demonstrates a balanced exploration–exploitation trade-off, evident in its convergence plots, showcasing both faster convergence and the discovery of optimal solutions during the search process.

Figure 2.

Figure 2

Graphical representation of convergence curves for functions ranging from F1 to F10 using CEC’17 with a dimensionality of 30.

Figure 3.

Figure 3

Graphical representation of convergence curves for functions ranging from F11 to F20 using CEC’17 with a dimensionality of 30.

Figure 4.

Figure 4

Graphical representation of convergence curves for functions ranging from F21 to F30 using CEC’17 with a dimensionality of 30.

Additionally, Figure 5, Figure 6 and Figure 7 shows box plots for the proposed algorithm and the other compared algorithms, highlighting the distribution of results and the degree of proximity and similarity across multiple runs. The box plot is utilized to represent the minimum, maximum, and mean values for each algorithm. Box plots serve as an effective visual representation for presenting the performance values obtained in the 1st, 2nd, and 3rd] quarters of the experiment. A horizontal line within the box denotes the median value, while the whiskers extending outside indicate the variability beyond the upper and lower quartiles. Ideally, smaller box sizes and lower variability are preferred. Figure 5, Figure 6 and Figure 7 illustrate box plots for mAEFA for the majority of the functions, displaying narrow boxes with the lowest median and variability.

Figure 5.

Figure 5

Box plots illustrating the distribution of results for functions spanning from F1 to F10 using CEC’17 with a dimensionality of 30.

Figure 6.

Figure 6

Box plots illustrating the distribution of results for functions spanning from F11 to F20 using CEC’17 with a dimensionality of 30.

Figure 7.

Figure 7

Box plots illustrating the distribution of results for functions spanning from F21 to F30 using CEC’17 with a dimensionality of 30.

5.1.2. Comparing mAEFA with Fitness–Distance Balance Variants

In this subsection we compare mAEFA with two other FDB variants, namely FDB-SOS [66] and dfDB-MRFO [69]. Table 6 presents the statistical analysis results in terms of minimum, maximum, average, and standard deviation values, along with the ranks of the algorithms. Upon examination, mAEFA consistently outperforms the other variants across several functions, as evidenced by its lower minimum and average values, as well as its higher ranks in most cases. For example, in functions F5–F9, F19, F21–F24, and F26, mAEFA demonstrates superior performance with lower minimum and average values compared to dfDB-MRFO and FDB-SOS, resulting in higher ranks. Additionally, mAEFA shows competitive performance in other functions, such as F1, F10, F15, F27, and F29, where it achieves comparable results to the other variants. Overall, the comparative analysis suggests that mAEFA exhibits favorable performance across a range of benchmark functions compared to dfDB-MRFO and FDB-SOS. These findings underscore the efficacy of mAEFA as a competitive algorithm for solving optimization problems, particularly in the context of the functions analyzed in this study.

Table 6.

Statistical outcomes comparing mAEFA against FDB variants on CEC’17 benchmark functions.

Fun. mAEFA dfDB-MRFO FDB-SOS Fun. mAEFA dfDB-MRFO FDB-SOS
F1 min 1.04 ×102 1.01 ×102 1.01 ×102 F3 min 7.63 ×104 1.46 ×104 5.50 ×103
max 1.82 ×104 1.04 ×104 1.80 ×104 max 8.84 ×104 4.51 ×104 2.09 ×104
avg 3.30 ×103 2.58 ×103 4.05 ×103 avg 8.40 ×104 2.83 ×104 1.20 ×104
STD 4.12 ×103 2.70 ×103 4.51 ×103 STD 3.00 ×103 7.58 ×103 3.66 ×103
rank 2 1 3 rank 3 2 1
F4 min 4.94 ×102 4.21 ×102 4.04 ×102 F5 min 5.17 ×102 6.52 ×102 5.73 ×102
max 5.74 ×102 5.25 ×102 5.30 ×102 max 5.41 ×102 7.38 ×102 7.63 ×102
avg 5.25 ×102 4.84 ×102 4.81 ×102 avg 5.25 ×102 6.96 ×102 6.36 ×102
STD 1.26 ×101 2.69 ×101 2.71 ×101 STD 5.71 ×100 1.86 ×101 4.75 ×101
rank 3 2 1 rank 1 3 2
F6 min 6.00 ×102 6.00 ×102 6.00 ×102 F7 min 7.41 ×102 9.24 ×102 7.98 ×102
max 6.02 ×102 6.03 ×102 6.28 ×102 max 7.64 ×102 9.82 ×102 1.02 ×103
avg 6.00 ×102 6.01 ×102 6.04 ×102 avg 7.49 ×102 9.52 ×102 8.90 ×102
STD 4.05 ×101 7.35 ×101 5.88 ×100 STD 4.95 ×100 1.59 ×101 6.56 ×101
rank 1 2 3 rank 1 3 2
F8 min 8.13 ×102 9.38 ×102 8.69 ×102 F9 min 9.00 ×102 9.02 ×102 1.23 ×103
max 8.42 ×102 1.02 ×103 1.04 ×103 max 9.73 ×102 1.61 ×103 4.19 ×103
avg 8.24 ×102 9.89 ×102 9.31 ×102 avg 9.14 ×102 1.00 ×103 2.39 ×103
STD 7.35 ×100 1.70 ×101 4.07 ×101 STD 2.07 ×101 1.52 ×102 8.57 ×102
rank 1 3 2 rank 1 2 3
F10 min 4.61 ×103 6.14 ×103 3.40 ×103 F11 min 1.23 ×103 1.16 ×103 1.12 ×103
max 5.32 ×103 7.82 ×103 6.21 ×103 max 9.04 ×103 1.29 ×103 1.26 ×103
avg 4.95 ×103 6.87 ×103 4.93 ×103 avg 3.51 ×103 1.23 ×103 1.16 ×103
STD 1.74 ×102 3.87 ×102 8.30 ×102 STD 1.66 ×103 2.84 ×101 3.07 ×101
rank 2 3 1 rank 3 2 1
F12 min 4.04 ×104 1.18 ×104 3.41 ×104 F13 min 3.27 ×103 1.60 ×103 1.32 ×103
max 1.51 ×107 7.61 ×105 1.51 ×106 max 5.35 ×104 5.56 ×104 6.05 ×104
avg 3.76 ×106 2.31 ×105 4.09 ×105 avg 2.63 ×104 1.31 ×104 1.52 ×104
STD 3.01 ×106 1.73 ×105 4.14 ×105 STD 1.24 ×104 1.45 ×104 1.47 ×104
rank 3 1 2 rank 3 1 2
F14 min 3.30 ×105 2.15 ×103 2.03 ×103 F15 min 2.09 ×103 1.88 ×103 1.61 ×103
max 2.20 ×106 1.25 ×105 7.76 ×104 max 2.27 ×104 1.77 ×105 3.31 ×104
avg 1.11 ×106 1.83 ×104 1.20 ×104 avg 7.14 ×103 2.42 ×104 5.73 ×103
STD 5.12 ×105 2.58 ×104 1.46 ×104 STD 4.47 ×103 4.27 ×104 6.80 ×103
rank 3 2 1 rank 2 3 1
F16 min 1.86 ×103 1.64 ×103 1.85 ×103 F17 min 1.77 ×103 1.77 ×103 1.76 ×103
max 3.35 ×103 3.26 ×103 3.10 ×103 max 2.62 ×103 2.32 ×103 2.50 ×103
avg 2.67 ×103 2.45 ×103 2.52 ×103 avg 2.13 ×103 1.91 ×103 2.11 ×103
STD 3.27 ×102 4.26 ×102 3.32 ×102 STD 1.85 ×102 1.39 ×102 2.02 ×102
rank 3 1 2 rank 3 1 2
F18 min 9.29 ×104 6.06 ×104 3.02 ×104 F19 min 2.90 ×103 1.97 ×103 1.96 ×103
max 2.59 ×106 1.27 ×106 6.15 ×105 max 1.73 ×104 3.93 ×105 3.36 ×104
avg 9.00 ×105 3.83 ×105 1.56 ×105 avg 6.61 ×103 3.11 ×104 7.98 ×103
STD 7.58 ×105 2.98 ×105 1.23 ×105 STD 3.44 ×103 7.56 ×104 6.88 ×103
rank 3 2 1 rank 1 3 2
F20 min 2.65 ×103 2.06 ×103 2.16 ×103 F21 min 2.31 ×103 2.45 ×103 2.34 ×103
max 3.12 ×103 2.72 ×103 2.67 ×103 max 2.35 ×103 2.51 ×103 2.43 ×103
avg 2.89 ×103 2.34 ×103 2.45 ×103 avg 2.33 ×103 2.48 ×103 2.38 ×103
STD 1.60 ×102 1.65 ×102 1.23 ×102 STD 9.82 ×100 1.74 ×101 2.34 ×101
rank 3 1 2 rank 1 3 2
F22 min 2.30 ×103 2.30 ×103 2.30 ×103 F23 min 2.66 ×103 2.81 ×103 2.70 ×103
max 2.30 ×103 2.31 ×103 5.28 ×103 max 2.73 ×103 2.88 ×103 2.82 ×103
avg 2.30 ×103 2.30 ×103 2.40 ×103 avg 2.70 ×103 2.84 ×103 2.76 ×103
STD 6.23 ×101 1.62 ×100 5.44 ×102 STD 1.73 ×101 1.65 ×101 2.71 ×101
rank 1 2 3 rank 1 3 2
F24 min 2.84 ×103 2.98 ×103 2.87 ×103 F25 min 2.89 ×103 2.88 ×103 2.88 ×103
max 2.88 ×103 3.06 ×103 3.01 ×103 max 3.00 ×103 2.94 ×103 2.94 ×103
avg 2.85 ×103 3.03 ×103 2.93 ×103 avg 2.93 ×103 2.89 ×103 2.89 ×103
STD 1.24 ×101 1.67 ×101 3.36 ×101 STD 2.69 ×101 1.46 ×101 1.33 ×101
rank 1 3 2 rank 3 2 1
F26 min 2.80 ×103 4.99 ×103 2.80 ×103 F27 min 3.20 ×103 3.19 ×103 3.22 ×103
max 5.08 ×103 5.81 ×103 7.39 ×103 max 3.27 ×103 3.27 ×103 3.27 ×103
avg 4.20 ×103 5.51 ×103 4.92 ×103 avg 3.24 ×103 3.22 ×103 3.25 ×103
STD 5.25 ×102 2.16 ×102 1.39 ×103 STD 1.33 ×101 1.69 ×101 1.56 ×101
rank 1 3 2 rank 2 1 3
F28 min 3.27 ×103 3.11 ×103 3.19 ×103 F29 min 3.51 ×103 3.48 ×103 3.62 ×103
max 3.42 ×103 3.26 ×103 3.26 ×103 max 4.30 ×103 4.12 ×103 4.32 ×103
avg 3.31 ×103 3.20 ×103 3.21 ×103 avg 3.87 ×103 3.77 ×103 3.88 ×103
STD 2.68 ×101 3.06 ×101 1.95 ×101 STD 1.83 ×102 1.75 ×102 1.71 ×102
rank 3 1 2 rank 2 1 3
F30 min 1.70 ×104 5.78 ×103 6.25 ×103
max 2.59 ×106 1.95 ×105 1.81 ×104
avg 3.14 ×105 3.40 ×104 1.03 ×104
STD 4.80 ×105 4.79 ×104 2.54 ×103

5.2. Experimental Series 2: Engineering Problems

In this section, we explore five distinct engineering challenges to assess the effectiveness of the proposed solution in dealing with problems that involve constraints. The following constraint problems, which are commonly encountered and widely utilized, have been considered: speed reducer design (SRD), pressure vessel design (PVD), cantilever beam design (CBD) and the multi-product batch plant (MPBP) problem, and Industrial Refrigeration System Optimal Design.

5.2.1. Speed Reducer Design Problem

This problem involves optimization with the objective of minimizing the weights associated with various design elements. The optimization process incorporates constraints related to gear teeth, stress, deflection ratios of bending, surface, and shafts [78]. The SRD is based on seven design variables (z1,z2,z3,z4,z5,z6,andz7) that aim to minimize the weight. These variables represent face width, teeth module, pinion teeth number, the length of the first shaft between the bearings, the length of the second shaft between the bearings, and the diameters of the first and second shafts.

The mAEFA algorithm attains the top rank with a significantly low standard deviation, as indicated by the results in Table 7. Furthermore, Table 8 illustrates its competitive performance across all seven variables, underscoring the noteworthy differentiation of mAEFA in comparison to alternative algorithms. Figure 8 illustrates a distinct divergence in convergence patterns among the algorithms, where mAEFA exhibits superior convergence by attaining exceptionally low values in the final iteration of the search process.

LetMinf(z)=0.7854z1z22(3.3333z32+14.9334z343.0934)1.508z1(z62+z72)+7.4777(z63+z73)+0.7854(z4z62+z5z72)Subjectto:R1(z)=27z1z22z310R2(z)=397.5z1z22z310R3(z)=1.93z43z2z3z641R4(z)=1.93z53z2z3z7410R5(z)=1110z63(745z4z2z3)2+16.9×10610R6(z)=185z73(745z5z2z3)2+157.5×10610R7(z)=z2z34010R8(z)=5z2z110R9(z)=z112z210R10(z)=1.5z6+1.9z410R11(z)=1.1z7+1.9z510where2.6z13.6,0.7z20.8,17z328,7.3x48.3,7.8z58.3,2.9z63.9,and5z75.5 (10)
Table 7.

Statistical outcomes comparing mAEFA against other metaheuristics applied to speed reducer design.

Algorithms Rank Best Worst Mean STD
mAEFA 1 2999.125 3276.964 3081.04 96.0022
AEFA 2 2998.704 3357.957 3161.609 125.7413
LCA 4 3084.932 3394.719 3217.231 81.89584
CHIO 3 3028.974 3438.588 3161.642 81.68969
GOA 8 3034.447 8843.277 4425.907 1503.485
GSA 5 3149.597 3818.236 3443.538 180.2837
POS 6 2993.701 5905.984 3487.556 687.8122
SAO 7 3230.902 8071.254 4153.676 956.0172
Table 8.

Results of mAEFA versus other metaheuristics on speed reducer design.

Best z1 z2 z3 z4 z5 z6 z7
mAEFA 2999.125 3.497593 0.7 17 7.342937 7.943631 3.350136 5.285752
AEFA 2998.704 3.49759 0.7 17 7.627354 7.805211 3.350677 5.285705
LCA 3084.932 3.483299 0.70015 17.05791 8.058002 8.247121 3.452092 5.311111
CHIO 3028.974 3.497172 0.7 17 7.51058 7.711363 3.44856 5.294154
GOA 3034.447 3.571421 0.701486 17 7.649347 7.714521 3.353149 5.288426
GSA 3149.597 3.592483 0.711502 17.12052 7.344929 7.683132 3.454113 5.282249
POS 2993.701 3.49765 0.7 17 7.3 7.712778 3.350306 5.28539
SAO 3230.902 3.6 2.6 3.565453 2.84373 2.951027 3.314187 2.715053
Figure 8.

Figure 8

Speed reducer design.

5.2.2. Pressure Vessel Design Problem

The PVD problem, introduced by Kannan and Kramer [79], is the second engineering design problem in this study, assessing the effectiveness of the proposed mAEFA. This problem revolves around optimizing the cost associated with PVD, aiming to minimize it. The cost is determined by four design variables: z1,z2,z3,andz4, which represent shell thickness, head thickness, inner radius, and cylinder length, respectively. Hashim et al. [80] provide an extensive description of the mathematical model for this problem. The PVD problem’s mathematical model can be defined using the following set of equations.

According to the findings presented in Table 9, the mAEFA algorithm obtains the first rank with a minimum mean value of 6412.152. Figure 9 demonstrates comparable behavior across all algorithms, with mAEFA consistently achieving minimum values during the later iterations. Table 10 presents the variable values obtained by each algorithm for this specific problem. Notably, mAEFA achieves competitive values for z1 and z3, and it secures the second-lowest value for z2. Table 10 clearly indicates that mAEFA surpasses all algorithms, except for the LCA algorithm, indicating a substantial difference in its performance.

LetMinf(z)=0.6224z1z3z4+1.7781z2z32+3.1661z12z4+19.84z12z3Subjectto:R1(z)=z1+0.0193zR2(z)=z2+0.00954z30R3(z)=πz32z4(4/3)πz33+1,296,0000R4(z)=z42400where0zi100,i=1,210zi200,i=3,4 (11)
Table 9.

Statistical results of mAEFA versus other metaheuristics on pressure vessel design problem.

Measures Best Worst Mean STD Rank
mAEFA 6412.152 15600.59 8646.057 1887.591 2
AEFA 6769.697 18818.75 9415.145 2473.888 3
LCA 6690.927 9928.012 8388.067 900.3348 1
CHIO 7381.037 13270.55 9741.683 1611.388 4
GOA 7411.902 40809.83 16388.75 8315.391 8
GSA 6390.632 40404.85 15418.54 7878.361 7
POS 6266.019 14690.99 10816.18 2322.216 5
SAO 7710.968 23390.18 13266.08 3334.069 6
Figure 9.

Figure 9

Pressure vessel design problem.

Table 10.

Statistical results of mAEFA compared with other algorithms on pressure vessel design problem.

Best z1 z2 z3 z4
mAEFA 6412.152 0.824149 0.407546 42.89068 185.7112
AEFA 6769.697 0.902527 0.445849 46.93715 146.8159
LCA 6690.927 0.838483 0.395264 44.07847 155.5208
CHIO 7381.037 0.926379 0.602382 42.96878 169.9356
GOA 7411.902 1.083653 0.572748 50 98.84785
GSA 6390.632 0.968216 0.477738 50.30862 98.92202
POS 6266.019 0.924327 0.443053 46.73207 126.5877
SAO 7710.968 2.527902 5.801292 1.559886 5.133647

5.2.3. Multi-Product Batch Plant Problem

In the MPBP problem, the production process begins after the customer submits their order, with each order representing a specific product. Throughout production, the batch size for each order remains constant. Each order is assigned due dates and release dates. At each stage, there are dedicated processing units that exclusively operate within that stage. The objective of this problem is to minimize the makespan, considering additional constraints such as unallowed unit assignments, order due dates, release dates, and storage issues. The problem’s formulation is delineated by Equations (12)–(23), with Equation (12) addressing the constraint related to order assignment, ensuring that each order (i) can only be processed on a single unit (j) at a specific step (s).

jJisZij=1,iIj (12)

Equations (13)–(15) reveal unit order sequencing, with Equations (13) and (15) indicating a singular first order for each unit j, and Equation (12) depicting sequence constraints for different orders i and i′ on the same unit j.

iIjZFij1 (13)
iIsXiis+jJisZij=1,IIs (14)
ZijZFij,iIj# (15)

Equations (16) and (17) represent unit assignment constraints. When integer variables Xiis or Xis are activated, it implies that orders i and i must be processed on the same unit j.

2(Xiis+Xis)+jJisJisZij+jJIsJisZijZij2,i>i,(i,i)Is (16)
ZijZij,jJisJij,(i,i)Is (17)

Equations (16)–(19) represent order timing constraints. Equation (18) pertains to timing constraints within one order’s various steps, while Equation (19) addresses timing constraints for multiple orders on the same unit. Additionally, when considering unit release time URj or order release time ORi, Equations (20) and (21) come into play. Equation (22) is applicable for cases involving due date DDi.

TisTis+jJisZijPTijs=nsis,sSi (18)
M(1Xiis)+TisTis+jJisZijPTijs=nsis,sSi (19)
TisjJisZFijURj,iIs (20)
TisORi,s=fsi (21)
Tis+jJisZijtijDDi,iI,s=Isi (22)

The goal is to minimize the makespan, and the objective function is defined as follows:

Make-span=minmaxTis+jJisZijPTij,s=Isi (23)

Meeting all constraints is straightforward except for Equation (22) when minimizing the makespan with metaheuristic algorithms. To accommodate Equation (22), a penalty function is employed, and Equations (24) and (25) are utilized to calculate the objective function in this study:

dimaxTis+jJisZijPTijDDi,iI,s=Isi (24)
Objectivefunction=minmaxTis+jJisZijPTij+MiIdi,s=Isi (25)

Equation (24) defines the penalty function, which is employed to penalize violations in Equation (25). When the completion time of each order exceeds its respective due date, Equation (24) comes into effect, leading to a reduction in the objective value in Equation (25).

According to the data presented in Table 11, the mAEFA algorithm secures the first position in terms of statistical ranking amongst all compared algorithms. Table 12 showcases that mAEFA attains competitive results across various problem variables. Figure 10 depicts that mAEFA, alongside some other algorithms, demonstrates similar convergence behavior reaching minimum values. However, both CHIO and SAO exhibit premature convergence behavior, indicating a tendency to get trapped in local minima.

Table 11.

Statistical results of mAEFA versus other metaheuristics on multi-product batch plant problem.

Measures Best Worst Mean STD Rank
mAEFA 63491.75 98704.83 80183.47 8834.615 1
AEFA 61081.05 101036.2 85863.96 9781.795 2
LCA 81552.61 109366.8 98317.28 6991.52 4
CHIO 74939.34 120685.6 97692.22 8345.171 3
GOA 125170.2 4.71 ×1010 3.39 ×109 1.03 ×1010 7
GSA 83258.78 157410.5 99175.74 11609.84 5
POS 76198.96 165947.4 112619.2 25070.13 6
SAO 80120.89 6.28 ×1010 5.12 ×109 1.26 ×1010 8
Table 12.

Results of mAEFA versus other metaheuristics on multi-product batch plant problem.

Best x1 x2 x3 x4 x5 x6 x7 x8 x9 x10
mAEFA 63491.75 0.52857 1.349136 1.036593 1745.955 1756.168 1404.395 19.99933 15.99943 233.0765 136.8181
AEFA 61081.05 0.753475 0.635496 0.530559 1555.817 1613.788 1423.447 20 16 227.9073 140.7617
LCA 81552.61 2.123418 0.523666 0.986531 1431.905 1882.782 1377.451 19.98424 8.146837 259.8943 81.26382
CHIO 74939.34 2.463464 1.862752 1.040579 1004.223 904.7103 781.3551 10.0301 8.258278 136.7337 58.49508
GOA 125170.2 1.638572 3.111043 2.266871 1145.717 1248.944 1296.439 7.316115 15.46341 140.0631 125.618
GSA 83258.78 1.07281 1.987149 1.491329 1548.196 1431.676 1419.545 14.63929 15.93547 270.3076 103.232
POS 76198.96 1.576433 1.501728 1.39028 746.8241 1203.952 864.5001 9.999428 8.189601 131.1378 55.74986
SAO 80120.89 2.324455 2.900101 1.435999 2.209186 1.393715 0.792087 3.341121 1.503093 3.419913 2.639791
Figure 10.

Figure 10

Multi-product batch plant problem.

5.2.4. Industrial Refrigeration System Optimal Design

A refrigeration system utilizes coolant to decrease the temperature of a hot stream, going through three distinct phases. Each phase incorporates a heat exchanger on one side and a boiling cooler on the other. The pumping current is determined by the heat exchanger’s surface area. Moreover, the boiling temperature for the refrigerant is established at the start of each phase. Designing an efficient cooling system entails calculating the surface areas of the three surfaces of the liquid cooling heat exchanger.

The refrigeration system is designed to dissipate 4186.8 J/kg°C of heat while pumping a flow rate of 10,800 kg per hour from an initial temperature of 10 °C to −55 °C. Operating for a minimum of 300 days per year, the refrigeration system’s key parameters include a refrigerant latent heat (λ) of 232.600 J/Kg and an overall heat transfer coefficient of 1130 J/s m2 °C. The primary design objective is to minimize the cost of the three steps, as specified by Equation (26).

Cost=i=13[ci(Ci)0.5+diMi] (26)

The optimization aims to obtain competitive values for design variables, including fluid temperatures, heat exchange area, and liquid refrigerant addition rates in each step, with a focus on minimizing costs. The optimization process places significant emphasis on the temperature of the liquid refrigerant in each step, described as follows:

Temp1=18 °C,Temp2=40 °C,Temp3=62 °C (27)

The temperature of the incoming fluid to the system is 10 °C, denoted as Temp0, while the temperature of the outgoing fluid from the system is −55 °C, represented as Temp3. It is essential for the output temperature at each step to be higher than the temperature of the refrigerant. As a result, the conditions for the design variables can be expressed as shown in Equations (28)–(33).

Temp0=10 °CTemp118 °C40TTC (28)
Temp1Temp240 °C (29)
Fi=Hi×Ci×ΔTempiln (30)

The log mean temperature difference at stage i is:

(ΔTempi)ln=Tempi1TempilnTempiTempRiTempi1TempRi (31)

The energy balance over refrigerant is:

Fi=λi×Mi (32)

λi is the penalty factor, where Fi is the ratio of heat flow, J/s. The energy balance over the fluid is:

Fi=V×kl×(Tempi1Tempi) (33)

where Kl is the specific heat of fluid, J/kg°C and V is the hot fluid pump ratio, kg/hr.

Based on the findings presented in Table 13, the proposed mAEFA algorithm achieves a competitive rank when compared to other algorithms. Additionally, as indicated in Table 14, mAEFA attains the smallest values for the majority of the problem variables and the best solution in comparison to other algorithms. Table 14 demonstrates that mAEFA yields significant results in comparison to all other algorithms. Convergence behavior, as depicted in Figure 11, varies among the algorithms in their approach towards the optimal solution. Notably, mAEFA demonstrates a favourable convergence pattern, reaching minimum values in later search iterations.

Table 13.

Statistical results of mAEFA versus other metaheuristics on design of industrial refrigeration system.

Measures Best Worst Mean STD Rank
mAEFA 5103.195 6845.299 5897.153 1243.806 2
AEFA 5854.164 7098.685 6584.685 1123.588 1
LCA 3846.410 27925.458 10666.124 5210.384 6
CHIO 2093.402 35616.788 19387.112 9003.481 7
GOA 8716.021 65297.497 9506.237 14314.970 5
GSA 7799.196 10860.561 4167.086 2660.564 3
POS 6053.37 146 ×108 5008.193 26683.805 4
SAO 4246.044 476 ×109 788 ×108 926 ×108 8
Table 14.

Results of mAEFA versus other metaheuristics on design of industrial refrigeration system.

mAEFA AEFA LCA CHIO GOA GSA POS SAO
x1 0.001 0.001 0.353413 0.001 0.001 0.143356 0.001 4.056261
x2 0.001 0.001 0.740792 0.001 0.116401 0.954376 0.001 3.897615
x3 2.936496 5 0.507325 0.172527 3.365744 2.404949 0.001 3.83318
x4 3.167783 0.001 4.210235 3.26012 0.01427 3.091417 0.582783 0.341573
x5 1.65173 5 1.659602 2.89829 2.575789 3.735812 0.001 0.159483
x6 4.565306 0.001 4.517057 1.692646 1.660094 1.984339 0.001 4.162318
x7 2.304624 5 1.896705 2.22803 2.992306 3.061548 1.667485 3.263213
x8 3.539738 5 2.371443 1.740014 3.971143 2.334031 2.171946 3.984312
x9 3.060001 5 4.956327 1.025065 4.629864 2.111681 4.99986 0.914673
x10 4.983777 2.04176 4.226494 5 4.642211 4.546502 1.999999 5
x11 2.626728 0.001 3.40886 3.963239 0.001 3.387399 0.001 0.729443
x12 0.06055 0.001 0.135978 0.205029 0.001 0.07326 0.001 4.060669
x13 0.068226 0.001 0.122193 0.120476 0.001 0.065854 0.001 4.240338
x14 0.797011 0.001 1.593305 3.023032 0.001 1.122458 0.010963 5
best 5103.195 5854.164 3846.410 2093.402 8716.021 779919.6 6053.37 14246.044
Figure 11.

Figure 11

Optimal design of Industrial refrigeration system.

6. Conclusions and Future Work

The main contributions of this research work can be summarized as follows:

  1. A modified mAEFA approach is developed by combining the original AEFA algorithm with Lévy flights, simulated annealing, and Adaptive s-best Mutation and Natural Survivor Method (NSM) mechanisms. The mAEFA utilizes Lévy flights distribution to enhance exploration and convergence, while the simulated annealing approach assists in avoiding local optima;

  2. The performance of the mAEFA algorithm is scrutinized on various benchmark functions and engineering design problems. Comparative analysis with popular metaheuristics demonstrates the superiority and competitiveness of mAEFA;

  3. In solving composite CEC’17 benchmark functions, mAEFA shows superior performance compared to other algorithms under comparison, except for specific functions. In unimodal, multimodal, and composite CEC’17 functions, mAEFA is either superior, equal, or ranked second among compared algorithms, with faster convergence observed;

  4. Box plot analysis and Wilcoxon rank-sum test results further validate the superiority and robustness of mAEFA over comparative algorithms in CEC’17 test functions;

  5. Performance testing of mAEFA on five engineering design problems demonstrates its superior capability to effectively solve practical optimization problems compared to other algorithms.

Nonetheless, similar to other metaheuristic algorithms, mAEFA faces limitations in solving all optimization problems across different domains, as highlighted by the No Free Lunch theorem. In future research, we aim to validate the effectiveness of the proposed mAEFA by testing it on more intricate continuous and discrete optimization problems. Moreover, there is potential to apply this algorithm to address various problems in diverse fields such as cloud task scheduling, image segmentation, engineering design, air quality prediction, finance, material science, and environmental sciences, among others.

In the future, a multiobjective version of the developed algorithm can be designed to tackle multiobjective problems. Furthermore, the authors intend to use the same combination to test other metaheuristic algorithms.

Author Contributions

Conceptualization, A.G.H. and F.A.H.; Methodology, A.G.H., S.K. and F.A.H.; Software, A.G.H. and F.A.H.; Validation, A.G.H. and A.P.; Formal analysis, A.G.H.; Investigation, A.G.H. and F.A.H.; Resources, A.G.H. and G.H.; Data curation, A.G.H. and G.H.; Writing—original draft, A.G.H., S.K., F.A.H., G.H. and A.P.; Writing—review & editing, A.G.H., A.P. and A.G.H.; Visualization, A.G.H. and F.A.H.; Supervision, A.P. and G.H.; Project administration, A.P. and G.H.; Funding acquisition, A.G.H. and A.P. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.García-Martínez C., Gutiérrez P.D., Molina D., Lozano M., Herrera F. Since CEC 2005 competition on real-parameter optimisation: A decade of research, progress and comparative analysis’s weakness. Soft Comput. 2017;21:5573–5583. doi: 10.1007/s00500-016-2471-9. [DOI] [Google Scholar]
  • 2.Mernik M., Liu S.H., Karaboga D., Črepinšek M. On clarifying misconceptions when comparing variants of the artificial bee colony algorithm by offering a new implementation. Inf. Sci. 2015;291:115–127. doi: 10.1016/j.ins.2014.08.040. [DOI] [Google Scholar]
  • 3.Abdullahi M., Ngadi M.A., Dishing S.I., Ahmad B.I.E. An efficient symbiotic organisms search algorithm with chaotic optimization strategy for multi-objective task scheduling problems in cloud computing environment. J. Netw. Comput. Appl. 2019;133:60–74. doi: 10.1016/j.jnca.2019.02.005. [DOI] [Google Scholar]
  • 4.Faris H., Mirjalili S., Aljarah I. Automatic selection of hidden neurons and weights in neural networks using grey wolf optimizer based on a hybrid encoding scheme. Int. J. Mach. Learn. Cybern. 2019;10:2901–2920. doi: 10.1007/s13042-018-00913-2. [DOI] [Google Scholar]
  • 5.Chen Y.P., Li Y., Wang G., Zheng Y.F., Xu Q., Fan J.H., Cui X.T. A novel bacterial foraging optimization algorithm for feature selection. Expert Syst. Appl. 2017;83:1–17. doi: 10.1016/j.eswa.2017.04.019. [DOI] [Google Scholar]
  • 6.Alweshah M., Khalaileh S.A., Gupta B.B., Almomani A., Hammouri A.I., Al-Betar M.A. The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput. Appl. 2020;34:11267–11281. doi: 10.1007/s00521-020-05210-0. [DOI] [Google Scholar]
  • 7.Fathi H., AlSalman H., Gumaei A., Manhrawy I.I., Hussien A.G., El-Kafrawy P. An Efficient Cancer Classification Model Using Microarray and High-Dimensional Data. Comput. Intell. Neurosci. 2021;2021:7231126. doi: 10.1155/2021/7231126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Kumar Y., Dahiya N., Malik S., Yadav G., Singh V. Natural Computing for Unsupervised Learning. Springer; Cham, Switzerland: 2019. Chemical Reaction-Based Optimization Algorithm for Solving Clustering Problems; pp. 147–162. [Google Scholar]
  • 9.Houssein E.H., Hosney M.E., Oliva D., Mohamed W.M., Hassaballah M. A novel hybrid Harris hawks optimization and support vector machines for drug design and discovery. Comput. Chem. Eng. 2020;133:106656. doi: 10.1016/j.compchemeng.2019.106656. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Bezdan T., Stoean C., Naamany A.A., Bacanin N., Rashid T.A., Zivkovic M., Venkatachalam K. Hybrid fruit-fly optimization algorithm with k-means for text document clustering. Mathematics. 2021;9:1929. doi: 10.3390/math9161929. [DOI] [Google Scholar]
  • 11.Besnassi M., Neggaz N., Benyettou A. Face detection based on evolutionary Haar filter. Pattern Anal. Appl. 2020;23:309–330. doi: 10.1007/s10044-019-00784-5. [DOI] [Google Scholar]
  • 12.Holland J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. MIT Press; Cambridge, MA, USA: 1992. [Google Scholar]
  • 13.Hansen N., Müller S.D., Koumoutsakos P. Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES) Evol. Comput. 2003;11:1–18. doi: 10.1162/106365603321828970. [DOI] [PubMed] [Google Scholar]
  • 14.Yang X.S., Deb S. Cuckoo search via Lévy flights; Proceedings of the 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC); Coimbatore, India. 9–11 December 2009; pp. 210–214. [Google Scholar]
  • 15.Geem Z.W., Kim J.H., Loganathan G.V. A new heuristic optimization algorithm: Harmony search. Simulation. 2001;76:60–68. doi: 10.1177/003754970107600201. [DOI] [Google Scholar]
  • 16.Karaboga D., Ozturk C. A novel clustering approach: Artificial Bee Colony (ABC) algorithm. Appl. Soft Comput. 2011;11:652–657. doi: 10.1016/j.asoc.2009.12.025. [DOI] [Google Scholar]
  • 17.Gandomi A.H., Alavi A.H. Krill herd: A new bio-inspired optimization algorithm. Commun. Nonlinear Sci. Numer. Simul. 2012;17:4831–4845. doi: 10.1016/j.cnsns.2012.05.010. [DOI] [Google Scholar]
  • 18.Assiri A.S., Hussien A.G., Amin M. Ant Lion Optimization: Variants, hybrids, and applications. IEEE Access. 2020;8:77746–77764. doi: 10.1109/ACCESS.2020.2990338. [DOI] [Google Scholar]
  • 19.Mirjalili S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015;89:228–249. doi: 10.1016/j.knosys.2015.07.006. [DOI] [Google Scholar]
  • 20.Askarzadeh A. A novel metaheuristic method for solving constrained engineering optimization problems: Crow search algorithm. Comput. Struct. 2016;169:1–12. doi: 10.1016/j.compstruc.2016.03.001. [DOI] [Google Scholar]
  • 21.Mirjalili S., Lewis A. The whale optimization algorithm. Adv. Eng. Softw. 2016;95:51–67. doi: 10.1016/j.advengsoft.2016.01.008. [DOI] [Google Scholar]
  • 22.Shareef H., Ibrahim A.A., Mutlag A.H. Lightning search algorithm. Appl. Soft Comput. 2015;36:315–333. doi: 10.1016/j.asoc.2015.07.028. [DOI] [Google Scholar]
  • 23.Mirjalili S., Gandomi A.H., Mirjalili S.Z., Saremi S., Faris H., Mirjalili S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017;114:163–191. doi: 10.1016/j.advengsoft.2017.07.002. [DOI] [Google Scholar]
  • 24.Heidari A.A., Mirjalili S., Faris H., Aljarah I., Mafarja M., Chen H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019;97:849–872. doi: 10.1016/j.future.2019.02.028. [DOI] [Google Scholar]
  • 25.Li M.D., Zhao H., Weng X.W., Han T. A novel nature-inspired algorithm for optimization: Virus colony search. Adv. Eng. Softw. 2016;92:65–88. doi: 10.1016/j.advengsoft.2015.11.004. [DOI] [Google Scholar]
  • 26.Hashim F.A., Hussien A.G. Snake Optimizer: A novel meta-heuristic optimization algorithm. Knowl.-Based Syst. 2022;242:108320. doi: 10.1016/j.knosys.2022.108320. [DOI] [Google Scholar]
  • 27.Yıldız B.S., Kumar S., Panagant N., Mehta P., Sait S.M., Yildiz A.R., Pholdee N., Bureerat S., Mirjalili S. A novel hybrid arithmetic optimization algorithm for solving constrained optimization problems. Knowl.-Based Syst. 2023;271:110554. doi: 10.1016/j.knosys.2023.110554. [DOI] [Google Scholar]
  • 28.Wolpert D.H., Macready W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997;1:67–82. doi: 10.1109/4235.585893. [DOI] [Google Scholar]
  • 29.Shayanfar H., Gharehchopogh F.S. Farmland fertility: A new metaheuristic algorithm for solving continuous optimization problems. Appl. Soft Comput. 2018;71:728–746. doi: 10.1016/j.asoc.2018.07.033. [DOI] [Google Scholar]
  • 30.Anosri S., Panagant N., Champasak P., Bureerat S., Thipyopas C., Kumar S., Pholdee N., Yıldız B.S., Yildiz A.R. A Comparative Study of State-of-the-art Metaheuristics for Solving Many-objective Optimization Problems of Fixed Wing Unmanned Aerial Vehicle Conceptual Design. Arch. Comput. Methods Eng. 2023;30:3657–3671. doi: 10.1007/s11831-023-09914-z. [DOI] [Google Scholar]
  • 31.Arora S., Singh H., Sharma M., Sharma S., Anand P. A new hybrid algorithm based on grey wolf optimization and crow search algorithm for unconstrained function optimization and feature selection. IEEE Access. 2019;7:26343–26361. doi: 10.1109/ACCESS.2019.2897325. [DOI] [Google Scholar]
  • 32.Kumar S., Tejani G.G., Pholdee N., Bureerat S., Mehta P. Hybrid heat transfer search and passing vehicle search optimizer for multi-objective structural optimization. Knowl.-Based Syst. 2021;212:106556. doi: 10.1016/j.knosys.2020.106556. [DOI] [Google Scholar]
  • 33.Yildiz B.S., Mehta P., Sait S.M., Panagant N., Kumar S., Yildiz A.R. A new hybrid artificial hummingbird-simulated annealing algorithm to solve constrained mechanical engineering problems. Mater. Test. 2022;64:1043–1050. doi: 10.1515/mt-2022-0123. [DOI] [Google Scholar]
  • 34.Kumar S., Tejani G.G., Mirjalili S. Modified symbiotic organisms search for structural optimization. Eng. Comput. 2019;35:1269–1296. doi: 10.1007/s00366-018-0662-y. [DOI] [Google Scholar]
  • 35.Houssein E.H., Mahdy M.A., Fathy A., Rezk H. A modified Marine Predator Algorithm based on opposition based learning for tracking the global MPP of shaded PV system. Expert Syst. Appl. 2021;183:115253. doi: 10.1016/j.eswa.2021.115253. [DOI] [Google Scholar]
  • 36.Deng W., Xu J., Zhao H. An improved ant colony optimization algorithm based on hybrid strategies for scheduling problem. IEEE Access. 2019;7:20281–20292. doi: 10.1109/ACCESS.2019.2897580. [DOI] [Google Scholar]
  • 37.Tubishat M., Idris N., Shuib L., Abushariah M.A., Mirjalili S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020;145:113122. doi: 10.1016/j.eswa.2019.113122. [DOI] [Google Scholar]
  • 38.Yadav A. AEFA: Artificial electric field algorithm for global optimization. Swarm Evol. Comput. 2019;48:93–108. [Google Scholar]
  • 39.Selem S.I., El-Fergany A.A., Hasanien H.M. Artificial electric field algorithm to extract nine parameters of triple-diode photovoltaic model. Int. J. Energy Res. 2021;45:590–604. doi: 10.1002/er.5756. [DOI] [Google Scholar]
  • 40.Houssein E.H., Hashim F.A., Ferahtia S., Rezk H. An efficient modified artificial electric field algorithm for solving optimization problems and parameter estimation of fuel cell. Int. J. Energy Res. 2021;45:20199–20218. doi: 10.1002/er.7103. [DOI] [Google Scholar]
  • 41.Anita, Yadav A., Kumar N. Application of artificial electric field algorithm for economic load dispatch problem; Proceedings of the International Conference on Soft Computing and Pattern Recognition; Hyderabad, India. 13–15 December 2019; Cham, Switzerland: Springer; 2019. pp. 71–79. [Google Scholar]
  • 42.Janjanam L., Saha S., Kar R., Mandal D. Volterra filter modelling of non-linear system using Artificial Electric Field algorithm assisted Kalman filter and its experimental evaluation. ISA Trans. 2020;125:614–630. doi: 10.1016/j.isatra.2020.09.010. [DOI] [PubMed] [Google Scholar]
  • 43.Anita, Yadav A., Kumar N., Kim J.H. Development of Discrete Artificial Electric Field Algorithm for Quadratic Assignment Problems; Proceedings of the International Conference on Harmony Search Algorithm; Istanbul, Turkey. 22–24 April 2020; Singapore: Springer; 2020. pp. 411–421. [Google Scholar]
  • 44.Cao M.T., Hoang N.D., Nhu V.H., Bui D.T. An advanced meta-learner based on artificial electric field algorithm optimized stacking ensemble techniques for enhancing prediction accuracy of soil shear strength. Eng. Comput. 2020;38:2185–2207. doi: 10.1007/s00366-020-01116-6. [DOI] [Google Scholar]
  • 45.Naderipour A., Abdul-Malek Z., Mustafa M.W.B., Guerrero J.M. A multi-objective artificial electric field optimization algorithm for allocation of wind turbines in distribution systems. Appl. Soft Comput. 2021;105:107278. doi: 10.1016/j.asoc.2021.107278. [DOI] [Google Scholar]
  • 46.Izci D., Ekinci S., Orenc S., Demirören A. Improved artificial electric field algorithm using Nelder-Mead simplex method for optimization problems; Proceedings of the 2020 4th International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT); Istanbul, Turkey. 22–24 October 2020; pp. 1–5. [Google Scholar]
  • 47.Petwal H., Rani R. An Improved Artificial Electric Field Algorithm for Multi-Objective Optimization. Processes. 2020;8:584. doi: 10.3390/pr8050584. [DOI] [Google Scholar]
  • 48.Yadav A., Kumar N. Artificial electric field algorithm for engineering optimization problems. Expert Syst. Appl. 2020;149:113308. [Google Scholar]
  • 49.Yadav A. Discrete artificial electric field algorithm for high-order graph matching. Appl. Soft Comput. 2020;92:106260. [Google Scholar]
  • 50.Sajwan A., Yadav A. A study of exploratory and stability analysis of artificial electric field algorithm. Appl. Intell. 2022;52:10805–10828. doi: 10.1007/s10489-021-02865-7. [DOI] [Google Scholar]
  • 51.Kahraman H.T., Katı M., Aras S., Taşci D.A. Development of the Natural Survivor Method (NSM) for designing an updating mechanism in metaheuristic search algorithms. Eng. Appl. Artif. Intell. 2023;122:106121. doi: 10.1016/j.engappai.2023.106121. [DOI] [Google Scholar]
  • 52.Hertz A., Werra D.d. Using tabu search techniques for graph coloring. Computing. 1987;39:345–351. doi: 10.1007/BF02239976. [DOI] [Google Scholar]
  • 53.Ekinci S., Izci D. Enhanced reptile search algorithm with Lévy flight for vehicle cruise control system design. Evol. Intell. 2023;16:1339–1351. doi: 10.1007/s12065-022-00745-8. [DOI] [Google Scholar]
  • 54.Syama S., Ramprabhakar J., Anand R., Guerrero J.M. A hybrid extreme learning machine model with lévy flight chaotic whale optimization algorithm for wind speed forecasting. Results Eng. 2023;19:101274. doi: 10.1016/j.rineng.2023.101274. [DOI] [Google Scholar]
  • 55.He Q., Liu H., Ding G., Tu L. A modified Lévy flight distribution for solving high-dimensional numerical optimization problems. Math. Comput. Simul. 2023;204:376–400. doi: 10.1016/j.matcom.2022.08.017. [DOI] [Google Scholar]
  • 56.Zhang B., Liu W., Cai Y., Zhou Z., Wang L., Liao Q., Fu Z., Cheng Z. State of health prediction of lithium-ion batteries using particle swarm optimization with Levy flight and generalized opposition-based learning. J. Energy Storage. 2024;84:110816. doi: 10.1016/j.est.2024.110816. [DOI] [Google Scholar]
  • 57.Hussien A.M., Hasanien H.M., Qais M.H., Alghuwainem S. Hybrid Transient Search Algorithm with Levy Flight for Optimal PI Controllers of Islanded Microgrids. IEEE Access. 2024;12:15075–15092. doi: 10.1109/ACCESS.2024.3357741. [DOI] [Google Scholar]
  • 58.Barua S., Merabet A. Lévy Arithmetic Algorithm: An enhanced metaheuristic algorithm and its application to engineering optimization. Expert Syst. Appl. 2024;241:122335. doi: 10.1016/j.eswa.2023.122335. [DOI] [Google Scholar]
  • 59.Pashaei E., Pashaei E. Hybrid binary COOT algorithm with simulated annealing for feature selection in high-dimensional microarray data. Neural Comput. Appl. 2023;35:353–374. doi: 10.1007/s00521-022-07780-7. [DOI] [Google Scholar]
  • 60.Xu R., Zhao C., Li J., Hu J., Hou X. A hybrid improved-whale-optimization–simulated-annealing algorithm for trajectory planning of quadruped robots. Electronics. 2023;12:1564. doi: 10.3390/electronics12071564. [DOI] [Google Scholar]
  • 61.Fontes D.B., Homayouni S.M., Gonçalves J.F. A hybrid particle swarm optimization and simulated annealing algorithm for the job shop scheduling problem with transport resources. Eur. J. Oper. Res. 2023;306:1140–1157. doi: 10.1016/j.ejor.2022.09.006. [DOI] [Google Scholar]
  • 62.Sajjad F., Rashid M., Zafar A., Zafar K., Fida B., Arshad A., Riaz S., Dutta A.K., Rodrigues J.J. An efficient hybrid approach for optimization using simulated annealing and grasshopper algorithm for IoT applications. Discov. Internet Things. 2023;3:7. doi: 10.1007/s43926-023-00036-3. [DOI] [Google Scholar]
  • 63.Agushaka J.O., Ezugwu A.E., Abualigah L., Alharbi S.K., Khalifa H.A.E.W. Efficient initialization methods for population-based metaheuristic algorithms: A comparative study. Arch. Comput. Methods Eng. 2023;30:1727–1787. doi: 10.1007/s11831-022-09850-4. [DOI] [Google Scholar]
  • 64.Kazimipour B., Li X., Qin A.K. A review of population initialization techniques for evolutionary algorithms; Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC); Beijing, China. 6–11 July 2014; pp. 2585–2592. [Google Scholar]
  • 65.Tharwat A., Schenck W. Population initialization techniques for evolutionary algorithms for single-objective constrained optimization problems: Deterministic vs. stochastic techniques. Swarm Evol. Comput. 2021;67:100952. doi: 10.1016/j.swevo.2021.100952. [DOI] [Google Scholar]
  • 66.Kahraman H.T., Aras S., Gedikli E. Fitness-distance balance (FDB): A new selection method for meta-heuristic search algorithms. Knowl.-Based Syst. 2020;190:105169. doi: 10.1016/j.knosys.2019.105169. [DOI] [Google Scholar]
  • 67.Ozkaya B., Kahraman H.T., Duman S., Guvenc U. Fitness-Distance-Constraint (FDC) based guide selection method for constrained optimization problems. Appl. Soft Comput. 2023;144:110479. doi: 10.1016/j.asoc.2023.110479. [DOI] [Google Scholar]
  • 68.Duman S., Kahraman H.T., Kati M. Economical operation of modern power grids incorporating uncertainties of renewable energy sources and load demand using the adaptive fitness-distance balance-based stochastic fractal search algorithm. Eng. Appl. Artif. Intell. 2023;117:105501. doi: 10.1016/j.engappai.2022.105501. [DOI] [Google Scholar]
  • 69.Kahraman H.T., Bakir H., Duman S., Katı M., Aras S., Guvenc U. Dynamic FDB selection method and its application: Modeling and optimizing of directional overcurrent relays coordination. Appl. Intell. 2022;52:4873–4908. doi: 10.1007/s10489-021-02629-3. [DOI] [Google Scholar]
  • 70.Kirkpatrick S., Gelatt C.D., Jr., Vecchi M.P. Optimization by simulated annealing. Science. 1983;220:671–680. doi: 10.1126/science.220.4598.671. [DOI] [PubMed] [Google Scholar]
  • 71.Wu G., Mallipeddi R., Suganthan P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real-Parameter Optimization. National University of Defense Technology; Changsha, China: Kyungpook National University; Daegu, Republic of Korea: Nanyang Technological University; Singapore: 2017. Technical Report. [Google Scholar]
  • 72.Al-Betar M.A., Alyasseri Z.A.A., Awadallah M.A., Abu Doush I. Coronavirus herd immunity optimizer (CHIO) Neural Comput. Appl. 2021;33:5011–5042. doi: 10.1007/s00521-020-05296-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Rashedi E., Nezamabadi-Pour H., Saryazdi S. GSA: A gravitational search algorithm. Inf. Sci. 2009;179:2232–2248. doi: 10.1016/j.ins.2009.03.004. [DOI] [Google Scholar]
  • 74.Salawudeen A.T., Mu’azu M.B., Yusuf A., Adedokun A.E. A Novel Smell Agent Optimization (SAO): An extensive CEC study and engineering application. Knowl.-Based Syst. 2021;232:107486. doi: 10.1016/j.knosys.2021.107486. [DOI] [Google Scholar]
  • 75.Saremi S., Mirjalili S., Lewis A. Grasshopper optimisation algorithm: Theory and application. Adv. Eng. Softw. 2017;105:30–47. doi: 10.1016/j.advengsoft.2017.01.004. [DOI] [Google Scholar]
  • 76.Kennedy J., Eberhart R. Particle swarm optimization; Proceedings of the ICNN’95-International Conference on Neural Networks; Perth, WA, Australia. 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  • 77.Houssein E.H., Oliva D., Samee N.A., Mahmoud N.F., Emam M.M. Liver Cancer Algorithm: A novel bio-inspired optimizer. Comput. Biol. Med. 2023;165:107389. doi: 10.1016/j.compbiomed.2023.107389. [DOI] [PubMed] [Google Scholar]
  • 78.Mezura-Montes E., Coello C.A.C. Useful infeasible solutions in engineering optimization with evolutionary algorithms; Proceedings of the Mexican International Conference on Artificial Intelligence; Monterrey, Mexico. 14–18 November 2005; Berlin/Heidelberg, Germany: Springer; 2005. pp. 652–662. [Google Scholar]
  • 79.Kannan B., Kramer S.N. An augmented Lagrange multiplier based method for mixed integer discrete continuous optimization and its applications to mechanical design. J. Mech. Des. 1994;116:405–411. doi: 10.1115/1.2919393. [DOI] [Google Scholar]
  • 80.Hashim F.A., Hussain K., Houssein E.H., Mabrouk M.S., Al-Atabany W. Archimedes optimization algorithm: A new metaheuristic algorithm for solving optimization problems. Appl. Intell. 2021;51:1531–1551. doi: 10.1007/s10489-020-01893-z. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data are available on request.


Articles from Biomimetics are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES