Sie können Operatoren mit Ihrer Suchanfrage kombinieren, um diese noch präziser einzugrenzen. Klicken Sie auf den Suchoperator, um eine Erklärung seiner Funktionsweise anzuzeigen.
Findet Dokumente, in denen beide Begriffe in beliebiger Reihenfolge innerhalb von maximal n Worten zueinander stehen. Empfehlung: Wählen Sie zwischen 15 und 30 als maximale Wortanzahl (z.B. NEAR(hybrid, antrieb, 20)).
Findet Dokumente, in denen der Begriff in Wortvarianten vorkommt, wobei diese VOR, HINTER oder VOR und HINTER dem Suchbegriff anschließen können (z.B., leichtbau*, *leichtbau, *leichtbau*).
Der Elefantenherdenoptimierungsalgorithmus (EHO) ist ein neuartiger metaheuristischer Optimierer, inspiriert durch die Erneuerung und Trennung von Elefantenpopulationen. Obwohl er nur wenige Parameter hat und leicht umzusetzen ist, leidet er unter einem Mangel an Ausbeutung, was zu langsamer Konvergenz führt. Dieser Aufsatz schlägt einen verbesserten EHO-Algorithmus vor, der als Manta-Rochen-Futtersuche und Gaußsche Mutation-basierte EHO für globale Optimierung (MGEHO) bezeichnet wird. Der Clan-aktualisierende Operator im ursprünglichen EHO-Algorithmus wird durch die Somersault-Foraging-Strategie von Mantelrochen ersetzt, die darauf abzielt, Patriarchenpositionen optimal anzupassen. Zusätzlich wird ein dynamischer Konvergenzfaktor gesetzt, um Exploration und Ausbeutung ins Gleichgewicht zu bringen. Die Gaußsche Mutation wird übernommen, um die Populationsdiversität zu verbessern, was MGEHO in die Lage versetzt, eine starke lokale Suchkapazität aufrechtzuerhalten.
Mit KI übersetzt
Abstract
The elephant herding optimization (EHO) algorithm is a novel metaheuristic optimizer inspired by the clan renewal and separation behaviors of elephant populations. Although it has few parameters and is easy to implement, it suffers from a lack of exploitation, leading to slow convergence. This paper proposes an improved EHO algorithm called manta ray foraging and Gaussian mutation-based EHO for global optimization (MGEHO). The clan updating operator in the original EHO algorithm is replaced by the somersault foraging strategy of manta rays, which aims to optimally adjust patriarch positions. Additionally, a dynamic convergence factor is set to balance exploration and exploitation. The gaussian mutation is adopted to enhance the population diversity, enabling MGEHO to maintain a strong local search capability. To evaluate the performances of different algorithms, 33 classical benchmark functions are chosen to verify the superiority of MGEHO. Also, the enhanced paradigm is compared with other advanced metaheuristic algorithms on 32 benchmark functions from IEEE CEC2014 and CEC2017. Furthermore, a scalability test, convergence analysis, statistical analysis, diversity analysis, and running time analysis demonstrate the effectiveness of MGEHO from various aspects. The results illustrate that MGEHO is superior to other algorithms in terms of solution accuracy and stability. Finally, MGEHO is applied to solve three real engineering problems. The comparison results show that this method is a powerful auxiliary tool for handling complex problems.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
1 Introduction
Optimization has long been a fundamental tool for applications in mathematics, medicine, engineering, the military, and many other fields. In general, optimization is an approach used to determine the best performance of a system that can be achieved based on optimized parameters to make the final solution as perfect and efficient as possible [1]. In optimization problems, the final goal is usually considered the maximization or minimization of some objective function (the function used to evaluate the nature or quality of the resulting solution). In this case, several steps are involved, such as defining a mathematical model for the system, identifying variable, defining the type of system needed, specifying constraints, and then finding the optimization response [2].
Traditional optimization techniques include the dynamic programming approach (DPA), stochastic search, the steepest descent method, and Newton’s method. Most of these methods are nonpolynomial and can find optimal solutions smoothly. Unfortunately, they are usually limited by the specific problem and the size of the given dataset. Furthermore, they are often highly complex and specialized in nature. In addition, the execution times and running costs are very high [3]. Therefore, such approaches are only suitable for solving small-scale problems. Currently, most real-world optimization problems are complex, highly nonlinear, and large-scale spatial, and the information metrics used are essentially multifaceted. On this basis, traditional methods are not suitable for such situations. The popular trend is to combine the multifaceted nature of the subject system with various proposed heuristics to solve difficult problems. Heuristics methods include nearest neighbor algorithms, evolutionary algorithms, insertion algorithms, memetic algorithms, dynamic relaxation, etc. By introducing greedy mechanisms, the number of search steps is fixed, and the number of iterations is reduced. With these advantages, heuristic methods are able to find near-optimal solutions with less time costs than those of other methods [4]. However, most of them are designed for specific problems. Due to the introduction of greedy mechanisms, these methods often fall into local optima when solving complex problems. Soon after the development of heuristic methods, metaheuristics, which are problem-independent adaptive strategies, emerged to provide a set of guidelines or strategies for developing heuristics. In this term, “meta” means using a higher level, usually one level higher than that of the given heuristic, to solve a variety of intractable and complex hard problems [5]. Specifically, metaheuristics perform the search process by using multiple agents, which are essentially systems that find the best solution based on constraints and rules over multiple iterations. The iterative process is stopped when a predefined criterion is satisfied, at which point the solution found is determined to be the best solution and the system reaches a fused state. Metaheuristics are not dependent on gradient information, are easy to execute and are highly flexible because they treat the optimization process as a black box [6]. As research has progressed, metaheuristics have been developed as powerful tools and are now widely used in various fields.
Anzeige
In general, metaheuristic algorithms are classified into four categories, namely, evolutionary-based algorithms, physics-based algorithms, population-based algorithms, and human-based algorithms. Inspired by Darwinian evolutionary theory, evolutionary-based algorithms were introduced in the early 1970s and usually draw from natural rules to find optimal solutions. By setting one or more operators (crossover, variation, or selection operators), such algorithms evaluate search agents during optimization [7]. Evolutionary-based algorithms include genetic algorithms (GA) [8], biogeography-based optimization algorithms (BBO) [9], differential evolutionary algorithms (DE) [10], etc. Physics-based algorithms are modelled according to physical rules. For example, the gravitational search algorithm (GSA) [11] and the equilibrium optimizer (EO) [12] are all physics-based approaches. Population-based algorithms simulate the group behaviour of animals, performing hunting based on cooperation. In population-based algorithms, multiple agents perform the search process together and share useful information with each other, thus improving operational efficiency. Representative algorithms include the artificial bee colony (ABC) algorithm [13], the sailfish optimizer (SFO) [14], whale optimization algorithm (WOA) [15], Harris hawks algorithm (HHO) [16], etc. Human-based algorithms generally simulate the special behaviour of humans. Such algorithms include the harmony search algorithm (HS) [17], teaching-learning-based optimization (TLBO) [18].
Elephant herding optimization (EHO) is a population-based algorithm proposed by Wang et al. [19] . Inspired by the nomadic herding behaviour of elephant populations, EHO is designed to solve global optimization problems. In EHO, an entire population is divided into different clans. Each clan is led by a female matriarch, while the male elephants live independently when they reach adulthood. EHO sets up two operators, a clan updating operator and a separation operator, to update the positions of the agents. A clan updating operator is implemented for each individual, and the next position of each individual is updated according to the current position in relation to the position of the clan leader. Subsequently, the agent with the worst position is replaced by the separation operator to speed up convergence and enhance the diversity of the population. In terms of advantages, EHO has fewer controlled parameters, is easier to implement, and has better performance than other optimization approaches, so it attracts much attention from mathematicians and engineers [20].
In response to the shortcomings of EHO, researchers have proposed many improvement ideas and optimization methods. For instance, Muthusamy et al. [21] introduced the sine cosine algorithm (SCA) and opposition-based learning (OBL) to update the positions of agents in the original EHO algorithm. Through this operation, the convergence speed and the searchability of the algorithm were enhanced. Li and Wang [20] proposed a new improved EHO approach based on dynamic topology and BBO methods. The operators of this EHO variant were updated, and a new separation operator was set, thus ensuring that the evolutionary process of the examined population moved in a better direction. Ismaeel et al. [22] proposed three versions of an unbiased algorithm, EEHO15, EEHO20, and EEHO25, to overcome the defect of unreasonable convergence in the original EHO algorithm. Three similar variants were proposed by Elhosseini et al. [23], namely cultural-based, alpha-tuning, and biased initialization EHO, with the aim of strengthening the global optimization capability of EHO. Balamurugan et al. [24] proposed an opposition-based EHO (OEHO) algorithm. In their work, opposition-based learning was adopted to enhance the performance of EHO. Xu et al. [25] proposed a novel algorithm, LFEHO, which combined Levy flight with EHO to overcome the problems of the original EHO algorithm falling into local optima and exhibiting poor convergence performance. Improved EHO algorithms such as IMEHO were presented by Li et al. [26] The robustness and diversity of a given population are optimized by using a new global learning-based evolution strategy to update the velocities and positions of agents. Reference [27] extended and improved the original EHO algorithm by designing six strategies to update the agents. The information obtained from the previous iterations of individuals was adopted for weighing operations. Therefore, the obtained results were better than those output by the original EHO algorithm. In addition, engineers extended the application of EHO to solve practical engineering problems in real production cycles.
In addition, MRFO is a novel optimization algorithm proposed by Zhao et al. [28]. It is a promising and powerful optimizer inspired by the predatory behaviour of manta rays. MRFO has three foraging operators, namely, chain foraging, cyclone foraging, and tumble foraging operators. Recently, researchers have introduced MRFO to other fields. For example, Ekinci et al. [29] combined the opposition-based (OBL) with MRFO and hybrid simulated annealing algorithm to provides MRFO a better exploration capability. Shaheen et al. [30] optimized the MRFO algorithm by using adaptive penalty parameters and introduced it to handle a thermal scheduling problem. In addition, MRFO has been extended to many fields, such as feature selection (FS) [31], fuel power generation [32], system identification [33], photovoltaic system operation [34], and other areas.
Anzeige
According to the no-free lunch (NFL) theorem [35], any algorithm that is applicable for solving a given problem cannot provide the best solution for all problems. In other words, there is a need to improve the metaheuristic algorithms that exist in reality. Maintaining a balance between exploration and exploitation is the common and fundamental feature of any optimization algorithm and the most challenging problem in metaheuristics. Among them, exploration is the process of exploring optimal solutions over a wide range of unknown regions, while the exploitation of explored regions allows an algorithm to effectively find high-quality solutions. Therefore, it is necessary to fine-tune exploration and exploitation, as they are the two main factors used to obtain satisfactory results. In addition, it has been shown that although some algorithms are very effective in solving certain types of problems or large-scale problems, more significant optimization results can be achieved by combining the advantages of different algorithms in a hybrid manner [36]. By combining them, the convergence performance of the resulting hybrid algorithm is improved and the balance between exploration and exploitation is better maintained, thus allowing the algorithm to avoid falling into local optima. The above facts have motivated researchers to continuously develop hybrid metaheuristic algorithms with good exploration and exploitation capabilities and have made our research highly relevant.
Like most metaheuristic algorithms, EHO is prone to falling into local optima when dealing with complex multipeaked problems, which stems from the stochasticity of such problems and the imbalance between exploration and exploitation. Therefore, these shortcomings become major issues that affect the performance of EHO. At the same time, EHO also has drawbacks such as a poor exploitation capability and a low convergence rate. To summarize, there are three defects in the EHO algorithm that need to be solved.
(a)
The position of the female matriarch is determined by all members of the clan. Under this circumstance, once the algorithm falls into a local optimum, the matriarch will lose her ability to explore.
(b)
The location update strategy for individual elephants is fixed and lacks a certain control mechanism for adjustments according to the iterative process. Therefore, it is difficult to strike a balance between exploration and exploitation when dealing with complex problems.
(c)
In addition, the separation operator does not guarantee that the worst individual in the clan is updated to a better position. To address the shortcomings of EHO, a manta ray foraging and Gaussian mutation-based EHO algorithm for global optimization called MGEHO is proposed. The contributions of this paper are summarized as follows:
1.
The position of search agents in EHO is optimized by hybridizing the tumbling foraging strategy in MRFO, which improves the ability of the original EHO algorithm to jump out of local optima. Additionally, the global optimization ability of the original algorithm is enhanced.
2.
To maintain a balance between exploration and exploitation, a nonlinear dynamic convergence factor is set to improve the clan updating operator in EHO and the Somersault foraging in MRFO. On this basis, the global exploration and local exploitation processes can be coordinated.
3.
Furthermore, Gaussian mutation is introduced to adjust the position of the worst individual of each clan. By using random numbers conforming to a Gaussian distribution instead of the original model parameters, new solutions in the vicinity of the candidate solutions are generated, and the diversity of the population is enhanced.
4.
To fully validate its effectiveness, MGEHO is tested on 33 benchmark functions and compared with 9 other well-known metaheuristics. Furthermore, performance tests are conducted on the algorithm in different dimensions. Moreover, nonparametric statistical tests are set up to highlight the differences between MGEHO and other competing approaches. A paired Wilcoxon symbolic rank test and a Friedman test at the 5\(\%\) significance level are invoked to test the compared algorithms. Besides, diversity tests of MGEHO and EHO are performed to verify the improved performance on exploration and exploitation.
5.
The performance of the proposed MGEHO is also evaluated and compared with selected advanced algorithms by using 32 modern CEC2014 and CEC2017 benchmark functions. Statistical tests are invoked to analyze the results of means and standard deviations. The Holm test is used as a post hoc test. In addition, the running time analysis is conducted to highlight the fundamental differences between the proposed and existing methods.
6.
In addition, the proposed MGEHO algorithm is used to solve three practical engineering problems: (a) a design problem regarding a gear train; (b) a design problem with a three-bar truss; and (c) a design problem with respect to a pressure vessel. The experimental results show that MGEHO has superior adaptability and the ability to handle optimization problems.
The rest of the paper is organized as follows: Sect. 2 reviews the original EHO and MRFO algorithms. Section 3 provides a detail of the proposed MGEHO method. Experiments and the analysis of the results are given in Sect. 4. Tests on practical engineering problems are conducted in Sect. 5. Section 6 summarizes the conclusions of this paper and plans for future work.
2 Preliminary
2.1 EHO
In general, elephants are gregarious creatures. Within a population, they are divided into different clans. Individual elephants from the same clan live together under the leadership of a female matriarch. After reaching adulthood, a fixed number of males leave the clan. EHO simulates these characteristics of elephant herd life by dividing the behaviours of searching agents into two parts, namely, a clan updating operator and separation operator.
2.1.1 Clan updating operator
The position of an elephant individual j in clan \({C_i}\) is updated in the following way:
where \({x_{\text {new},{c_i},j}}\) and \({x_{{c_i},j}}\) are the new and original positions of elephant individual j in clan \({C_i}\), respectively. \(\alpha\) and r are both random numbers within [0,1]. In addition, \({x_{\text {best},{c_i}}}\) is the best individual of clan \({C_i}\), which is updated by the following equation:
where \({x_{\text {center},{c_i}}}\) represents the central positions of all individuals in clan \({C_i}\) and \({n_{{c_i}}}\) stands for the number of elephants in clan \({C_i}\). \(\beta\) affects the extent to which \({x_{\text {center},{c_i}}}\) acts on \({x_{\text {new},{c_i},j}}\), \(\beta \in [0,1]\).
2.1.2 Separation operator
A separation operator is proposed to update the worst individual in each clan by the following equation:
In MRFO, three manta ray foraging strategies are simulated, namely chain foraging, cyclone foraging, and somersault foraging. Similar to other metaheuristics, MRFO also performs initialization first and then updates each agent sequentially through iteration to find the best solution.
2.2.1 Chain foraging
In the first stage, manta rays form a foraging chain sequentially from front to back, and the position of each search agent is updated by the current optimal solution and the solutions generated during previous iterations. For a set of individuals \({x_i}\), the process of position updating is represented as follows:
where \(x_i^d(t)\) represents the position of ith individual at time t in the dth dimension. r represents a random number within [0,1]. N is the size of the population. \(x_{\text {best}}^d\) is the \(\alpha\) stands for a weight coefficient determined in the following way:
During this stage, manta rays swim forward in a spiral pattern. Unlike the spiral foraging approach in the WOA, each individual not only approaches its prey along the spiral path but also adjusts to the positions of the current and previous best agents and keeps swimming forward. The mathematical model for this process is represented as follows:
where \({r_1}\) is a random number, \({r_1} \in [0,1]\). T denotes the maximum number of iterations. In addition, agents at this stage update their positions by searching for random positions in the space to enhance their exploration abilities in the following manner:
where \(x_{\text {rand}}^d\) denotes a random position in dimension d and \(x_{\text {rand}}^d = x_{\min }^d + r \times (x_{\max }^d - x_{\min }^d)\). \({x_{\max }}\) and \({x_{\min }}\) represent the upper and lower bounds of the search space, respectively.
2.2.3 Somersault foraging
In somersault foraging, the MRFO algorithm treats food as a hub, and each agent makes a flipping motion along the hub. On this basis, agents can update their positions around the current optimal solution to enhance their exploitation capabilities. The mathematical model of this process is shown as follows:
$$\begin{aligned}&x_i^d(t + 1) = x_i^d(t) + S \times [{r_2} \times x_{\text {best}}^d(t) - {r_3} \times x_i^d(t)]\nonumber \\&\quad i = 1,2, \ldots \ldots N, \end{aligned}$$
(10)
where S is the somersault factor, which is used to indicate the range of each flip. In MRFO, the parameter is set to 2. \({r_2}\) and \({r_3}\) are both random numbers within [0,1].
In EHO, the position update strategies of the clan leader and other members are different. Specifically, the position of the patriarch is determined by the action factor \(\beta\) and the central positions of all agents in the same clan during each iteration. When \(\beta\) is small, the patriarch moves near the origin. When \(\beta\) is large, the patriarch suddenly moves to the center of the clan. This kind of strategy lacks guidance towards the global optimal solution, so the algorithm may fall into a local optimum, which affects its convergence efficiency and reduces its exploration ability. Second, the search process of EHO is fixed and lacks an effective way to control the balance between exploration and exploitation, and this largely affects the performance of the algorithm. Third, the worst individuals in the clan are randomly replaced by the separation operator, which makes it difficult to ensure the improvement of population diversity.
Therefore, MGEHO is proposed for the purpose of improving the above shortcomings of EHO. In this section, three aspects of the newly developed strategy are explained in detail.
3.1 Hybrid MRFO with EHO
First, we update the clan updating operator in EHO. Specifically, the somersault foraging strategy is introduced to update the position of the patriarch in each clan, as shown in Fig. 1. According to this novel behaviour, the patriarch can be updated to a position symmetric to the pivot point with the global optimal solution as the pivot point. As the distance between the position of each agent and that of the optimal solution decreases, the fluctuation of the current position decreases, and the range of somersault foraging decreases. Unlike opposition-based learning, somersault foraging is centered around the global optimal solution, so it is more directional, and the convergence rate of the algorithm is further enhanced. An agent that falls into a local optimum can be compared with its updated agent so that it can jump out of the local optimum.
Second, after finishing the clan updating and separation operations, chain foraging and cyclone foraging are executed. Later, the elite agents are selected for retention by comparing the fitness values of the agents generated in the ith iteration.
3.2 Nonlinear convergence factor
In the original EHO algorithm, there is a lack of effective control over diversity and convergence, which limits the ability of the algorithm to find the best convergence rate. For this reason, a nonlinear convergence factor is introduced that can change dynamically with the iterative process. In the early iterations of the algorithm, the weights are relatively large, which can enhance the global search ability and diversity of the population. In later iterations, the weights decrease so that the algorithm can focus on the explored areas, improve the local search ability, and accelerate the convergence rate. The convergence factor is expressed as follows:
where \({\omega _u}\) and \({\omega _l}\) are the upper and lower bounds of the convergence factor, respectively. k is the impact factor. When k takes different values, the corresponding changes in the convergence factor are shown in Fig. 2. In the early iterations, the convergence factor is larger and scales down more, which will allow the population to perform a broad exploration. In later iterations, the convergence factor decreases and slows, which allows MGEHO to perform a fine local search. As observed in Fig. 2, when the impact factor k changes, \(\omega\) decreases in a nonlinear manner, and the rate of decrease gradually accelerates as the impact factor increases. The choice of the influencing factor depends on the choices and requirements of the specific problem. On this basis, the position of elephant individual j in clan \({C_i}\) is updated in the following way:
Gaussian mutation [37] was proposed by Back and Schwefe to improve the search efficiency of heuristic algorithms. Generally, Gaussian mutation generates new solutions in the vicinity of the candidate solutions. It can explore every corner of the search space in small steps during the execution of the search, thus enhancing the population diversity. As a successful strategy, it has been widely used in algorithms such as the GA and PSO [38]. When performing the mutation operation, the original values are replaced with eligible random numbers as the mean and variance of a Gaussian distribution [39]. The expression of the Gaussian distribution function is as follows:
where \({\sigma ^2}\) is the variance value of the candidate solutions. \(\delta\) represents the mean value. After applying a Gaussian distribution to MGEHO, the worst individual position of each clan is updated to explore a larger spatial region and increase the diversity of the candidate solutions. The equation for applying the Gaussian distribution is shown as follows:
One of the main measures of the optimization algorithm is the execution time. The modification of MGEHO consists of the following three main parts: mixture of MRFO’s operators, adaptive inertia weights, and Gaussian variants. When the population size N, dimension of the optimization problem and maximum number of iterations are set to N, D and T, respectively, the complexity of MGEHO can be calculated by the following way.
In MGEHO, the time used to initialize N search agents is \(O(N*D)\). The time to execute the clan-updating operator with adaptive inertia weights is \(O(N*D)\). Next, the time to execute the separating operator with Gaussian variation is O(N). Finally, the number of iterations is reduced because the MRFO operator is embedded within each iteration and no additional iterations are needed to be performed again. The time required to execute the MRFO operator is \(O(N*D)\). When T iterations are performed, the total time complexity of MGEHO is \(O(N*D+T(N*D+N+N*D)) =O (N*(D+T+ 2*T*D))\), which is higher than the \(O(T*N*D)\) of the basic EHO.
4 Experimental results
In this section, experimental results are presented. To verify the effectiveness and applicability of MGEHO, experiments are conducted on 33 classical benchmark functions, 32 CEC2014 and CEC2017 benchkmark functions. To ensure the fairness of the experiments, each algorithm is run 30 times independently on the benchmark functions to maintain stability. In addition, the population size N for all algorithms is set to 30, and the maximum number of iterations \({t_{\max }}\) is set to 500. Statistical tests are conducted to examine the differences between the different algorithms. The environmental conditions are an Intel (R) Core (TM) i5-9300H CPU @ 2.40 GHz, 16 GB of RAM, and the Windows 10 operating system, and the simulation experiments are set up on the MATLAB R2019b platform.
4.1 Tests on 33 benchmark functions
4.1.1 Benchmark functions
In this subsection, 33 classical benchmark functions from reference [40] are selected, and the basic information regarding these functions is shown in Tables 1, 2 and 3. Among them, F1–F10 are single-peaked functions that have only one global optimal solution in the global range and can test the convergence efficiency and exploration capability of each algorithm. F11–F23 are high-dimensional multipeaked functions that have multiple local extremes in the search space and are used to test the ability of each algorithm to search globally and jump out of local optima. F24–F33 are fixed-dimensional multipeak functions that are used to test the ability of each algorithm to maintain a balance between exploration and exploitation.
The selection of the parameters can make an impact on the algorithm. \(\omega _l,\omega _u\) and k are three controlled parameters in MGEHO. In this subsection, value tests are performed to verify the effects of different parameters. During the designment of MGEHO, various values of these parameters are set for this experiment. It is worth noting that the interval of \(\omega _l,\omega _u\) and k is continuous and has numerous values. Here, \(\left[ {{\omega _l},{\omega _u}} \right]\) is defined as \(\left\{ {\left[ {0.1,0.7} \right] ,\left[ {0.2,0.8} \right] ,\left[ {0.3,0.9} \right] } \right\}\) and \(k={1,2,3}\). Since there are 3 values for each of the 2 sets of parameters, the full-order factor will have \(3*3=9\) design combinations. The 9 MGEHO variants are named after MGEHO1 to MGEHO9, and the settings of variants are shown in Table 4. Two functions, namely F5, F9, F11, F17, F29, F30, are employed from each category of unimodal, multimodal and fixed dimension multimodal functions The original EHO is selected for comparison with the other 9 different variants. Each algorithm is averaged over 30 independent runs of the fitness of the functions.
Table 4
Combination of MGEHO’s parameters
Parameters
Range
Algorithms
MGEHO1
MGEHO2
MGEHO3
MGEHO4
MGEHO5
MGEHO6
MGEHO7
MGEHO8
MGEHO9
\(\left[ {{\omega _l},{\omega _u}} \right]\)
[0.1,0.7]
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
[0.2,0.8]
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
[0.3,0.9]
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
k
1
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
2
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
3
\(\checkmark\)
\(\checkmark\)
\(\checkmark\)
Table 5
Comparison results of EHO with 9 variants of MGEHO
Function
EHO1
EHO2
EHO3
EHO4
EHO5
EHO6
EHO7
EHO8
EHO9
EHO
f1
Avg
1.17E\(-\)240
8.79E\(-\)308
0.00E\(+\)00
2.03E\(-\)191
6.53E\(-\)251
4.09E\(-\)279
7.26E\(-\)156
5.37E\(-\)206
1.27E\(-\)229
9.67E\(-\)04
Std
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
3.40E\(-\)155
0.00E\(+\)00
0.00E\(+\)00
7.89E\(-\)04
Rank
5
2
1
8
4
3
9
7
6
10
p value
1.21E\(-\)12
6.58E\(-\)05
NaN
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
f2
Avg
5.76E\(-\)05
3.31E\(-\)05
3.85E\(-\)05
8.09E\(-\)05
7.25E\(-\)05
3.60E\(-\)05
1.67E\(-\)04
6.91E\(-\)05
7.08E\(-\)05
1.07E\(-\)03
Std
4.58E\(-\)05
3.02E\(-\)05
2.78E\(-\)05
4.92E\(-\)05
5.57E\(-\)05
3.24E\(-\)05
1.94E\(-\)04
5.52E\(-\)05
5.75E\(-\)05
1.52E\(-\)03
Rank
4
1
3
8
7
2
9
5
6
10
p value
2.92E\(-\)05
NaN
2.12E\(-\)04
7.66E\(-\)05
1.37E\(-\)06
5.69E\(-\)03
2.13E\(-\)07
3.03E\(-\)05
4.07E\(-\)05
8.48E\(-\)09
f3
Avg
\(-\)1.24E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.25E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.25E\(+\)04
\(-\)1.25E\(+\)04
\(-\)1.24E\(+\)04
\(-\)1.26E\(+\)04
Std
1.39E\(-\)01
4.08E\(-\)02
8.54E\(-\)02
5.34E\(-\)01
2.96E\(-\)02
1.63E\(-\)03
1.37E\(-\)01
2.16E\(+\)01
7.24E\(-\)01
7.41E\(+\)00
Rank
10
3
4
6
2
1
7
8
9
5
p value
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
8.12E\(-\)04
NaN
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
2.71E\(-\)02
f4
Avg
5.85E\(-\)248
0.00E\(+\)00
0.00E\(+\)00
2.17E\(-\)204
5.12E\(-\)240
3.83E\(-\)289
2.07E\(-\)131
5.42E\(-\)205
1.50E\(-\)226
2.33E\(-\)04
Std
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
1.14E\(-\)130
0.00E\(+\)00
0.00E\(+\)00
3.12E\(-\)04
Rank
4
1
1
8
5
3
9
7
6
10
p value
1.21E\(-\)12
NaN
NaN
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
f5
Avg
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.40E\(+\)00
Std
8.45E\(-\)04
5.35E\(-\)04
1.06E\(-\)03
2.68E\(-\)04
1.35E\(-\)03
1.12E\(-\)03
6.38E\(-\)05
8.79E\(-\)05
3.15E\(-\)04
3.83E\(-\)01
Rank
2
4
1
6
5
3
9
8
7
10
p value
1.33E\(-\)03
1.81E\(-\)04
NaN
5.19E\(-\)07
7.96E\(-\)05
5.94E\(-\)02
1.33E\(-\)10
3.82E\(-\)10
1.16E\(-\)07
3.02E\(-\)11
f6
Avg
\(-\)3.26E\(+\)00
\(-\)3.29E\(+\)00
\(-\)3.31E\(+\)00
\(-\)3.26E\(+\)00
\(-\)3.26E\(+\)00
\(-\)3.26E\(+\)00
\(-\)3.25E\(+\)00
\(-\)3.26E\(+\)00
\(-\)3.26E\(+\)00
\(-\)2.19E\(+\)00
Std
6.85E\(-\)02
5.93E\(-\)03
3.40E\(-\)03
6.96E\(-\)02
6.76E\(-\)02
6.31E\(-\)02
6.69E\(-\)02
6.23E\(-\)02
6.19E\(-\)02
4.15E\(-\)01
Rank
3
2
1
4
5
6
9
7
8
10
p value
7.73E\(-\)04
3.51E\(-\)04
NaN
9.33E\(-\)04
8.42E\(-\)06
4.38E\(-\)07
7.06E\(-\)08
1.76E\(-\)08
2.01E\(-\)08
3.02E\(-\)11
Average rank
4.6667
2.1667
1.8333
6.6667
4.6667
3.0000
8.6667
7.0000
7.0000
9.1667
Overall rank
4
2
1
6
4
3
9
7
7
10
According to the results in Table 5, the different combinations of the proposed MGEHO outperform the original EHO, and the obtained superiority is proved by Friedman’s test results and the associated p values. Considering the overall ranking, the strongest contender is MGEHO3, followed by MGEHO2 and MGEHO6. In other words, MGEHO with \(\left[ {{\omega _l},{\omega _u}} \right] = \left[ {0.1,0.7} \right]\) and \(k=3\) has the best performance among the nine combinations. Based on that, MGEHO with \(\left[ {{\omega _l},{\omega _u}} \right] = \left[ {0.1,0.7} \right]\) and \(k=3\) is adopted for the subsequent experiments.
4.1.3 Comparison with other algorithms
MGEHO is compared with 9 other metaheuristics. These include EHO [19], enhanced EHO based on the \(\gamma\) value (EEHO15) [22], OEHO [24], LFEHO [25], the IGWO [41], the EO [12], HHO [16], the WOA [15], and the SFO [14]. To make the experimental process fairer and more reliable, the parameters are set within each of the selected algorithms, as shown in Table 6. The other parameters are kept consistent except for those in Table 6.
Table 6
Parameter settings of different algorithms
Algorithm
Parameter
Range
EHO
\(\alpha\)
0.5
\(\beta\)
0.2
c
5
\({n_{{c_i}}}\)
6
EEHO15
\(\alpha\)
0.5
\(\beta\)
0.2
c
5
\({n_{{c_i}}}\)
6
\(\gamma\)
0.015
OEHO
\(\alpha\)
0.5
\(\beta\)
0.2
N
5
\({n_j}\)
6
LFEHO
\(\alpha\)
0.5
\({\beta _1}\)
0.1
\({\beta _2}\)
1.5
N
5
\({n_j}\)
6
IGWO
a
(2,0)
\(\overrightarrow{A}\)
[\(-a,a\)]
\(\overrightarrow{C}\)
[0,2]
EO
\({a_1}\)
2
\({a_2}\)
1
GP
0.5
HHO
\(\beta\)
1.5
\({E_0}\)
[\(-\)1,1]
WOA
a
(2,0)
\({a_2}\)
(2,1)
b
1
SFO
A
4
e
0.001
Initial population of sailfish
9
Initial population of sardines
21
MGEHO
\(\alpha\)
0.5
c
5
\({n_{{c_i}}}\)
6
\({\omega _l}\)
0.1
\({\omega _u}\)
0.7
k
3
4.1.4 Scalability test for MGEHO
Aiming to verify the adaptability of MGEHO under different conditions, the dimensions of F1–F23 are expanded from 30, 100 to 200, and the rest of the conditions are kept constant. After the experiments are completed, the mean value (Ave) and standard deviation (Std) are chosen as the evaluation indexes. The mean value reflects the accuracy and quality of a given algorithm, and the standard deviation reflects its stability. When solving a minimization problem, the smaller the mean value is, the better the performance of the algorithm. Similarity, the smaller the standard deviation is, the more stable the operation of the algorithm. The results are presented in Tables 7, 8, 9, and 10. In addition, the best solution in each table is shown in bold.
When the dimensionality of F1–F23 is 30, the means and standard deviations recorded after 30 independent operations are shown in Table 7, where boldface indicates the relative optimal value of the corresponding comparison algorithm. It is clear that MGEHO performs better than the other approaches in general. In the single-peak function test, MGEHO obtains global optimal solutions on F1, F2, F4, F8, and F10 and the best solution among the 10 algorithms on F5 and F9 with the best performance stability. HHO achieves better solutions on F6, and the EO obtains a better solution on F7. This indicates that the exploratory nature of MGEHO is enhanced by incorporating somersault foraging from MRFO. Additionally, by combining the convergence factor and Gaussian perturbation, the population diversity is improved, enabling the algorithm to further develop the exploration space to obtain the best solutions. Regarding the multipeak functions, MGEHO achieves the global optimal solutions on F12, F14, F19, F20, F22, and F23 and outperforms the other algorithms on F13, F17, and F18. This indicates that MGEHO has a strong ability to jump out of local optima; this stems from the clan updating operator and the convergence factor increasing the diversity and randomness of the population.
Table 7
Comparison of results obtained on the F1–F23 benchmark functions (\(d = 30\))
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
Mean
0.00E\(+\)00
2.98E\(-\)05
5.08E\(-\)06
8.07E\(-\)05
3.18E\(-\)03
1.93E\(-\)28
9.09E\(-\)41
4.30E\(-\)92
7.74E\(-\)11
3.62E\(-\)70
Std
0.00E\(+\)00
3.59E\(-\)05
8.57E\(-\)06
2.26E\(-\)04
4.30E\(-\)03
2.76E\(-\)28
1.94E\(-\)40
2.35E\(-\)91
1.13E\(-\)10
1.96E\(-\)69
F2
Mean
0.00E\(+\)00
1.58E\(-\)05
1.13E\(-\)06
7.76E\(-\)05
8.79E\(-\)04
3.57E\(-\)29
7.29E\(-\)42
5.28E\(-\)96
2.25E\(-\)09
1.55E\(-\)75
Std
0.00E\(+\)00
2.43E\(-\)05
4.58E\(-\)06
1.52E\(-\)04
1.24E\(-\)03
6.78E\(-\)29
1.81E\(-\)41
2.47E\(-\)95
6.38E\(-\)09
5.15E\(-\)75
F3
Mean
6.68E\(-\)294
3.64E\(-\)03
6.17E\(-\)04
1.08E\(-\)02
2.17E\(-\)02
6.56E\(-\)18
4.11E\(-\)24
3.83E\(-\)49
4.60E\(-\)05
1.44E\(-\)49
Std
0.00E\(+\)00
3.55E\(-\)03
6.15E\(-\)04
1.52E\(-\)02
2.13E\(-\)02
3.90E\(-\)18
4.65E\(-\)24
2.08E\(-\)48
4.15E\(-\)05
7.71E\(-\)49
F4
Mean
0.00E\(+\)00
6.95E\(-\)02
2.84E\(-\)03
7.47E\(-\)02
7.57E\(-\)01
1.43E\(-\)03
2.98E\(-\)09
1.53E\(-\)70
3.08E\(-\)08
4.31E\(+\)04
Std
0.00E\(+\)00
2.47E\(-\)01
4.53E\(-\)03
2.91E\(-\)01
2.82E\(+\)00
3.22E\(-\)03
6.43E\(-\)09
8.27E\(-\)70
6.37E\(-\)08
1.20E\(+\)04
F5
Mean
4.46E\(-\)297
2.19E\(-\)03
3.14E\(-\)04
1.20E\(-\)03
1.27E\(-\)02
1.95E\(-\)05
4.23E\(-\)10
7.45E\(-\)50
1.47E\(-\)06
5.23E\(+\)01
Std
0.00E\(+\)00
3.49E\(-\)03
4.08E\(-\)04
1.26E\(-\)03
1.29E\(-\)02
1.70E\(-\)05
8.89E\(-\)10
2.18E\(-\)49
1.09E\(-\)06
2.86E\(+\)01
F6
Mean
1.19E\(+\)00
1.45E\(-\)01
3.23E\(-\)02
1.23E\(-\)01
6.50E\(+\)00
2.43E\(+\)01
2.54E\(+\)01
9.80E\(-\)03
3.55E\(-\)02
2.81E\(+\)01
Std
4.95E\(+\)00
1.96E\(-\)01
3.11E\(-\)02
1.72E\(-\)01
9.93E\(+\)00
4.73E\(-\)01
2.13E\(-\)01
1.52E\(-\)02
3.45E\(-\)02
4.99E\(-\)01
F7
Mean
3.67E\(-\)03
1.58E\(-\)03
4.14E\(-\)04
1.10E\(-\)03
1.85E\(-\)01
2.44E\(-\)02
1.03E\(-\)05
1.59E\(-\)04
2.89E\(-\)02
3.55E\(-\)01
Std
7.03E\(-\)03
2.91E\(-\)03
5.54E\(-\)04
1.97E\(-\)03
2.04E\(-\)01
7.42E\(-\)02
9.38E\(-\)06
2.87E\(-\)04
9.06E\(-\)02
2.08E\(-\)01
F8
Mean
0.00E\(+\)00
1.31E\(-\)14
7.21E\(-\)17
1.59E\(-\)11
3.07E\(-\)11
3.59E\(-\)50
3.54E\(-\)73
1.17E\(-\)192
1.97E\(-\)20
7.38E\(-\)107
Std
0.00E\(+\)00
6.21E\(-\)14
3.57E\(-\)16
6.92E\(-\)11
7.48E\(-\)11
1.46E\(-\)49
1.15E\(-\)72
0.00E\(+\)00
5.42E\(-\)20
4.04E\(-\)106
F9
Mean
7.01E\(-\)05
1.34E\(-\)03
1.97E\(-\)03
1.44E\(-\)03
4.47E\(-\)03
2.73E\(-\)03
1.05E\(-\)03
1.77E\(-\)04
5.05E\(-\)04
4.06E\(-\)03
Std
5.65E\(-\)05
1.86E\(-\)03
1.56E\(-\)03
2.27E\(-\)03
6.81E\(-\)03
1.09E\(-\)03
5.60E\(-\)04
2.33E\(-\)04
4.36E\(-\)04
4.50E\(-\)03
F10
Mean
0.00E\(+\)00
1.06E\(-\)09
1.03E\(-\)11
9.92E\(-\)08
3.33E\(-\)09
1.54E\(-\)91
3.77E\(-\)131
3.51E\(-\)124
2.38E\(-\)13
9.70E\(-\)105
Std
0.00E\(+\)00
2.98E\(-\)09
2.54E\(-\)11
1.08E\(-\)07
6.06E\(-\)09
5.93E\(-\)91
2.06E\(-\)130
1.92E\(-\)123
5.23E\(-\)13
5.24E\(-\)104
F11
Mean
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)1.26E\(+\)04
\(-\)8.19E\(+\)03
\(-\)8.94E\(+\)03
\(-\)1.25E\(+\)04
\(-\)3.67E\(+\)03
\(-\)1.07E\(+\)04
Std
1.41E\(+\)01
3.41E\(+\)00
2.20E\(-\)01
8.40E\(+\)00
6.85E\(+\)00
1.56E\(+\)03
6.36E\(+\)02
2.79E\(+\)02
2.96E\(+\)02
1.88E\(+\)03
F12
Mean
0.00E\(+\)00
9.67E\(-\)05
3.29E\(-\)06
1.85E\(-\)03
4.04E\(-\)03
1.89E\(+\)01
0.00E\(+\)00
0.00E\(+\)00
3.54E\(-\)07
0.00E\(+\)00
Std
0.00E\(+\)00
2.53E\(-\)04
8.45E\(-\)06
3.52E\(-\)03
8.68E\(-\)03
1.05E\(+\)01
0.00E\(+\)00
0.00E\(+\)00
1.18E\(-\)06
0.00E\(+\)00
F13
Mean
8.88E\(-\)16
2.76E\(-\)03
4.71E\(-\)04
2.85E\(-\)03
1.11E\(-\)02
6.12E\(-\)14
8.23E\(-\)15
8.88E\(-\)16
4.79E\(-\)06
4.09E\(-\)15
Std
0.00E\(+\)00
3.02E\(-\)03
7.06E\(-\)04
3.48E\(-\)03
1.39E\(-\)02
1.13E\(-\)14
2.07E\(-\)15
0.00E\(+\)00
5.69E\(-\)06
2.70E\(-\)15
F14
Mean
0.00E\(+\)00
1.82E\(-\)04
4.93E\(-\)06
6.54E\(-\)05
1.19E\(-\)02
3.78E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
4.84E\(-\)12
1.39E\(-\)02
Std
0.00E\(+\)00
5.07E\(-\)04
1.15E\(-\)05
1.00E\(-\)04
3.00E\(-\)02
8.63E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
6.71E\(-\)12
5.29E\(-\)02
F15
Mean
9.02E\(-\)04
2.62E\(-\)05
2.07E\(-\)05
3.82E\(-\)05
3.37E\(-\)03
2.52E\(-\)04
5.45E\(-\)07
1.07E\(-\)05
2.84E\(-\)02
2.12E\(-\)02
Std
3.29E\(-\)03
4.00E\(-\)05
2.45E\(-\)05
5.59E\(-\)05
6.19E\(-\)03
1.17E\(-\)03
7.30E\(-\)07
1.24E\(-\)05
9.67E\(-\)02
1.69E\(-\)02
F16
Mean
2.12E\(-\)03
9.73E\(-\)04
3.10E\(-\)04
9.44E\(-\)04
3.03E\(-\)02
1.30E\(-\)01
2.53E\(-\)02
9.31E\(-\)05
7.92E\(-\)05
4.76E\(-\)01
Std
8.10E\(-\)03
1.83E\(-\)03
3.51E\(-\)04
1.91E\(-\)03
4.68E\(-\)02
1.69E\(-\)01
4.08E\(-\)02
1.67E\(-\)04
8.95E\(-\)05
2.67E\(-\)01
F17
Mean
1.11E\(-\)293
3.02E\(-\)04
6.37E\(-\)05
5.33E\(-\)04
2.49E\(-\)03
3.25E\(-\)04
1.38E\(-\)08
5.06E\(-\)07
4.10E\(-\)06
1.72E\(-\)45
Std
0.00E\(+\)00
3.85E\(-\)04
9.42E\(-\)05
5.77E\(-\)04
2.41E\(-\)03
2.45E\(-\)04
7.56E\(-\)08
2.77E\(-\)06
3.66E\(-\)06
9.40E\(-\)45
F18
Mean
\(-\)1.33E\(-\)15
1.09E\(-\)07
5.01E\(-\)09
8.01E\(-\)05
1.79E\(-\)05
\(-\)1.33E\(-\)15
−1.33E\(-\)15
−1.33E\(-\)15
2.29E\(-\)09
−1.32E\(-\)15
Std
0.00E\(+\)00
1.43E\(-\)07
6.91E\(-\)09
1.98E\(-\)04
2.62E\(-\)05
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
4.20E\(-\)09
8.11E\(-\)17
F19
Mean
0.00E\(+\)00
9.02E\(-\)01
2.16E\(-\)01
9.90E\(-\)01
1.48E\(+\)00
6.25E\(-\)07
2.01E\(-\)11
7.83E\(-\)26
9.51E\(-\)02
4.12E\(-\)28
Std
0.00E\(+\)00
4.91E\(-\)01
1.30E\(-\)01
6.23E\(-\)01
8.79E\(-\)01
5.43E\(-\)07
1.36E\(-\)11
3.50E\(-\)25
5.73E\(-\)02
2.06E\(-\)27
F20
Mean
0.00E\(+\)00
4.09E\(+\)00
4.50E\(-\)01
7.73E\(+\)00
5.16E\(+\)02
4.49E\(-\)25
5.20E\(-\)37
2.50E\(-\)91
8.13E\(-\)06
1.20E\(-\)68
Std
0.00E\(+\)00
6.95E\(+\)00
1.24E\(+\)00
2.64E\(+\)01
1.05E\(+\)03
1.16E\(-\)24
1.88E\(-\)36
1.36E\(-\)90
1.34E\(-\)05
6.40E\(-\)68
F21
Mean
\(-\)9.87E\(-\)01
\(-\)9.83E\(-\)01
\(-\)9.95E\(-\)01
\(-\)9.82E\(-\)01
−7.72E\(-\)01
0.00E\(+\)00
−3.33E\(-\)02
\(-\)9.96E\(-\)01
\(-\)9.42E\(-\)01
0.00E\(+\)00
Std
2.28E\(-\)02
3.43E\(-\)02
8.36E\(-\)03
2.45E\(-\)02
3.15E\(-\)01
0.00E\(+\)00
1.82E\(-\)01
5.04E\(-\)03
4.23E\(-\)02
0.00E\(+\)00
F22
Mean
0.00E\(+\)00
8.13E\(-\)03
4.65E\(-\)04
6.37E\(-\)03
4.56E\(-\)02
2.17E\(-\)01
1.07E\(-\)01
1.84E\(-\)47
1.40E\(-\)03
1.17E\(-\)01
Std
0.00E\(+\)00
7.31E\(-\)03
1.18E\(-\)03
1.85E\(-\)02
5.42E\(-\)02
3.79E\(-\)02
2.54E\(-\)02
1.00E\(-\)46
1.76E\(-\)03
7.46E\(-\)02
F23
Mean
0.00E\(+\)00
1.75E\(-\)03
9.50E\(-\)06
7.28E\(-\)04
2.97E\(-\)02
2.67E\(-\)01
4.37E\(-\)02
0.00E\(+\)00
2.37E\(-\)07
1.19E\(-\)01
Std
0.00E\(+\)00
2.67E\(-\)03
2.32E\(-\)05
8.97E\(-\)04
2.66E\(-\)02
8.89E\(-\)02
4.00E\(-\)12
0.00E\(+\)00
5.81E\(-\)07
1.21E\(-\)01
Table 8
Comparison of results obtained on the F24–F33 benchmark functions
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F24
Mean
9.98E\(-\)01
9.98E\(-\)01
9.98E\(-\)01
9.98E\(-\)01
9.98E\(-\)01
9.98E\(-\)01
1.06E\(+\)00
1.53E\(+\)00
6.70E\(+\)00
2.08E\(+\)00
Std
1.10E\(-\)17
8.88E\(-\)04
2.20E\(-\)09
1.27E\(-\)04
2.95E\(-\)05
1.13E\(-\)16
3.62E\(-\)01
1.29E\(+\)00
3.31E\(+\)00
2.59E\(+\)00
F25
Mean
4.08E\(-\)04
1.62E\(-\)03
1.30E\(-\)03
1.65E\(-\)03
8.02E\(-\)04
1.64E\(-\)03
2.45E\(-\)03
3.53E\(-\)04
3.87E\(-\)04
1.35E\(-\)03
Std
2.28E\(-\)04
1.82E\(-\)04
5.16E\(-\)04
1.35E\(-\)04
5.48E\(-\)04
5.09E\(-\)03
6.08E\(-\)03
4.31E\(-\)05
9.72E\(-\)05
3.40E\(-\)03
F26
Mean
\(-\)1.03E\(+\)00
\(-\)7.28E\(-\)01
\(-\)1.03E\(+\)00
\(-\)8.01E\(-\)01
\(-\)1.02E\(+\)00
\(-\)1.03E\(+\)00
\(-\)1.03E\(+\)00
\(-\)1.03E\(+\)00
\(-\)1.02E\(+\)00
\(-\)1.03E\(+\)00
Std
1.07E\(-\)18
3.69E\(-\)01
1.95E\(-\)04
3.78E\(-\)01
8.12E\(-\)03
6.18E\(-\)16
6.52E\(-\)16
1.35E\(-\)09
1.84E\(-\)02
9.89E\(-\)10
F27
Mean
4.00E\(-\)01
4.69E\(-\)01
3.98E\(-\)01
4.37E\(-\)01
4.68E\(-\)01
3.98E\(-\)01
3.98E\(-\)01
3.98E\(-\)01
3.99E\(-\)01
3.98E\(-\)01
Std
2.38E\(-\)03
1.00E\(-\)01
6.52E\(-\)04
4.44E\(-\)02
1.11E\(-\)01
1.76E\(-\)13
0.00E\(+\)00
1.47E\(-\)05
8.03E\(-\)04
2.49E\(-\)05
F28
Mean
3.00E\(+\)00
2.23E\(+\)01
7.57E\(+\)00
1.66E\(+\)01
9.82E\(+\)00
3.00E\(+\)00
3.00E\(+\)00
3.00E\(+\)00
7.21E\(+\)00
3.00E\(+\)00
Std
1.97E\(-\)04
1.05E\(+\)01
1.04E\(+\)01
1.04E\(+\)01
9.89E\(+\)00
1.45E\(-\)15
1.70E\(-\)15
1.63E\(-\)07
7.11E\(+\)00
7.09E\(-\)05
F29
Mean
\(-\)3.86E\(+\)00
\(-\)3.28E\(+\)00
\(-\)3.77E\(+\)00
\(-\)3.58E\(+\)00
\(-\)3.78E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.86E\(+\)00
\(-\)3.83E\(+\)00
\(-\)3.85E\(+\)00
Std
3.46E\(-\)06
4.32E\(-\)01
8.66E\(-\)02
2.00E\(-\)01
1.24E\(-\)01
2.52E\(-\)15
1.44E\(-\)03
2.16E\(-\)03
3.16E\(-\)02
2.04E\(-\)02
F30
Mean
\(-\)3.22E\(+\)00
\(-\)2.04E\(+\)00
\(-\)2.80E\(+\)00
\(-\)2.13E\(+\)00
\(-\)3.16E\(+\)00
\(-\)3.30E\(+\)00
\(-\)3.27E\(+\)00
\(-\)3.09E\(+\)00
\(-\)2.92E\(+\)00
\(-\)3.22E\(+\)00
Std
7.75E\(-\)02
4.88E\(-\)01
3.29E\(-\)01
5.03E\(-\)01
1.33E\(-\)01
4.84E\(-\)02
6.05E\(-\)02
1.23E\(-\)01
2.22E\(-\)01
1.28E\(-\)01
F31
Mean
\(-\)1.02E\(+\)01
\(-\)1.01E\(+\)01
\(-\)1.02E\(+\)01
\(-\)1.01E\(+\)01
\(-\)1.01E\(+\)01
\(-\)9.98E\(+\)00
\(-\)8.47E\(+\)00
\(-\)5.22E\(+\)00
\(-\)1.00E\(+\)01
\(-\)8.27E\(+\)00
Std
1.68E\(-\)05
1.82E\(-\)02
4.94E\(-\)04
3.27E\(-\)02
2.81E\(-\)02
9.22E\(-\)01
2.68E\(+\)00
9.17E\(-\)01
1.96E\(-\)01
2.54E\(+\)00
F32
Mean
\(-\)1.04E\(+\)01
\(-\)1.04E\(+\)01
\(-\)1.04E\(+\)01
\(-\)1.04E\(+\)01
\(-\)1.04E\(+\)01
\(-\)1.04E\(+\)01
\(-\)9.65E\(+\)00
\(-\)5.25E\(+\)00
\(-\)1.03E\(+\)01
\(-\)7.40E\(+\)00
Std
7.30E\(-\)06
1.25E\(-\)02
4.23E\(-\)04
3.87E\(-\)02
3.48E\(-\)02
2.16E\(-\)05
1.97E\(+\)00
9.24E\(-\)01
1.57E\(-\)01
3.09E\(+\)00
F33
Mean
\(-\)1.05E\(+\)01
\(-\)1.05E\(+\)01
\(-\)1.05E\(+\)01
\(-\)1.05E\(+\)01
\(-\)1.05E\(+\)01
\(-\)1.05E\(+\)01
\(-\)9.10E\(+\)00
\(-\)5.11E\(+\)00
\(-\)1.04E\(+\)01
\(-\)7.41E\(+\)00
Std
7.57E\(-\)11
2.16E\(-\)02
4.40E\(-\)04
5.27E\(-\)02
4.87E\(-\)02
4.93E\(-\)09
2.69E\(+\)00
1.09E\(+\)00
1.37E\(-\)01
3.24E\(+\)00
Table 9
Comparison of results obtained on the F1–F23 benchmark functions (\(d = 100\))
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
Mean
0.00E\(+\)00
2.35E\(-\)04
1.69E\(-\)05
2.90E\(-\)04
1.24E\(-\)02
2.19E\(-\)12
5.27E\(-\)29
7.66E\(-\)91
2.82E\(-\)10
5.66E\(-\)70
Std
0.00E\(+\)00
3.57E\(-\)04
3.59E\(-\)05
4.76E\(-\)04
1.48E\(-\)02
1.64E\(-\)12
8.97E\(-\)29
4.19E\(-\)90
5.11E\(-\)10
3.09E\(-\)69
F2
Mean
0.00E\(+\)00
2.56E\(-\)04
3.51E\(-\)06
8.64E\(-\)04
5.23E\(-\)03
8.79E\(-\)13
1.31E\(-\)29
6.63E\(-\)93
1.14E\(-\)08
2.40E\(-\)71
Std
0.00E\(+\)00
8.98E\(-\)04
8.34E\(-\)06
1.10E\(-\)03
1.45E\(-\)02
6.42E\(-\)13
1.21E\(-\)29
3.63E\(-\)92
1.59E\(-\)08
1.18E\(-\)70
F3
Mean
4.21E\(-\)296
1.30E\(-\)02
2.48E\(-\)03
3.87E\(-\)02
9.13E\(-\)02
1.78E\(-\)08
1.62E\(-\)17
6.00E\(-\)51
1.02E\(-\)04
5.75E\(-\)49
Std
0.00E\(+\)00
1.35E\(-\)02
3.18E\(-\)03
3.67E\(-\)02
8.45E\(-\)02
6.00E\(-\)09
9.77E\(-\)18
1.86E\(-\)50
8.64E\(-\)05
2.04E\(-\)48
F4
Mean
0.00E\(+\)00
1.19E\(+\)00
7.87E\(-\)02
1.89E\(+\)00
1.65E\(+\)01
5.58E\(+\)03
5.28E\(+\)00
3.07E\(-\)55
7.45E\(-\)07
1.11E\(+\)06
Std
0.00E\(+\)00
2.41E\(+\)00
2.23E\(-\)01
3.37E\(+\)00
3.54E\(+\)01
2.42E\(+\)03
1.59E\(+\)01
1.68E\(-\)54
1.02E\(-\)06
3.45E\(+\)05
F5
Mean
2.84E\(-\)297
1.31E\(-\)03
2.29E\(-\)04
9.94E\(-\)04
1.88E\(-\)02
4.09E\(+\)00
1.52E\(-\)03
5.06E\(-\)48
1.68E\(-\)06
7.21E\(+\)01
Std
0.00E\(+\)00
1.68E\(-\)03
3.39E\(-\)04
1.12E\(-\)03
1.89E\(-\)02
2.35E\(+\)00
2.25E\(-\)03
2.44E\(-\)47
1.57E\(-\)06
2.63E\(+\)01
F6
Mean
3.92E\(+\)00
1.23E\(+\)00
1.19E\(-\)01
5.65E\(-\)01
2.05E\(+\)01
9.76E\(+\)01
9.61E\(+\)01
5.29E\(-\)02
1.01E\(-\)01
9.81E\(+\)01
Std
7.34E\(+\)00
4.19E\(+\)00
1.42E\(-\)01
1.14E\(+\)00
3.63E\(+\)01
1.14E\(+\)00
7.88E\(-\)01
6.91E\(-\)02
1.26E\(-\)01
2.44E\(-\)01
F7
Mean
1.30E\(-\)01
5.20E\(-\)03
8.97E\(-\)04
4.72E\(-\)03
1.11E\(+\)00
7.94E\(+\)00
3.81E\(+\)00
6.53E\(-\)04
2.27E\(-\)01
4.16E\(+\)00
Std
6.86E\(-\)01
9.39E\(-\)03
1.25E\(-\)03
1.48E\(-\)02
1.98E\(+\)00
8.86E\(-\)01
5.74E\(-\)01
1.13E\(-\)03
7.90E\(-\)01
1.50E\(+\)00
F8
Mean
0.00E\(+\)00
6.36E\(-\)14
1.97E\(-\)17
7.02E\(-\)10
2.83E\(-\)11
2.32E\(-\)22
1.22E\(-\)50
3.38E\(-\)195
1.73E\(-\)18
7.77E\(-\)108
Std
0.00E\(+\)00
3.30E\(-\)13
6.67E\(-\)17
2.62E\(-\)09
9.63E\(-\)11
3.98E\(-\)22
2.45E\(-\)50
0.00E\(+\)00
7.94E\(-\)18
4.06E\(-\)107
F9
Mean
5.57E\(-\)05
2.15E\(-\)03
2.74E\(-\)03
2.04E\(-\)03
6.03E\(-\)03
1.21E\(-\)02
2.49E\(-\)03
1.53E\(-\)04
6.10E\(-\)04
3.95E\(-\)03
Std
4.85E\(-\)05
3.97E\(-\)03
2.12E\(-\)03
3.62E\(-\)03
7.87E\(-\)03
3.48E\(-\)03
1.09E\(-\)03
1.76E\(-\)04
5.23E\(-\)04
4.94E\(-\)03
F10
Mean
0.00E\(+\)00
9.31E\(-\)10
6.97E\(-\)12
1.96E\(-\)07
4.72E\(-\)09
1.02E\(-\)60
1.58E\(-\)128
1.78E\(-\)122
1.83E\(-\)13
2.81E\(-\)105
Std
0.00E\(+\)00
1.94E\(-\)09
1.46E\(-\)11
3.95E\(-\)07
1.13E\(-\)08
4.63E\(-\)60
8.32E\(-\)128
9.73E\(-\)122
3.61E\(-\)13
1.07E\(-\)104
F11
Mean
\(-\)4.19E\(+\)04
\(-\)4.19E\(+\)04
\(-\)4.19E\(+\)04
\(-\)4.19E\(+\)04
\(-\)4.19E\(+\)04
\(-\)1.56E\(+\)04
\(-\)2.60E\(+\)04
\(-\)4.19E\(+\)04
\(-\)6.80E\(+\)03
\(-\)3.33E\(+\)04
Std
8.39E\(+\)01
1.26E\(+\)01
4.46E\(-\)01
1.48E\(+\)01
1.58E\(+\)01
6.91E\(+\)03
1.36E\(+\)03
5.45E\(+\)00
6.38E\(+\)02
5.16E\(+\)03
F12
Mean
0.00E\(+\)00
4.52E\(-\)04
3.12E\(-\)05
5.36E\(-\)03
2.15E\(-\)02
1.35E\(+\)02
0.00E\(+\)00
0.00E\(+\)00
1.29E\(-\)06
0.00E\(+\)00
Std
0.00E\(+\)00
1.54E\(-\)03
1.03E\(-\)04
7.83E\(-\)03
5.20E\(-\)02
5.47E\(+\)01
0.00E\(+\)00
0.00E\(+\)00
3.21E\(-\)06
0.00E\(+\)00
F13
Mean
8.88E\(-\)16
2.00E\(-\)03
3.06E\(-\)04
3.00E\(-\)03
8.88E\(-\)03
1.88E\(-\)07
3.51E\(-\)14
8.88E\(-\)16
5.09E\(-\)06
4.09E\(-\)15
Std
0.00E\(+\)00
2.11E\(-\)03
5.29E\(-\)04
3.17E\(-\)03
8.78E\(-\)03
7.00E\(-\)08
6.56E\(-\)15
0.00E\(+\)00
4.17E\(-\)06
2.35E\(-\)15
F14
Mean
0.00E\(+\)00
2.12E\(-\)04
5.20E\(-\)05
2.08E\(-\)04
2.55E\(-\)02
1.28E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
5.17E\(-\)12
9.21E\(-\)03
Std
0.00E\(+\)00
3.23E\(-\)04
1.85E\(-\)04
5.61E\(-\)04
4.32E\(-\)02
4.03E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
8.32E\(-\)12
5.04E\(-\)02
F15
Mean
1.57E\(-\)05
3.12E\(-\)05
1.18E\(-\)05
1.04E\(-\)05
2.03E\(-\)03
2.04E\(-\)01
3.78E\(-\)02
4.47E\(-\)06
3.10E\(-\)03
3.77E\(-\)02
Std
1.96E\(-\)05
6.72E\(-\)05
1.13E\(-\)05
1.21E\(-\)05
4.45E\(-\)03
6.42E\(-\)02
1.13E\(-\)02
5.97E\(-\)06
7.98E\(-\)03
1.84E\(-\)02
F16
Mean
2.72E\(-\)03
5.90E\(-\)04
3.48E\(-\)04
6.05E\(-\)04
5.96E\(-\)02
5.99E\(+\)00
5.86E\(+\)00
1.76E\(-\)04
1.71E\(-\)04
3.06E\(+\)00
Std
4.44E\(-\)03
9.10E\(-\)04
3.53E\(-\)04
9.08E\(-\)04
9.26E\(-\)02
6.21E\(-\)01
1.15E\(+\)00
2.47E\(-\)04
1.76E\(-\)04
9.92E\(-\)01
F17
Mean
3.97E\(-\)287
8.17E\(-\)04
3.65E\(-\)04
1.58E\(-\)03
8.09E\(-\)03
4.92E\(-\)03
4.64E\(-\)18
6.80E\(-\)06
1.38E\(-\)05
1.05E\(-\)50
Std
0.00E\(+\)00
8.78E\(-\)04
5.41E\(-\)04
2.07E\(-\)03
8.11E\(-\)03
3.49E\(-\)03
4.64E\(-\)18
3.73E\(-\)05
1.52E\(-\)05
2.77E\(-\)50
F18
Mean
1.95E\(-\)14
1.51E\(-\)06
2.39E\(-\)08
3.29E\(-\)04
7.91E\(-\)05
4.07E\(-\)14
1.95E\(-\)14
1.95E\(-\)14
9.74E\(-\)09
1.95E\(-\)14
Std
0.00E\(+\)00
5.00E\(-\)06
4.11E\(-\)08
1.30E\(-\)03
2.55E\(-\)04
6.10E\(-\)15
0.00E\(+\)00
0.00E\(+\)00
2.67E\(-\)08
0.00E\(+\)00
F19
Mean
0.00E\(+\)00
2.62E\(+\)00
6.26E\(-\)01
2.95E\(+\)00
4.60E\(+\)00
3.18E\(-\)01
4.41E\(-\)08
3.11E\(-\)25
2.94E\(-\)01
1.87E\(-\)28
Std
0.00E\(+\)00
1.19E\(+\)00
2.77E\(-\)01
1.52E\(+\)00
2.58E\(+\)00
1.28E\(-\)01
1.50E\(-\)08
1.23E\(-\)24
1.86E\(-\)01
7.87E\(-\)28
F20
Mean
0.00E\(+\)00
1.28E\(+\)01
3.21E\(+\)00
2.66E\(+\)01
1.30E\(+\)03
3.77E\(-\)09
1.34E\(-\)25
9.03E\(-\)84
5.33E\(-\)05
8.69E\(-\)68
Std
0.00E\(+\)00
2.23E\(+\)01
8.32E\(+\)00
4.29E\(+\)01
2.13E\(+\)03
4.31E\(-\)09
2.05E\(-\)25
4.95E\(-\)83
1.52E\(-\)04
2.69E\(-\)67
F21
Mean
\(-\)9.61E\(-\)01
\(-\)9.17E\(-\)01
\(-\)1.00E\(+\)00
\(-\)9.51E\(-\)01
\(-\)5.54E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
\(-\)9.67E\(-\)01
\(-\)8.77E\(-\)01
0.00E\(+\)00
Std
9.22E\(-\)02
1.76E\(-\)01
1.75E\(-\)06
1.03E\(-\)01
3.64E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
5.46E\(-\)02
9.57E\(-\)02
0.00E\(+\)00
F22
Mean
0.00E\(+\)00
1.86E\(-\)02
4.70E\(-\)03
1.58E\(-\)02
9.47E\(-\)02
4.93E\(-\)01
1.80E\(-\)01
3.88E\(-\)48
1.28E\(-\)03
1.10E\(-\)01
Std
0.00E\(+\)00
3.04E\(-\)02
1.85E\(-\)02
2.62E\(-\)02
8.69E\(-\)02
5.21E\(-\)02
4.07E\(-\)02
1.75E\(-\)47
1.75E\(-\)03
7.12E\(-\)02
F23
Mean
0.00E\(+\)00
9.02E\(-\)03
4.04E\(-\)05
5.49E\(-\)03
5.01E\(-\)02
4.70E\(-\)01
1.04E\(-\)01
0.00E\(+\)00
1.58E\(-\)05
1.08E\(-\)01
Std
0.00E\(+\)00
1.56E\(-\)02
1.25E\(-\)04
9.54E\(-\)03
5.35E\(-\)02
1.70E\(-\)02
1.11E\(-\)01
0.00E\(+\)00
3.26E\(-\)05
1.19E\(-\)01
Table 10
Comparison of results obtained on the F1–F23 benchmark functions (\(d = 200\))
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
Mean
0.00E\(+\)00
8.88E\(-\)04
2.67E\(-\)04
1.91E\(-\)03
1.09E\(-\)01
1.29E\(-\)02
1.64E\(-\)22
1.92E\(-\)95
8.32E\(-\)10
8.88E\(-\)70
Std
0.00E\(+\)00
1.43E\(-\)03
7.05E\(-\)04
2.56E\(-\)03
2.35E\(-\)01
5.16E\(-\)03
3.32E\(-\)22
8.02E\(-\)95
1.33E\(-\)09
3.42E\(-\)69
F2
Mean
0.00E\(+\)00
4.67E\(-\)04
6.73E\(-\)05
8.18E\(-\)03
5.01E\(-\)02
2.48E\(-\)07
8.41E\(-\)26
3.78E\(-\)92
2.84E\(-\)08
5.14E\(-\)71
Std
0.00E\(+\)00
9.95E\(-\)04
2.28E\(-\)04
3.28E\(-\)02
6.65E\(-\)02
1.61E\(-\)07
1.60E\(-\)25
2.07E\(-\)91
3.61E\(-\)08
2.51E\(-\)70
F3
Mean
8.14E\(-\)296
3.29E\(-\)02
5.96E\(-\)03
5.95E\(-\)02
1.98E\(-\)01
1.67E\(-\)05
1.71E\(-\)15
3.67E\(-\)51
3.12E\(-\)04
1.02E\(-\)47
Std
0.00E\(+\)00
4.06E\(-\)02
9.22E\(-\)03
5.82E\(-\)02
1.92E\(-\)01
4.71E\(-\)06
1.28E\(-\)15
9.51E\(-\)51
3.17E\(-\)04
5.54E\(-\)47
F4
Mean
0.00E\(+\)00
5.88E\(+\)00
1.12E\(+\)00
1.93E\(+\)01
9.24E\(+\)01
7.60E\(+\)04
9.27E\(+\)02
1.76E\(-\)45
1.05E\(-\)05
4.57E\(+\)06
Std
0.00E\(+\)00
1.22E\(+\)01
3.45E\(+\)00
6.54E\(+\)01
1.48E\(+\)02
1.56E\(+\)04
1.22E\(+\)03
9.24E\(-\)45
2.61E\(-\)05
1.18E\(+\)06
F5
Mean
4.87E\(-\)297
1.60E\(-\)03
2.35E\(-\)04
1.01E\(-\)03
1.31E\(-\)02
3.12E\(+\)01
2.94E\(+\)01
3.96E\(-\)48
1.18E\(-\)06
8.16E\(+\)01
Std
0.00E\(+\)00
1.48E\(-\)03
2.21E\(-\)04
1.12E\(-\)03
1.43E\(-\)02
5.43E\(+\)00
2.74E\(+\)01
1.37E\(-\)47
1.07E\(-\)06
1.98E\(+\)01
F6
Mean
5.16E\(+\)00
8.81E\(-\)01
3.22E\(-\)01
9.04E\(-\)01
5.08E\(+\)01
1.98E\(+\)02
1.97E\(+\)02
8.22E\(-\)02
1.47E\(-\)01
1.98E\(+\)02
Std
1.06E\(+\)01
2.62E\(+\)00
3.02E\(-\)01
1.58E\(+\)00
7.17E\(+\)01
5.11E\(-\)01
8.04E\(-\)01
1.09E\(-\)01
1.32E\(-\)01
1.73E\(-\)01
F7
Mean
1.21E\(-\)01
9.45E\(-\)03
1.94E\(-\)03
7.17E\(-\)03
2.91E\(+\)00
2.64E\(+\)01
2.04E\(+\)01
8.92E\(-\)04
5.34E\(-\)02
1.09E\(+\)01
Std
3.02E\(-\)01
1.53E\(-\)02
2.96E\(-\)03
1.66E\(-\)02
4.27E\(+\)00
1.46E\(+\)00
1.25E\(+\)00
1.27E\(-\)03
1.79E\(-\)01
2.87E\(+\)00
F8
Mean
0.00E\(+\)00
6.79E\(-\)14
4.67E\(-\)16
1.65E\(-\)09
1.02E\(-\)09
9.44E\(-\)13
7.06E\(-\)43
3.18E\(-\)189
6.07E\(-\)18
1.54E\(-\)105
Std
0.00E\(+\)00
2.92E\(-\)13
1.63E\(-\)15
3.68E\(-\)09
3.94E\(-\)09
2.00E\(-\)12
3.51E\(-\)42
0.00E\(+\)00
3.31E\(-\)17
8.44E\(-\)105
F9
Mean
7.52E\(-\)05
3.42E\(-\)03
2.80E\(-\)03
2.84E\(-\)03
7.21E\(-\)03
3.34E\(-\)02
3.24E\(-\)03
1.39E\(-\)04
6.24E\(-\)04
3.77E\(-\)03
Std
8.54E\(-\)05
5.10E\(-\)03
2.74E\(-\)03
6.27E\(-\)03
9.01E\(-\)03
8.59E\(-\)03
1.07E\(-\)03
8.56E\(-\)05
5.88E\(-\)04
3.10E\(-\)03
F10
Mean
0.00E\(+\)00
6.41E\(-\)10
3.72E\(-\)11
9.22E\(-\)08
2.29E\(-\)09
1.28E\(-\)35
1.15E\(-\)104
1.09E\(-\)122
5.10E\(-\)13
1.53E\(-\)107
Std
0.00E\(+\)00
1.35E\(-\)09
1.08E\(-\)10
1.44E\(-\)07
4.63E\(-\)09
7.03E\(-\)35
6.32E\(-\)104
4.56E\(-\)122
1.47E\(-\)12
6.19E\(-\)107
F11
Mean
\(-\)8.37E\(+\)04
\(-\)8.38E\(+\)04
\(-\)8.38E\(+\)04
\(-\)8.38E\(+\)04
\(-\)8.38E\(+\)04
\(-\)2.82E\(+\)04
\(-\)4.29E\(+\)04
\(-\)8.37E\(+\)04
\(-\)9.91E\(+\)03
\(-\)6.90E\(+\)04
Std
1.22E\(+\)02
8.68E\(+\)01
3.31E\(+\)00
3.61E\(+\)01
5.21E\(+\)01
1.40E\(+\)04
2.69E\(+\)03
3.43E\(+\)02
7.11E\(+\)02
1.34E\(+\)04
F12
Mean
0.00E\(+\)00
4.55E\(-\)04
9.20E\(-\)06
5.98E\(-\)03
1.01E\(-\)01
3.05E\(+\)02
0.00E\(+\)00
0.00E\(+\)00
3.41E\(-\)06
1.52E\(-\)14
Std
0.00E\(+\)00
8.89E\(-\)04
2.10E\(-\)05
9.91E\(-\)03
2.91E\(-\)01
8.73E\(+\)01
0.00E\(+\)00
0.00E\(+\)00
8.23E\(-\)06
8.30E\(-\)14
F13
Mean
8.88E\(-\)16
2.32E\(-\)03
2.28E\(-\)04
1.46E\(-\)03
1.51E\(-\)02
4.29E\(-\)05
1.05E\(-\)13
8.88E\(-\)16
3.21E\(-\)06
3.49E\(-\)15
Std
0.00E\(+\)00
2.59E\(-\)03
3.77E\(-\)04
1.20E\(-\)03
1.46E\(-\)02
1.19E\(-\)05
2.93E\(-\)14
0.00E\(+\)00
3.57E\(-\)06
2.63E\(-\)15
F14
Mean
0.00E\(+\)00
4.54E\(-\)04
3.64E\(-\)05
9.25E\(-\)05
1.84E\(-\)02
4.39E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
1.75E\(-\)11
0.00E\(+\)00
Std
0.00E\(+\)00
1.03E\(-\)03
1.26E\(-\)04
1.30E\(-\)04
2.41E\(-\)02
9.35E\(-\)03
0.00E\(+\)00
0.00E\(+\)00
4.33E\(-\)11
0.00E\(+\)00
F15
Mean
5.43E\(-\)05
9.31E\(-\)06
8.41E\(-\)06
1.21E\(-\)05
1.77E\(-\)03
6.07E\(+\)00
2.14E\(-\)01
4.06E\(-\)06
1.39E\(-\)02
7.16E\(-\)02
Std
1.31E\(-\)04
1.46E\(-\)05
8.37E\(-\)06
2.42E\(-\)05
3.57E\(-\)03
5.69E\(+\)00
3.53E\(-\)02
6.33E\(-\)06
4.54E\(-\)02
2.98E\(-\)02
F16
Mean
1.98E\(-\)03
2.71E\(-\)03
7.21E\(-\)04
1.20E\(-\)03
2.60E\(-\)01
2.13E\(+\)01
1.84E\(+\)01
2.13E\(-\)04
2.73E\(-\)04
7.04E\(+\)00
Std
3.40E\(-\)03
7.21E\(-\)03
9.95E\(-\)04
4.59E\(-\)03
5.63E\(-\)01
2.90E\(+\)00
4.33E\(-\)01
2.11E\(-\)04
2.28E\(-\)04
2.20E\(+\)00
F17
Mean
1.08E\(-\)278
2.03E\(-\)03
3.80E\(-\)04
3.33E\(-\)03
1.37E\(-\)02
1.87E\(-\)02
2.33E\(-\)16
6.12E\(-\)05
3.15E\(-\)05
3.87E\(-\)50
Std
0.00E\(+\)00
2.40E\(-\)03
6.76E\(-\)04
3.97E\(-\)03
1.45E\(-\)02
8.60E\(-\)03
1.58E\(-\)16
2.22E\(-\)04
5.14E\(-\)05
1.48E\(-\)49
F18
Mean
\(-\)1.42E\(-\)14
2.04E\(-\)06
7.21E\(-\)08
6.50E\(-\)04
2.30E\(-\)04
2.33E\(-\)10
\(-\)1.42E\(-\)14
\(-\)1.42E\(-\)14
1.53E\(-\)08
\(-\)1.42E\(-\)14
Std
0.00E\(+\)00
6.15E\(-\)06
1.17E\(-\)07
1.40E\(-\)03
4.01E\(-\)04
1.40E\(-\)10
0.00E\(+\)00
0.00E\(+\)00
1.73E\(-\)08
0.00E\(+\)00
F19
Mean
0.00E\(+\)00
4.90E\(+\)00
2.03E\(+\)00
6.45E\(+\)00
1.18E\(+\)01
3.42E\(+\)00
4.61E\(-\)07
1.46E\(-\)24
5.46E\(-\)01
4.25E\(-\)29
Std
0.00E\(+\)00
3.36E\(+\)00
1.09E\(+\)00
4.48E\(+\)00
6.92E\(+\)00
7.36E\(-\)01
1.58E\(-\)07
6.10E\(-\)24
3.78E\(-\)01
1.43E\(-\)28
F20
Mean
0.00E\(+\)00
1.08E\(+\)01
7.51E\(-\)01
3.24E\(+\)01
3.46E\(+\)03
3.79E\(-\)04
5.28E\(-\)22
1.99E\(-\)92
2.01E\(-\)05
1.18E\(-\)67
Std
0.00E\(+\)00
1.87E\(+\)01
1.67E\(+\)00
1.06E\(+\)02
5.82E\(+\)03
2.11E\(-\)04
6.99E\(-\)22
9.48E\(-\)92
2.45E\(-\)05
4.23E\(-\)67
F21
Mean
\(-\)7.72E\(-\)01
\(-\)8.13E\(-\)01
\(-\)1.00E\(+\)00
\(-\)8.75E\(-\)01
\(-\)4.94E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
\(-\)5.01E\(-\)01
\(-\)8.48E\(-\)01
0.00E\(+\)00
Std
3.10E\(-\)01
2.79E\(-\)01
1.04E\(-\)05
2.28E\(-\)01
3.61E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
4.70E\(-\)01
1.81E\(-\)01
0.00E\(+\)00
F22
Mean
0.00E\(+\)00
2.70E\(-\)02
2.28E\(-\)03
2.04E\(-\)02
9.41E\(-\)02
7.77E\(-\)01
2.00E\(-\)01
3.49E\(-\)49
1.35E\(-\)03
1.17E\(-\)01
Std
0.00E\(+\)00
3.45E\(-\)02
5.49E\(-\)03
3.03E\(-\)02
9.52E\(-\)02
9.35E\(-\)02
3.29E\(-\)10
1.33E\(-\)48
1.72E\(-\)03
6.47E\(-\)02
F23
Mean
0.00E\(+\)00
1.08E\(-\)02
1.06E\(-\)04
7.32E\(-\)03
7.92E\(-\)02
4.97E\(-\)01
2.41E\(-\)01
0.00E\(+\)00
6.43E\(-\)05
8.93E\(-\)02
Std
0.00E\(+\)00
1.69E\(-\)02
2.22E\(-\)04
1.38E\(-\)02
7.37E\(-\)02
1.10E\(-\)03
1.11E\(-\)01
0.00E\(+\)00
1.15E\(-\)04
1.09E\(-\)01
The experimental results on the fixed-dimension multimodal benchmark functions are shown in Table 8. Through the data shown in the table, MGEHO achieves superior solutions on F24, F26, F31, F32, F33 and the best average result on F30. The IGWO performs best on F27 and F28 and runs most stably on F29 and F30. HHO achieves the best result on F25, and the solution obtained on F29 is the closest to the global solution. This indicates that by introducing the nonlinear convergence factor, an algorithm can effectively control the balance between exploration and exploitation by maintaining diversification and intensification throughout the search process.
Table 9 shows the test results of different algorithms on the F1–F23 benchmark functions when the dimensionality is 100. As the dimensionality increases, algorithms tend to need more information to search for optimal solutions; thus, increasing the dimensionality tests the ability of each algorithm to adapt to problems with different variants. The comparison results illustrate that the proposed MGEHO algorithm outperforms the other nine algorithms. Among the 23 benchmark functions, MGEHO surpasses the other algorithms on 17 of them. For F11, EEHO15 can perform better. LFEHO obtains the best solutions on F10 and F23, the IGWO has the best performance on F21, and the EO achieves the global optimal solutions on F12, F14, and F21. HHO has better results than those of the other algorithms on F6, F7, F12, F14, and F15. The SFO is able to provide the best solution on F16, and the WOA has the best results on F12 and F21. In summary, MGEHO has certain reproducibility and stability advantages with respect to solving complex high-dimensional problems.
Table 10 shows the mean and standard deviation results obtained by different algorithms when the dimensionality of F1–F23 is increased to 200. The performances of the algorithms fluctuate differently as the dimensionality increases. Regarding the high-dimensional single-peaked problem, MGEHO finds the best solutions for eight of the ten benchmark functions. In the high-dimensional multipeaked problem, MGEHO outperforms the other algorithms on 9 of the 13 benchmark functions. Therefore, the robustness and accuracy of the original algorithm are improved by mixing MRFO with EHO. It can be concluded that MGEHO is still the best-performing one among the nine other algorithms.
4.1.5 Convergence curve analysis
Convergence curves can reflect the abilities and convergence speeds of algorithms when dealing with optimization problems. The curves obtained with 30 dimensions are shown in Fig. 3. From these plots, it can be observed that MGEHO achieves the best solutions, which are more accurate than those of any other algorithm on both the single-peak and multipeak functions, thereby exhibiting superior performance. The results on the fixed-size functions indicate that MGEHO also ranks among the top algorithms in terms of its ability to balance exploration and exploitation.
García et al. [42] pointed out that when evaluating the performances of metaheuristic algorithms, it is not sufficient to make comparisons based only on their mean and standard deviation values. At the same time, there are inevitable chance factors that affect the experimental results during the experimental process. To examine the differences between different algorithms, the nonparametric Wilcoxon signed-rank test [43] and Friedman test [44] at the \(5\%\) significance level are invoked in the experiments.
Table 11
Nonparametric statistical comparison based on the Wilcoxon signed-rank test for F1–F33
Function
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F2
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F3
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
F4
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F5
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
F6
7.84E\(-\)04
1.70E\(-\)02
5.30E\(-\)06
1.04E\(-\)04
5.57E\(-\)10
5.57E\(-\)10
5.86E\(-\)06
4.36E\(-\)02
3.02E\(-\)11
F7
2.06E\(-\)01
9.88E\(-\)03
7.24E\(-\)06
1.07E\(-\)07
9.52E\(-\)04
1.07E\(-\)09
8.29E\(-\)06
6.97E\(-\)03
3.02E\(-\)11
F8
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F8
1.73E\(-\)07
8.15E\(-\)11
4.31E\(-\)08
4.18E\(-\)09
3.02E\(-\)11
3.02E\(-\)11
9.07E\(-\)03
2.03E\(-\)07
1.33E\(-\)10
F9
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F10
3.79E\(-\)03
3.50E\(-\)09
7.48E\(-\)06
5.30E\(-\)01
3.02E\(-\)11
3.02E\(-\)11
8.68E\(-\)03
3.02E\(-\)11
9.26E\(-\)09
F11
9.21E\(-\)05
1.26E\(-\)06
1.16E\(-\)07
2.68E\(-\)06
3.02E\(-\)11
3.02E\(-\)11
1.49E\(-\)04
3.02E\(-\)11
3.02E\(-\)11
F12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
NaN
NaN
1.21E\(-\)12
NaN
F13
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.03E\(-\)12
8.71E\(-\)14
NaN
1.21E\(-\)12
1.02E\(-\)07
F14
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
5.58E\(-\)03
NaN
NaN
1.21E\(-\)12
1.61E\(-\)01
F15
2.71E\(-\)02
5.01E\(-\)02
7.73E\(-\)02
2.13E\(-\)05
7.01E\(-\)02
5.97E\(-\)09
1.27E\(-\)02
9.71E\(-\)01
1.33E\(-\)10
F16
3.79E\(-\)04
3.63E\(-\)01
6.20E\(-\)01
2.00E\(-\)06
6.77E\(-\)05
9.82E\(-\)01
1.32E\(-\)04
3.01E\(-\)04
3.02E\(-\)11
F17
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
F18
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
NaN
NaN
NaN
1.21E\(-\)12
3.34E\(-\)01
F19
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F20
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F21
4.64E\(-\)05
3.37E\(-\)04
6.41E\(-\)05
2.13E\(-\)05
1.21E\(-\)12
3.02E\(-\)11
2.51E\(-\)02
1.53E\(-\)05
1.21E\(-\)12
F22
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.18E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
F23
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
1.21E\(-\)12
9.42E\(-\)13
NaN
1.21E\(-\)12
3.87E\(-\)12
F24
4.46E\(-\)01
1.17E\(-\)09
2.71E\(-\)01
1.33E\(-\)02
8.87E\(-\)12
3.11E\(-\)10
5.87E\(-\)04
3.02E\(-\)11
2.32E\(-\)02
F25
4.08E\(-\)11
1.55E\(-\)09
3.69E\(-\)11
8.20E\(-\)07
8.48E\(-\)09
4.92E\(-\)01
4.20E\(-\)01
4.06E\(-\)02
4.12E\(-\)06
F26
3.02E\(-\)11
4.46E\(-\)04
3.02E\(-\)11
7.39E\(-\)11
1.01E\(-\)11
4.08E\(-\)12
3.02E\(-\)11
3.69E\(-\)11
3.02E\(-\)11
F27
5.00E\(-\)09
1.78E\(-\)04
1.10E\(-\)08
4.11E\(-\)07
1.72E\(-\)12
1.21E\(-\)12
4.98E\(-\)11
3.04E\(-\)01
7.39E\(-\)11
F28
3.02E\(-\)11
1.47E\(-\)07
3.02E\(-\)11
7.39E\(-\)11
2.41E\(-\)11
2.53E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
3.01E\(-\)04
F29
3.02E\(-\)11
1.17E\(-\)05
3.02E\(-\)11
4.94E\(-\)05
1.41E\(-\)11
1.44E\(-\)10
4.46E\(-\)04
2.49E\(-\)06
3.95E\(-\)06
F30
3.02E\(-\)11
3.50E\(-\)09
3.02E\(-\)11
1.91E\(-\)01
7.04E\(-\)07
6.13E\(-\)03
1.11E\(-\)04
3.50E\(-\)09
1.09E\(-\)01
F31
4.06E\(-\)02
1.87E\(-\)05
8.77E\(-\)01
1.37E\(-\)01
1.07E\(-\)09
7.70E\(-\)03
3.02E\(-\)11
5.09E\(-\)08
4.44E\(-\)07
F32
5.83E\(-\)03
1.58E\(-\)04
1.19E\(-\)06
2.81E\(-\)02
3.02E\(-\)11
8.69E\(-\)07
3.02E\(-\)11
1.07E\(-\)09
4.31E\(-\)08
F33
2.64E\(-\)05
1.61E\(-\)06
3.85E\(-\)03
6.67E\(-\)03
3.02E\(-\)11
3.81E\(-\)04
3.02E\(-\)11
8.10E\(-\)10
3.20E\(-\)09
\(+\)/=/\(-\)
31/2/0
31/2/0
29/4/0
30/3/0
31/2/0
28/5/0
27/6/0
31/2/0
29/4/0
Table 12
Nonparametric statistical comparison based on the Friedman rank test for F1–F23
Function
Dimensionality
F1
F2
F3
F4
F5
F6
F7
F8
F9
F10
F11
F12
F13
MGEHO
30
1
1
1
1
1
4.5
6.4
1
1.63
1
4.57
1
1
100
1
1
1
1
1
4.63
4.63
1
1.43
1
4.3
1
1
200
1
1
1
1
1
4.6
5.07
1
1.7
1
4.4
1
1
EHO
30
8.43
8.47
8.2
7.13
7.4
4.43
5.43
7.7
5.33
8.37
3.97
7
8.6
100
8.57
8.37
8.1
5.87
5.77
4.27
4.07
7.87
5.33
8.2
3.53
6.93
8.67
200
7.27
8.23
8.17
5.37
5.73
3.97
4
6.83
5.37
8.1
3.97
6.93
8.5
EEHO15
30
7.5
7.27
7.17
6.03
6.4
3.47
4.83
7
7.1
7.13
1.5
6
7.23
100
7.33
7.07
7.27
4.6
4.57
3.27
3.53
6.83
6.73
6.93
1.7
5.93
7.3
200
6.47
7.37
7.27
4.4
4.27
3.93
2.87
6
6.03
7.3
1.63
5.73
6.83
OEHO
30
8.23
8.7
8.83
7.23
7.33
4.1
5.37
9.5
5.47
9.87
3.77
8.03
8.47
100
8.33
8.97
8.9
6.1
5.9
3.93
3.77
9.63
5.1
9.83
4.23
8.17
8.5
200
7.5
8.7
8.67
5.63
5.17
4.13
3.6
9.17
4.57
9.77
3.67
8.07
8.5
LFEHO
30
9.83
9.53
9.7
8.53
8.83
6.5
8.77
9.33
6.23
8.43
4.17
8.6
9.67
100
9.77
9.5
9.63
7.53
7.87
6.73
6.57
9.07
6.53
8.73
4.27
8.5
9.53
200
9.47
9.63
9.77
6.77
6.83
5.87
6.73
9.3
7.07
8.67
4.1
8.73
9.83
IGWO
30
5
5
5
6.03
5.03
7.87
4.6
5
8.47
5
8.47
10
5
100
5.1
5
5
9
9
9.17
10
5.37
9.73
5
8.87
10
5.03
200
9.3
6.07
5.1
9
8.57
9.4
10
8.3
9.97
5
8.77
10
6.33
EO
30
4
4
4
3.17
3
8.8
1.9
4
6.17
2.17
8.23
2
3.77
100
4
4
4
5.87
5.9
7.97
8.27
4
7.07
2.33
8.1
2
4
200
4
4
4
7.77
8.5
8.47
9
4
6.9
3.13
8.23
2
4
HHO
30
2
2
2.47
2
2
2.07
3.97
2
2.57
2.83
3.27
3
2
100
2
2
2.27
2
2
2.43
2.8
2
2.37
2.8
3.07
3
2
200
2
2
2.3
2
2
2.37
2.7
2
2.5
2.1
3.4
3
2
SFO
30
6
6.03
6.1
3.87
4
3.4
4.03
6.47
4.3
6.2
10
5.37
6.03
100
5.9
6.1
6.1
3.03
3
3.1
2.9
6.23
4.3
6.3
10
5.47
5.97
200
5
5
6.03
3.07
3
3.23
3.07
5.4
4.27
6.17
10
5.53
5
WOA
30
3
3
2.53
10
10
9.87
9.7
3
7.73
4
7.07
4
3.23
100
3
3
2.73
10
10
9.5
8.47
3
6.4
3.87
6.93
4
3
200
3
3
2.7
10
9.93
9.03
7.97
3
6.63
3.77
6.83
4
3
Function
Dimensionality
F14
F15
F16
F17
F18
F19
F20
F21
F22
F23
Avg. rank
Overall rank
MGEHO
30
1
5.83
4.87
1
1
1
1
7
1
1
2.21
1
100
1
4.03
4.87
1
1
1
1
7.53
1
1
2.06
1
200
1
4.07
4.23
1
1.03
1
1
6.83
1
1
2.09
1
EHO
30
8.17
4.6
4.43
7.73
8.17
8.73
8.23
7.33
6.13
5.63
6.94
8
100
8.43
3.6
3.67
7.3
7.93
8.7
8.37
6.97
5.57
5.77
6.6
7
200
8.47
2.93
3.9
7.23
7.43
8
8.43
7
5.7
5.53
6.43
7
EEHO15
30
7.03
4.7
4.87
6.4
6.73
6.97
7.5
10
3.7
3.9
6.11
7
100
7.4
3.97
3.6
6.43
6.93
6.73
7.43
9.97
3.9
3.83
5.79
6
200
7.13
3.87
3.93
6.13
6.93
6.43
7.53
10
3.77
3.87
5.72
6
OEHO
30
8.17
4.8
4.63
8.33
9.4
8.87
8.43
6.33
5.17
5.37
7.15
9
100
8.03
3.47
3.7
7.83
9.37
8.77
8.4
7.27
5.53
5.5
6.92
8
200
7.97
3.5
3.13
7.53
9.5
8.43
8.2
7.33
5.57
5.57
6.74
8
LFEHO
30
9.5
8.4
7.3
9.33
9.1
9.23
9.83
4.33
6.83
7.67
8.25
10
100
9.63
6
6.23
9.17
9.4
9.47
9.77
4.7
7.5
7.53
7.98
10
200
9.53
6.03
6.73
8.93
9.33
9.47
9.8
5.23
7
7.7
7.95
10
IGWO
30
3.8
5.03
7.57
8.1
2
5
5
1
10
9.93
6
6
100
5.8
10
9.53
9.1
5
5.63
5
1
10
10
7.28
9
200
6.9
10
9.93
9.83
5.23
7.6
6.03
1.03
10
10
7.83
9
EO
30
2.77
1.7
5.17
3.97
3
4
4
3
8.2
7.87
4.3
3
100
2
8.6
9.47
3.97
2
4
4
2
8.53
8.07
5.22
4
200
2
8.97
9.07
3.9
2.03
4
4
2.03
8.7
8.57
5.4
5
HHO
30
3.77
4.3
3.13
2.63
4
2.93
2
8
2
2
2.91
2
100
3
2.57
3
2.63
3
3
2
7.23
2
2
2.66
2
200
3
2.97
2.9
2.87
3.03
3
2
5.97
2
2
2.64
2
SFO
30
5.7
5.8
3.13
5.07
6.6
6.2
6
6
4.6
3.3
5.4
4
100
5.47
4.43
2.93
5.1
6.37
5.7
6.03
5.33
3.93
3.8
5.11
3
200
5
4.7
3.17
5.07
6.43
5.07
5
6.5
4.13
3.83
4.99
3
WOA
30
5.1
9.83
9.9
2.43
5
2.07
3
2
7.37
8.33
5.75
5
100
4.23
8.33
8
2.47
4
2
3
3
7.03
7.5
5.37
5
200
4
7.97
8
2.5
4.03
2
3
3.07
7.13
6.93
5.21
4
The results of the Wilcoxon signed-rank test are shown in Table 11, where the symbols “+/=/–” represent that MGEHO is better than, equal to, or worse than the currently compared algorithm. p values less than 0.05 indicate that there are statistically significant differences in the results. In the table, NaN indicates that the algorithm achieved the theoretically best solution for the corresponding problem; therefore, the comparison is not meaningful. In addition, the bold values in the table indicate values greater than 5\(\%\). According to the experimental results, for the comparison between MGEHO and EHO and its variants, there are no statistically significant differences between MGEHO and EHO on F7 and F24; between MGEHO and EEHO15 on F15 and F16; between MGEHO and OEHO on F15, F16, F24, and F31; and between MGEHO and LFEHO on F10, F30, and F31. According to the comparison with other metaheuristics, there are no statistically significant differences between MGEHO and the IGWO on F15; between MGEHO and the EO on F16 and F25; between MGEHO and HHO on F25; between MGEHO and the SFO on F15 and F27; and between MGEHO and the WOA on F14, F18, and F30. Overall, MGEHO is significantly different from most of the algorithms in terms of the obtained results. In addition, MGEHO outperforms the other algorithms on most of the benchmark functions. To make the results more convincing, the Friedman test at a significance level of \(p = 0.05\) is performed. The differences between the proposed MGEHO algorithm and other algorithms are tested by comparing the average rankings of the different algorithms on all benchmark functions; these rankings are shown in Tables 12 and 13. According to the literature, the algorithm with the lowest ranking is considered the most efficient algorithm. By comparison, MGEHO ranks first with different dimensions (\(d = 30,100,200\)), proving that the proposed algorithm has the largest competitive advantage.
Table 13
Nonparametric statistical comparison based on the Friedman rank test for F24–F33
Function
F24
F25
F26
F27
F28
F29
F30
F31
F32
F33
Avg. rank
Overall rank
MGEHO
2.13
3.57
1.53
5.2
4.87
2.97
3.87
1.83
1.2
1.73
2.89
1
EHO
3
8.63
9.3
9
9.33
9.37
9.4
5.7
5.43
4.77
7.39
10
EEHO15
5.63
7.73
2.8
3.2
6.4
7.07
7.37
3.2
2.5
2.43
4.83
4
OEHO
2.6
9.07
9.13
8.7
8.67
8.87
9.33
5.13
5.4
5.5
7.24
9
LFEHO
3.4
6.13
7.7
8.7
7.93
6.1
4.7
5.57
5.67
5.8
6.17
7
IGWO
7.57
1.63
4.97
4.23
1.53
4.2
1.73
1.87
2.33
1.93
3.2
2
EO
8.37
3.97
5.77
5.23
1.53
4.67
2.73
3.6
3.93
4.17
4.4
3
HHO
6.37
3.7
3.43
2.9
2.93
2.37
5.6
9.8
9.73
9.7
5.65
6
SFO
9.87
4.47
7.07
5.4
7.57
5.93
6.83
7.7
7.73
7.73
7.03
8
WOA
6.07
6.1
3.3
2.43
4.23
3.47
3.43
7.6
8.07
8.23
5.29
5
4.1.7 Diversity analysis of MGEHO
To measure the performance of exploration and exploitation in metaheuristics, some researchers have tried indirect research methods, such as diversity analysis. In terms of the exploration and exploitation process, high diversity means that the solution is traversing and can spread across locations in the search space, thus allowing for extensive exploration. Reduced diversity then means that the agents search within a certain neighborhood and focuses on the process of exploitation. Diversity is calculated by the following equation:
where Div represents the overall diversity of the population, \({\text {Div}_j}\) is the diversity of the data set consisting of the data in j-th dimension, m is the dimension of the search agent, and n is the population size, \(\text {median} \left( {{x^j}} \right)\) stands for the median of the data in j-th dimension, and \(x_i^j\) is the j-th dimension of i-th individual.
Figure 4 reflects the difference in diversity between the original EHO and MGEHO. The horizontal axis indicates the number of iterations and the vertical axis indicates the average distance between particles. It can be seen that both algorithms exhibit high diversity in the beginning phase due to the random initialization operation. Then, the diversity changes under the influence of different algorithmic mechanisms As the iterations run, the search agents traverse the whole search space and the average distance gradually becomes smaller, making the diversity decrease, and the algorithms gradually turn to the exploitation, which is consistent with the above analysis. In general, the diversity of MGEHO is significantly enhanced based on EHO. The reason lies in the additional exploration and variation strategies applied in MGEHO. These strategies lead to an enhanced level of diversity in the population, and therefore the probability of the algorithm falling into a locally optimal solution during the iterations is reduced.
4.1.8 Sensitivity analysis on high-dimensional problems
To further evaluate the effectiveness of MGEHO when optimizing complex high-dimensional problems, 23 benchmarks of F1–F23 are expanded to two higher dimension types, 500, 1000. The varying dimensions are set to evaluate the dimension impact on the algorithm. Under the consideration of fairness, other conditions remain the same as in the previous experiments. Tables 14 and 15 shows the results of MGEHO with other algorithms of 23 benchmark functions on 500 and 1000 dimensions. The best solutions for each function are in bold. Moreover, based on the Friedman test in Table 16, the average ranking of MGEHO is 2.11 and 2.25 on 500 and 1000 dimensions, which are lower than the ranking results of all involved variants of the algorithm. From this aspect, it can be confirmed that MGEHO is superior than the above algorithms in optimizing complex high-dimensional problems.
Table 14
Comparison of results obtained on the F1–F23 benchmark functions (\(d=500\))
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
Mean
0.00E\(+\)00
1.37E\(-\)03
6.60E\(-\)05
1.23E\(-\)03
1.61E\(-\)01
1.10E\(-\)02
1.23E\(-\)22
1.11E\(-\)94
4.12E\(-\)09
5.82E\(-\)69
Std
0.00E\(+\)00
2.31E\(-\)04
3.99E\(-\)03
2.58E\(-\)01
2.42E\(-\)03
1.09E\(-\)22
6.09E\(-\)94
8.65E\(-\)09
3.18E\(-\)68
0.00E\(+\)00
F2
Mean
0.00E\(+\)00
2.94E\(-\)03
1.13E\(-\)04
4.22E\(-\)02
2.16E\(-\)01
2.34E\(-\)02
1.07E\(-\)22
4.40E\(-\)94
1.50E\(-\)07
5.94E\(-\)70
Std
0.00E\(+\)00
6.93E\(-\)03
2.89E\(-\)04
5.16E\(-\)02
4.51E\(-\)01
9.67E\(-\)03
1.13E\(-\)22
8.12E\(-\)94
2.14E\(-\)07
1.95E\(-\)69
F3
Mean
0.00E\(+\)00
8.16E\(-\)02
1.30E\(-\)02
1.19E\(-\)01
4.30E\(-\)01
5.01E\(-\)01
7.22E\(-\)14
5.09E\(-\)50
4.91E\(-\)04
3.09E\(-\)48
Std
0.00E\(+\)00
6.34E\(-\)02
1.24E\(-\)02
1.05E\(-\)01
4.20E\(-\)01
9.37E\(-\)01
5.75E\(-\)14
1.57E\(-\)49
5.16E\(-\)04
5.14E\(-\)48
F4
Mean
0.00E\(+\)00
5.46E\(+\)01
4.36E\(+\)00
1.73E\(+\)02
1.47E\(+\)02
6.64E\(+\)05
1.99E\(+\)04
1.87E\(-\)47
2.98E\(-\)04
2.94E\(+\)07
Std
0.00E\(+\)00
7.03E\(+\)01
6.39E\(+\)00
5.21E\(+\)02
1.58E\(+\)02
5.55E\(+\)04
1.85E\(+\)04
5.89E\(-\)47
3.98E\(-\)04
1.01E\(+\)07
F5
Mean
0.00E\(+\)00
7.06E\(-\)04
1.75E\(-\)04
7.36E\(-\)04
2.69E\(-\)02
5.59E\(+\)01
8.04E\(+\)01
4.70E\(-\)49
1.02E\(-\)06
8.39E\(+\)01
Std
0.00E\(+\)00
5.07E\(-\)04
1.98E\(-\)04
8.36E\(-\)04
2.59E\(-\)02
3.06E\(+\)00
1.85E\(+\)01
1.12E\(-\)48
9.56E\(-\)07
2.44E\(+\)01
F6
Mean
5.77E\(+\)01
7.72E\(-\)01
8.22E\(-\)01
3.97E\(+\)00
9.36E\(+\)01
9.01E\(+\)03
4.98E\(+\)02
4.46E\(-\)01
5.90E\(-\)01
4.96E\(+\)02
Std
1.55E\(+\)02
1.51E\(+\)00
1.43E\(+\)00
4.43E\(+\)00
1.58E\(+\)02
5.75E\(+\)03
1.16E\(-\)01
5.72E\(-\)01
9.79E\(-\)01
3.15E\(-\)01
F7
Mean
4.58E\(-\)01
9.01E\(-\)03
2.38E\(-\)03
8.73E\(-\)03
1.22E\(+\)01
8.70E\(+\)01
8.63E\(+\)01
3.45E\(-\)03
1.33E\(-\)01
2.95E\(+\)01
Std
4.12E\(-\)01
1.24E\(-\)02
2.52E\(-\)03
1.09E\(-\)02
1.27E\(+\)01
2.60E\(+\)00
1.98E\(+\)00
6.65E\(-\)03
2.15E\(-\)01
4.26E\(+\)00
F8
Mean
0.00E\(+\)00
6.98E\(-\)14
9.39E\(-\)17
1.86E\(-\)09
5.98E\(-\)08
1.92E\(-\)03
1.54E\(-\)37
3.02E\(-\)200
3.60E\(-\)17
4.92E\(-\)107
Std
0.00E\(+\)00
7.18E\(-\)14
2.49E\(-\)16
5.02E\(-\)09
1.64E\(-\)07
2.50E\(-\)03
2.73E\(-\)37
0.00E\(+\)00
1.13E\(-\)16
1.54E\(-\)106
F9
Mean
4.11E\(-\)05
1.43E\(-\)03
4.27E\(-\)03
2.38E\(-\)03
1.03E\(-\)02
6.43E\(-\)01
4.12E\(-\)03
1.46E\(-\)04
7.45E\(-\)04
4.73E\(-\)03
Std
3.38E\(-\)05
1.96E\(-\)03
3.31E\(-\)03
2.95E\(-\)03
1.10E\(-\)02
2.78E\(-\)01
1.48E\(-\)03
1.02E\(-\)04
8.95E\(-\)04
5.30E\(-\)03
F10
Mean
0.00E\(+\)00
8.21E\(-\)11
9.31E\(-\)12
1.57E\(-\)07
3.80E\(-\)09
1.82E\(-\)19
1.15E\(-\)01
3.13E\(-\)124
3.72E\(-\)14
9.19E\(-\)111
Std
0.00E\(+\)00
1.29E\(-\)10
2.21E\(-\)11
2.22E\(-\)07
5.92E\(-\)09
5.62E\(-\)19
2.59E\(-\)01
9.88E\(-\)124
3.61E\(-\)14
2.86E\(-\)110
F11
Mean
\(-\)2.09E\(+\)05
\(-\)2.09E\(+\)05
\(-\)2.09E\(+\)05
\(-\)2.09E\(+\)05
\(-\)2.09E\(+\)05
\(-\)3.96E\(+\)04
\(-\)7.57E\(+\)04
\(-\)2.09E\(+\)05
\(-\)1.48E\(+\)04
\(-\)1.81E\(+\)05
Std
1.73E\(+\)01
8.53E\(+\)01
4.83E\(+\)00
1.31E\(+\)02
2.10E\(+\)01
2.25E\(+\)04
3.20E\(+\)03
3.57E\(+\)01
1.07E\(+\)03
2.79E\(+\)04
F12
Mean
0.00E\(+\)00
4.14E\(-\)04
3.62E\(-\)05
3.58E\(-\)02
7.96E\(-\)02
4.27E\(+\)02
0.00E\(+\)00
0.00E\(+\)00
5.46E\(-\)06
0.00E\(+\)00
Std
0.00E\(+\)00
7.34E\(-\)04
8.78E\(-\)05
6.56E\(-\)02
7.36E\(-\)02
7.35E\(+\)01
0.00E\(+\)00
0.00E\(+\)00
6.58E\(-\)06
0.00E\(+\)00
F13
Mean
8.88E\(-\)16
5.55E\(-\)04
3.02E\(-\)04
1.25E\(-\)03
2.35E\(-\)02
5.42E\(-\)03
6.54E\(-\)13
8.88E\(-\)16
3.29E\(-\)06
4.80E\(-\)15
Std
0.00E\(+\)00
4.22E\(-\)04
3.85E\(-\)04
1.26E\(-\)03
3.53E\(-\)02
1.67E\(-\)03
3.28E\(-\)13
0.00E\(+\)00
4.21E\(-\)06
2.02E\(-\)15
F14
Mean
0.00E\(+\)00
3.94E\(-\)04
2.08E\(-\)05
1.12E\(-\)04
1.45E\(-\)02
1.26E\(-\)02
9.99E\(-\)17
0.00E\(+\)00
2.58E\(-\)12
0.00E\(+\)00
Std
0.00E\(+\)00
5.39E\(-\)04
3.59E\(-\)05
1.42E\(-\)04
2.21E\(-\)02
2.42E\(-\)02
3.51E\(-\)17
0.00E\(+\)00
2.78E\(-\)12
0.00E\(+\)00
F15
Mean
2.14E\(-\)04
9.54E\(-\)06
5.60E\(-\)06
3.35E\(-\)06
8.30E\(-\)04
1.34E\(+\)03
5.92E\(-\)01
1.55E\(-\)06
7.12E\(-\)05
6.86E\(-\)02
Std
3.36E\(-\)04
1.23E\(-\)05
6.60E\(-\)06
3.22E\(-\)06
1.71E\(-\)03
3.91E\(+\)03
2.13E\(-\)02
2.02E\(-\)06
2.05E\(-\)04
3.15E\(-\)02
F16
Mean
4.09E\(-\)02
1.11E\(-\)03
2.01E\(-\)03
6.89E\(-\)03
5.15E\(-\)01
2.62E\(+\)03
4.94E\(+\)01
4.23E\(-\)04
8.17E\(-\)04
1.64E\(+\)01
Std
3.43E\(-\)02
2.33E\(-\)03
2.02E\(-\)03
1.33E\(-\)02
1.07E\(+\)00
2.45E\(+\)03
1.63E\(-\)01
6.47E\(-\)04
9.41E\(-\)04
4.11E\(+\)00
F17
Mean
0.00E\(+\)00
9.26E\(-\)03
7.44E\(-\)04
6.70E\(-\)03
6.33E\(-\)02
5.02E\(-\)01
2.02E\(-\)14
5.49E\(-\)40
8.67E\(-\)05
6.46E\(-\)50
Std
0.00E\(+\)00
1.83E\(-\)02
8.45E\(-\)04
4.67E\(-\)03
5.82E\(-\)02
5.52E\(-\)01
1.00E\(-\)14
1.74E\(-\)39
8.79E\(-\)05
1.79E\(-\)49
F18
Mean
\(-\)4.41E\(-\)13
4.97E\(-\)06
5.27E\(-\)07
1.31E\(-\)03
7.58E\(-\)04
4.18E\(-\)06
\(-\)4.41E\(-\)13
\(-\)4.41E\(-\)13
8.86E\(-\)08
\(-\)4.38E\(-\)13
Std
0.00E\(+\)00
5.98E\(-\)06
1.57E\(-\)06
2.98E\(-\)03
1.04E\(-\)03
9.42E\(-\)07
0.00E\(+\)00
0.00E\(+\)00
1.57E\(-\)07
3.43E\(-\)15
F19
Mean
0.00E\(+\)00
1.26E\(+\)01
3.26E\(+\)00
1.57E\(+\)01
2.41E\(+\)01
2.48E\(+\)01
3.87E\(-\)06
4.99E\(-\)24
1.33E\(+\)00
2.64E\(-\)29
Std
0.00E\(+\)00
5.18E\(+\)00
1.60E\(+\)00
7.03E\(+\)00
1.43E\(+\)01
2.84E\(+\)00
8.38E\(-\)07
9.37E\(-\)24
7.65E\(-\)01
6.19E\(-\)29
F20
Mean
0.00E\(+\)00
9.81E\(+\)01
8.83E\(+\)01
1.42E\(+\)02
2.89E\(+\)03
1.12E\(+\)01
5.42E\(-\)19
2.71E\(-\)91
1.35E\(-\)04
5.31E\(-\)67
Std
0.00E\(+\)00
1.18E\(+\)02
2.41E\(+\)02
2.50E\(+\)02
3.15E\(+\)03
3.83E\(+\)00
7.97E\(-\)19
6.16E\(-\)91
1.57E\(-\)04
1.53E\(-\)66
F21
Mean
\(-\)9.00E\(-\)01
\(-\)5.23E\(-\)01
\(-\)1.00E\(+\)00
\(-\)6.66E\(-\)01
\(-\)2.57E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
\(-\)1.26E\(-\)54
\(-\)5.59E\(-\)01
0.00E\(+\)00
Std
1.02E\(-\)01
4.28E\(-\)01
3.19E\(-\)05
3.79E\(-\)01
3.60E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
3.99E\(-\)54
3.99E\(-\)01
0.00E\(+\)00
F22
Mean
0.00E\(+\)00
3.09E\(-\)02
1.08E\(-\)03
5.51E\(-\)02
1.36E\(-\)01
1.60E\(+\)00
2.30E\(-\)01
1.73E\(-\)47
3.32E\(-\)03
7.99E\(-\)02
Std
0.00E\(+\)00
4.13E\(-\)02
2.13E\(-\)03
5.89E\(-\)02
1.14E\(-\)01
1.33E\(-\)01
4.83E\(-\)02
5.41E\(-\)47
4.55E\(-\)03
4.21E\(-\)02
F23
Mean
0.00E\(+\)00
1.89E\(-\)02
9.51E\(-\)04
1.54E\(-\)02
1.99E\(-\)01
5.00E\(-\)01
3.01E\(-\)01
0.00E\(+\)00
6.04E\(-\)05
5.63E\(-\)02
Std
0.00E\(+\)00
2.01E\(-\)02
2.53E\(-\)03
2.07E\(-\)02
1.61E\(-\)01
1.27E\(-\)05
1.27E\(-\)13
0.00E\(+\)00
8.43E\(-\)05
8.84E\(-\)02
Table 15
Comparison of results obtained on the F1–F23 benchmark functions (\(d=1000\))
Function
MGEHO
EHO
EEHO15
OEHO
LFEHO
IGWO
EO
HHO
SFO
WOA
F1
Mean
0.00E\(+\)00
1.40E\(-\)03
1.40E\(-\)05
5.36E\(-\)04
2.21E\(-\)01
3.28E\(+\)00
5.58E\(-\)21
2.08E\(-\)90
3.98E\(-\)09
1.58E\(-\)73
Std
0.00E\(+\)00
2.44E\(-\)03
3.36E\(-\)05
5.43E\(-\)04
4.94E\(-\)01
1.30E\(+\)00
4.30E\(-\)21
6.58E\(-\)90
5.40E\(-\)09
3.77E\(-\)73
F2
Mean
0.00E\(+\)00
1.35E\(-\)02
3.22E\(-\)04
7.07E\(-\)02
7.95E\(-\)01
9.46E\(+\)00
1.96E\(-\)20
4.53E\(-\)92
1.44E\(-\)06
7.33E\(-\)71
Std
0.00E\(+\)00
1.63E\(-\)02
8.65E\(-\)04
9.09E\(-\)02
1.13E\(+\)00
1.87E\(+\)00
2.04E\(-\)20
1.43E\(-\)91
1.96E\(-\)06
2.05E\(-\)70
F3
Mean
0.00E\(+\)00
8.44E\(-\)02
3.19E\(-\)02
3.83E\(-\)01
9.62E\(-\)01
1.74E\(+\)07
4.20E\(-\)13
2.17E\(-\)49
8.67E\(-\)04
1.61E\(-\)49
Std
0.00E\(+\)00
5.53E\(-\)02
2.74E\(-\)02
4.43E\(-\)01
1.24E\(+\)00
3.67E\(+\)06
1.66E\(-\)13
3.67E\(-\)49
5.60E\(-\)04
2.70E\(-\)49
F4
Mean
0.00E\(+\)00
1.99E\(+\)02
1.29E\(+\)02
9.22E\(+\)02
2.88E\(+\)03
2.78E\(+\)06
2.37E\(+\)05
2.96E\(-\)18
9.80E\(-\)04
1.19E\(+\)08
Std
0.00E\(+\)00
3.15E\(+\)02
2.38E\(+\)02
1.26E\(+\)03
3.17E\(+\)03
3.75E\(+\)05
2.66E\(+\)05
9.35E\(-\)18
1.86E\(-\)03
6.18E\(+\)07
F5
Mean
0.00E\(+\)00
1.15E\(-\)03
3.72E\(-\)04
1.18E\(-\)03
1.44E\(-\)02
6.30E\(+\)01
7.89E\(+\)01
1.39E\(-\)46
8.27E\(-\)07
7.69E\(+\)01
Std
0.00E\(+\)00
1.30E\(-\)03
3.46E\(-\)04
1.29E\(-\)03
1.77E\(-\)02
1.99E\(+\)00
1.51E\(+\)01
4.34E\(-\)46
7.58E\(-\)07
2.29E\(+\)01
F6
Mean
3.46E\(+\)00
4.84E\(+\)01
5.39E\(+\)00
5.24E\(+\)00
2.48E\(+\)02
2.40E\(+\)07
9.97E\(+\)02
5.23E\(-\)01
1.35E\(+\)00
9.94E\(+\)02
Std
6.21E\(+\)00
6.53E\(+\)01
1.09E\(+\)01
9.93E\(+\)00
4.22E\(+\)02
1.63E\(+\)07
8.63E\(-\)02
7.42E\(-\)02
1.96E\(+\)00
1.04E\(+\)00
F7
Mean
5.27E\(-\)01
1.96E\(-\)02
5.24E\(-\)03
1.00E\(-\)01
1.44E\(+\)00
2.04E\(+\)02
2.07E\(+\)02
1.48E\(-\)02
1.15E\(+\)00
5.93E\(+\)01
Std
5.06E\(-\)01
1.38E\(-\)02
6.25E\(-\)03
1.81E\(-\)01
2.70E\(+\)00
2.04E\(+\)00
1.34E\(+\)00
1.35E\(-\)02
2.05E\(+\)00
1.67E\(+\)01
F8
Mean
0.00E\(+\)00
1.25E\(-\)14
9.30E\(-\)17
1.99E\(-\)09
4.48E\(-\)06
2.14E\(+\)02
4.24E\(-\)35
2.54E\(-\)204
5.39E\(-\)18
2.80E\(-\)105
Std
0.00E\(+\)00
2.18E\(-\)14
9.91E\(-\)17
4.44E\(-\)09
1.00E\(-\)05
1.32E\(+\)02
5.48E\(-\)35
0.00E\(+\)00
1.18E\(-\)17
6.20E\(-\)105
F9
Mean
6.87E\(-\)05
1.16E\(-\)02
4.30E\(-\)03
1.07E\(-\)02
8.31E\(-\)03
1.45E\(+\)02
6.57E\(-\)03
1.36E\(-\)04
3.49E\(-\)04
4.57E\(-\)03
Std
3.08E\(-\)05
1.64E\(-\)02
6.90E\(-\)03
1.87E\(-\)02
1.21E\(-\)02
4.22E\(+\)01
9.92E\(-\)04
1.09E\(-\)04
2.47E\(-\)04
4.25E\(-\)03
F10
Mean
0.00E\(+\)00
4.21E\(-\)09
6.92E\(-\)12
4.16E\(-\)08
6.62E\(-\)09
1.01E\(-\)16
2.40E\(-\)01
1.09E\(-\)118
4.12E\(-\)13
6.41E\(-\)110
Std
0.00E\(+\)00
8.82E\(-\)09
1.24E\(-\)11
7.89E\(-\)08
1.09E\(-\)08
1.86E\(-\)16
2.26E\(-\)01
2.43E\(-\)118
8.58E\(-\)13
1.43E\(-\)109
F11
Mean
\(-\)4.19E\(+\)05
\(-\)4.19E\(+\)05
\(-\)4.19E\(+\)05
\(-\)4.19E\(+\)05
\(-\)4.19E\(+\)05
\(-\)1.04E\(+\)05
\(-\)1.11E\(+\)05
\(-\)4.19E\(+\)05
\(-\)2.45E\(+\)04
\(-\)3.05E\(+\)05
Std
7.33E\(-\)02
2.41E\(+\)02
9.39E\(-\)01
1.68E\(+\)02
2.11E\(+\)02
3.85E\(+\)04
1.01E\(+\)04
7.13E\(+\)01
4.58E\(+\)03
2.16E\(+\)04
F12
Mean
0.00E\(+\)00
8.89E\(-\)04
1.40E\(-\)04
8.88E\(-\)02
7.81E\(-\)02
9.39E\(+\)02
3.55E\(-\)16
0.00E\(+\)00
2.47E\(-\)06
0.00E\(+\)00
Std
0.00E\(+\)00
9.85E\(-\)04
2.61E\(-\)04
1.94E\(-\)01
6.20E\(-\)02
2.00E\(+\)02
7.94E\(-\)16
0.00E\(+\)00
2.82E\(-\)06
0.00E\(+\)00
F13
Mean
8.88E\(-\)16
1.02E\(-\)03
4.27E\(-\)04
1.88E\(-\)03
3.27E\(-\)02
6.60E\(-\)02
1.29E\(-\)12
8.88E\(-\)16
9.75E\(-\)06
4.44E\(-\)15
Std
0.00E\(+\)00
9.93E\(-\)04
4.62E\(-\)04
2.35E\(-\)03
2.49E\(-\)02
2.23E\(-\)02
6.68E\(-\)13
0.00E\(+\)00
8.26E\(-\)06
3.55E\(-\)15
F14
Mean
0.00E\(+\)00
4.85E\(-\)04
1.32E\(-\)05
1.37E\(-\)04
3.10E\(-\)03
2.11E\(-\)01
1.78E\(-\)16
0.00E\(+\)00
7.63E\(-\)13
0.00E\(+\)00
Std
0.00E\(+\)00
6.28E\(-\)04
2.31E\(-\)05
1.61E\(-\)04
3.73E\(-\)03
1.98E\(-\)02
6.08E\(-\)17
0.00E\(+\)00
6.32E\(-\)13
0.00E\(+\)00
F15
Mean
4.00E\(-\)05
8.17E\(-\)06
1.12E\(-\)05
1.35E\(-\)05
4.24E\(-\)04
1.47E\(+\)07
8.26E\(-\)01
9.04E\(-\)07
2.96E\(-\)05
9.99E\(-\)02
Std
3.71E\(-\)05
1.37E\(-\)05
1.79E\(-\)05
1.82E\(-\)05
6.87E\(-\)04
5.63E\(+\)06
2.02E\(-\)02
7.90E\(-\)07
4.70E\(-\)05
4.38E\(-\)02
F16
Mean
2.85E\(-\)02
1.47E\(-\)03
2.83E\(-\)03
9.27E\(-\)03
1.30E\(+\)00
5.77E\(+\)07
9.94E\(+\)01
5.43E\(-\)04
1.58E\(-\)03
3.19E\(+\)01
Std
4.48E\(-\)02
2.61E\(-\)03
5.70E\(-\)03
8.27E\(-\)03
1.74E\(+\)00
3.21E\(+\)07
2.31E\(-\)01
6.77E\(-\)04
1.94E\(-\)03
7.89E\(+\)00
F17
Mean
0.00E\(+\)00
2.06E\(-\)03
5.57E\(-\)03
1.16E\(-\)02
7.38E\(-\)02
4.59E\(+\)00
1.28E\(-\)13
6.90E\(-\)51
1.57E\(-\)04
9.15E\(-\)48
Std
0.00E\(+\)00
1.53E\(-\)03
8.38E\(-\)03
1.14E\(-\)02
5.16E\(-\)02
3.81E\(+\)00
1.29E\(-\)13
1.28E\(-\)50
1.41E\(-\)04
1.53E\(-\)47
F18
Mean
1.41E\(-\)12
4.05E\(-\)06
1.41E\(-\)07
5.23E\(-\)04
1.85E\(-\)04
6.43E\(-\)04
1.42E\(-\)12
1.41E\(-\)12
1.82E\(-\)07
1.42E\(-\)12
Std
0.00E\(+\)00
5.41E\(-\)06
2.78E\(-\)07
8.07E\(-\)04
2.05E\(-\)04
1.42E\(-\)04
0.00E\(+\)00
0.00E\(+\)00
1.69E\(-\)07
7.78E\(-\)15
F19
Mean
0.00E\(+\)00
2.18E\(+\)01
4.11E\(+\)00
4.03E\(+\)01
5.96E\(+\)01
8.55E\(+\)01
9.93E\(-\)06
1.33E\(-\)25
3.66E\(+\)00
1.82E\(-\)28
Std
0.00E\(+\)00
8.27E\(+\)00
2.41E\(+\)00
1.22E\(+\)01
3.36E\(+\)01
6.28E\(+\)00
3.53E\(-\)06
2.75E\(-\)25
2.39E\(+\)00
3.59E\(-\)28
F20
Mean
0.00E\(+\)00
6.79E\(+\)01
3.67E\(+\)00
7.02E\(+\)00
1.06E\(+\)04
3.27E\(+\)03
2.41E\(-\)17
1.43E\(-\)87
1.81E\(-\)04
4.92E\(-\)66
Std
0.00E\(+\)00
5.13E\(+\)01
5.45E\(+\)00
1.36E\(+\)01
1.58E\(+\)04
9.41E\(+\)02
2.04E\(-\)17
3.21E\(-\)87
8.59E\(-\)05
7.09E\(-\)66
F21
Mean
\(-\)9.77E\(-\)01
\(-\)6.94E\(-\)01
\(-\)8.00E\(-\)01
\(-\)4.79E\(-\)01
\(-\)1.88E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
\(-\)5.57E\(-\)01
0.00E\(+\)00
Std
2.39E\(-\)02
2.97E\(-\)01
4.47E\(-\)01
4.44E\(-\)01
4.03E\(-\)01
0.00E\(+\)00
0.00E\(+\)00
0.00E\(+\)00
4.16E\(-\)01
0.00E\(+\)00
F22
Mean
0.00E\(+\)00
1.78E\(-\)03
8.76E\(-\)05
4.80E\(-\)03
4.18E\(-\)02
2.00E\(-\)01
9.99E\(-\)02
8.50E\(-\)53
2.03E\(-\)03
1.20E\(-\)01
Std
0.00E\(+\)00
2.30E\(-\)03
6.86E\(-\)05
4.58E\(-\)03
3.56E\(-\)02
2.62E\(-\)09
9.35E\(-\)12
1.71E\(-\)52
2.14E\(-\)03
8.36E\(-\)02
F23
Mean
0.00E\(+\)00
2.32E\(-\)02
4.07E\(-\)04
2.50E\(-\)02
1.30E\(-\)01
5.00E\(-\)01
3.29E\(-\)01
0.00E\(+\)00
2.20E\(-\)04
8.64E\(-\)02
Std
0.00E+00
1.99E-02
8.07E-04
1.80E-02
8.89E-02
3.00E-07
6.34E-02
0.00E\(+\)00
2.38E-04
1.21E-01
Table 16
Results of Friedman rank test of high-dimensional problems
Function
Dimensionality
F1
F2
F3
F4
F5
F6
F7
F8
F9
F10
F11
F12
F13
MGEHO
500
1.00
1.00
1.00
1.00
1.00
4.10
6.20
1.00
1.30
1.00
2.20
2.50
1.55
1000
1.00
1.00
1.00
1.00
1.00
3.00
5.60
3
5.60
1.00
2.00
1.00
1.80
EHO
500
7.40
6.91
7.70
5.90
5.50
3.20
3.20
7.90
8.20
6.90
6.50
7.00
6.40
1000
7.40
7.40
7.00
4.90
5.50
3.80
2.80
4.4
4.00
6.60
7.20
7.60
3.40
EEHO15
500
6.50
6.36
6.20
4.70
4.30
3.10
2.80
7.40
4.60
7.40
4.80
6.80
7.30
1000
6.10
6.00
6.40
4.90
4.50
3.80
3.40
3.8
2.80
6.20
4.80
5.80
1.80
OEHO
500
7.37
8.09
7.60
5.10
5.20
4.50
3.50
8.40
5.50
9.60
4.50
8.20
7.40
1000
7.40
7.90
7.90
5.70
5.30
3.80
3.40
3.8
3.40
8.20
5.60
8.40
4.40
LFEHO
500
9.50
9.45
8.50
6.20
7.00
6.40
6.50
8.20
6.60
8.70
4.30
8.70
9.60
1000
9.00
8.60
8.70
6.40
6.70
6.40
5.60
6.4
5.60
8.80
6.20
8.00
4.80
IGWO
500
9.23
7.35
10.00
9.00
8.30
10.00
9.60
10.00
10.00
4.50
8.90
10.00
9.30
1000
10.00
10.00
10.00
9.00
8.20
10.00
9.00
10
9.00
10.00
10.00
4.00
8.60
EO
500
4.00
4.00
4.00
8.00
9.20
9.00
9.40
4.00
7.10
6.20
8.10
2.50
4.00
1000
4.00
4.00
4.00
8.00
9.50
9.00
10.00
9
10.00
4.00
7.40
10.00
8.40
HHO
500
2.00
2.00
2.30
2.00
2.00
3.10
2.90
2.00
2.50
2.10
3.20
2.50
1.55
1000
2.00
2.00
2.40
2.00
2.00
3.00
3.20
3
3.20
2.00
1.80
2.20
4.80
SFO
500
5.00
5.00
5.00
3.10
3.00
3.70
3.10
5.40
4.50
5.70
10.00
5.60
5.00
1000
5.10
5.10
5.00
3.10
3.00
3.60
3.40
3.6
3.40
5.20
3.20
5.20
10.00
WOA
500
3.00
3.00
2.70
10.00
9.50
7.90
7.80
3.00
6.10
2.90
7.00
2.50
2.90
1000
3.00
3.00
2.60
10.00
9.30
8.00
8.00
8
8.00
3.00
6.80
2.80
7.00
Function
Dimensionality
F14
F15
F16
F17
F18
F19
F20
F21
F22
F23
Avg. rank
Overall rank
MGEHO
500
1.55
2.05
5.20
1.00
2.65
1.00
1.00
6.20
1.00
1.65
2.11
1
1000
2.40
1.70
2.00
5.80
1.70
1.00
1.00
5.00
1.00
1.60
2.25
1
EHO
500
7.30
7.80
2.80
7.50
7.30
7.50
8.30
7.00
5.70
6.00
6.54
7
1000
7.00
7.40
7.80
4.00
6.60
7.00
8.00
7.20
5.40
5.80
6.02
6
EEHO15
500
6.40
6.40
4.40
5.70
5.90
5.80
7.00
9.90
3.60
3.90
5.79
6
1000
6.00
6.40
6.20
4.40
5.60
5.80
6.20
8.60
4.40
3.60
5.18
5
OEHO
500
7.40
7.20
3.50
7.60
9.60
8.20
7.80
7.50
6.10
5.60
6.81
8
1000
8.20
7.20
7.40
3.60
8.80
8.40
6.80
7.00
5.60
6.00
6.28
7
LFEHO
500
9.60
9.20
5.70
9.00
9.00
8.90
9.50
5.70
7.50
8.10
7.90
9
1000
8.80
9.40
8.60
4.40
8.00
9.00
9.40
9.00
7.40
7.40
7.51
9
IGWO
500
9.30
9.40
10.00
9.90
7.60
9.40
7.40
2.45
10.00
10.00
8.71
10
1000
10.00
9.60
10.00
10.00
9.60
9.60
9.60
2.70
10.00
10.00
9.04
10
EO
500
4.00
3.85
9.00
4.00
2.65
4.00
4.00
2.45
8.80
8.70
5.56
5
1000
2.80
4.00
4.00
9.00
3.70
4.00
4.00
2.70
8.40
9.00
6.36
8
HHO
500
1.55
2.05
2.50
2.40
2.65
3.00
2.00
2.75
2.00
1.65
2.32
2
1000
2.40
1.70
2.00
2.60
1.70
3.00
2.00
2.70
2.00
1.60
2.44
2
SFO
500
5.00
5.00
3.30
5.30
5.60
5.20
5.00
6.60
4.10
3.90
4.96
4
1000
5.00
5.00
5.00
3.20
6.40
5.20
5.00
7.40
3.60
3.80
4.76
3
WOA
500
2.90
2.05
8.00
2.60
2.05
2.00
3.00
2.45
6.20
5.50
4.53
3
1000
2.40
2.60
2.00
8.00
2.90
2.00
3.00
2.70
7.20
6.20
5.10
4
4.2 Experiments on CEC2014 and CEC2017 benchmark functions
To further verify the performance of MGEHO when dealing with complex problems, experiments are conducted on the CEC2014 and CEC2017 composite benchmark functions. Besides, seven advanced metaheuristic algorithms were selected for comparison, they are the improved elephant herding optimization (EHOI) [21], the RUNge Kutta optimizer (RUN) [45], the gradient-based optimizer (GBO) [46], the Hunger Games Search (HGS) [47], the Slime mould algorithm (SMA) [48], the an elite evolutionary strategy based Harris hawks optimization (EESHHO) [49], the multi-trial vector-based differential evolution (MTDE) [50]. It should be noted that the parameter settings of these algorithms are selected from the recent reported literature. Moreover, the population size N for all algorithms is set to 30, the spatial dimension is 30, and the maximum number of iterations \({t_{\max }}\) is 1000. The results are based on 30 runs.
4.2.1 Benchmark functions
In this subsection, 32 benchmark functions from CEC2014 and CEC2017 [51] are employed in this section to demonstrate the efficacy of MGEHO. The descriptions of these functions are presented in Table 17, where D represents the dimension of the function and “search range” is the parameter that constrains the frontier of the search space of the corresponding function. In general, these functions cover four categories: F1– F3 and F16–F17 are unimodal functions, F4–F10 and F18 are multimodal functions, F11–F14 and F19–F28 are mixed-state functions, and F29–F32 are compositional functions. It is well known that unimodal functions are suitable for confirming the exploratory ability of the method because there exists only one optimal solution. Unlike unimodal functions, there are a large number of locally optimal solutions in the multimodal case. The number of these solutions increases with dimensionality, which is used to test the capability of the exploitation and jumping out of the local optimum.
Table 17
Details of 32 CEC2014 and CEC2017 benchmark functions
Type
No.
Description
Fmin
CEC 2014 Test functions (Search range: \([-100,100]^D\))
Unimodal functions
F1
Rotated High Conditioned Elliptic Function
100
F2
Rotated Bent Cigar Function
200
F3
Rotated Discus Function
300
Simple Multimodal functions
F4
Shifted and Rotated Rosenbrock’s Function
400
F5
Shifted and Rotated Ackley’s Function
500
F6
Shifted and Rotated Schwefel’s Function
1100
F7
Shifted and Rotated Katsuura Function
1200
F8
Shifted and Rotated HappyCat Function
1300
F9
Shifted and Rotated HGBat Function
1400
F10
Shifted and Rotated Expanded Griewank’s plus Rosenbrock’s Function
1500
Hybrid functions
F11
Hybrid Function 1 (\({N}=3\))
1700
F12
Hybrid Function 2 (\({N}=3\))
1800
F13
Hybrid Function 3 (\({N}=4\))
1900
F14
Hybrid Function 4 (\({N}=4\))
2000
F15
Hybrid Function 5 (\({N}=5\))
2100
CEC 2017 Test functions (Search range: \([-100,100]^D\))
Unimodal functions
F16
Shifted and Rotated Bent Cigar Function
100
F17
Shifted and Rotated Zakharov Function
300
Simple Multimodal functions
F18
Shifted and Rotated Rastrigin’s Function
500
Hybrid functions
F19
Hybrid Function 1(\({N}=3\))
1100
F20
Hybrid Function 2(\({N}=3\))
1200
F21
Hybrid Function 3(\({N}=3\))
1300
F22
Hybrid Function 4(\({N}=4\))
1400
F23
Hybrid Function 5(\({N}=4\))
1500
F24
Hybrid Function 6(\({N}=4\))
1600
F25
Hybrid Function 6(\({N}=5\))
1700
F26
Hybrid Function 6(\({N}=5\))
1800
F27
Hybrid Function 6(\({N}=5\))
1900
F28
Hybrid Function 6(\({N}=6\))
2000
Composition functions
F29
Composition Function 2(\({N}=3\))
2200
F30
Composition Function 5(\({N}=5\))
2500
F31
Composition Function 8(\({N}=6\))
2800
F32
Composition Function 10(\({N}=3\))
3000
4.2.2 Experimental results and analysis
The comparison results are shown in Table 18. The table reports the mean, standard deviation obtained for 30 runs. The best solutions for each function are in bold. According to the results, it is clear that MGEHO outperforms the comparison algorithm on most of the functions. Compared to the other metaheuristics, MGEHO obtained the first ranking on 18 functions, which is better than any of the comparison algorithms. Although both MGEHO and EHOI are hybrid GWOs, the former is significantly superior to the latter. The main reason is that MGEHO not only uses a mixture of EHO and MRFO as the information sharing method but also utilizes two other information enhancement methods, which contributes to the performance of the algorithm.
Table 18
Results on CEC2014 and CEC2017 benchmark functions
Function
MGEHO
EHOI
RUN
GBO
HGS
SMA
EESHHO
MTDE
F1
Mean
1.34E\(+\)06
1.88E\(+\)07
2.21E\(+\)06
2.16E\(+\)06
1.26E\(+\)07
8.81E\(+\)06
1.53E\(+\)07
5.82E\(+\)08
Std
7.85E\(+\)05
1.96E\(+\)07
1.15E\(+\)06
1.05E\(+\)06
1.05E\(+\)07
5.57E\(+\)06
9.93E\(+\)06
2.00E\(+\)08
F2
Mean
3.31E\(+\)03
1.11E\(+\)07
2.22E\(+\)04
1.09E\(+\)04
4.90E\(+\)05
2.21E\(+\)04
7.22E\(+\)06
1.79E\(+\)10
Std
3.48E\(+\)03
5.18E\(+\)07
1.06E\(+\)04
9.59E\(+\)03
1.53E\(+\)06
1.22E\(+\)04
1.46E\(+\)07
4.29E\(+\)09
F3
Mean
1.84E\(+\)03
2.42E\(+\)04
4.89E\(+\)03
2.43E\(+\)03
1.80E\(+\)04
6.52E\(+\)03
9.31E\(+\)03
9.34E\(+\)04
Std
8.90E\(+\)02
1.94E\(+\)04
1.96E\(+\)03
1.67E\(+\)03
1.65E\(+\)04
6.69E\(+\)03
5.62E\(+\)03
1.43E\(+\)04
F4
Mean
4.74E\(+\)02
5.53E\(+\)02
5.06E\(+\)02
5.15E\(+\)02
5.37E\(+\)02
5.32E\(+\)02
5.43E\(+\)02
2.87E\(+\)03
Std
1.36E\(+\)01
5.16E\(+\)01
2.90E\(+\)01
4.62E\(+\)01
3.66E\(+\)01
3.61E\(+\)01
4.51E\(+\)01
8.97E\(+\)02
F5
Mean
5.18E\(+\)02
5.21E\(+\)02
5.21E\(+\)02
5.21E\(+\)02
5.20E\(+\)02
5.21E\(+\)02
5.20E\(+\)02
5.21E\(+\)02
Std
5.80E\(-\)02
2.53E\(-\)01
2.36E\(-\)01
2.54E\(-\)01
6.98E\(-\)02
7.05E\(-\)02
7.06E\(-\)02
6.18E\(-\)02
F6
Mean
4.80E\(+\)03
6.21E\(+\)03
4.44E\(+\)03
5.09E\(+\)03
4.24E\(+\)03
4.46E\(+\)03
4.49E\(+\)03
9.20E\(+\)03
Std
6.37E\(+\)02
8.72E\(+\)02
5.82E\(+\)02
5.79E\(+\)02
4.41E\(+\)02
6.81E\(+\)02
5.20E\(+\)02
4.52E\(+\)02
F7
Mean
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
1.20E\(+\)03
Std
8.53E\(-\)01
9.06E\(-\)01
9.00E\(-\)01
2.19E\(-\)01
8.44E\(-\)02
2.27E\(-\)01
9.54E\(-\)02
7.34E\(-\)01
F8
Mean
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
1.30E\(+\)03
Std
1.23E\(-\)01
1.41E\(-\)01
8.36E\(-\)02
1.18E\(-\)01
1.12E\(-\)01
1.03E\(-\)01
1.06E\(-\)01
4.65E\(-\)01
F9
Mean
1.40E\(+\)03
1.40E\(+\)03
1.40E\(+\)03
1.40E\(+\)03
1.40E\(+\)03
1.40E\(+\)03
1.40E\(+\)03
1.46E\(+\)03
Std
5.10E\(-\)02
1.90E\(-\)01
1.16E\(-\)01
4.90E\(-\)02
3.58E\(-\)01
3.75E\(-\)01
1.10E\(-\)01
1.29E\(+\)01
F10
Mean
1.51E\(+\)03
1.59E\(+\)03
1.55E\(+\)03
1.52E\(+\)03
1.52E\(+\)03
1.51E\(+\)03
1.53E\(+\)03
3.57E\(+\)04
Std
7.19E\(+\)00
4.12E\(+\)01
1.83E\(+\)01
7.34E\(+\)00
5.19E\(+\)00
3.75E\(+\)00
9.03E\(+\)00
3.24E\(+\)04
F11
Mean
1.97E\(+\)05
3.30E\(+\)06
1.70E\(+\)05
1.41E\(+\)05
1.73E\(+\)06
1.74E\(+\)06
1.81E\(+\)06
2.45E\(+\)07
Std
1.57E\(+\)05
2.66E\(+\)06
1.58E\(+\)05
1.16E\(+\)05
1.20E\(+\)06
1.06E\(+\)06
1.47E\(+\)06
1.07E\(+\)07
F12
Mean
3.80E\(+\)03
2.39E\(+\)04
3.68E\(+\)03
8.57E\(+\)03
1.27E\(+\)04
2.54E\(+\)04
3.74E\(+\)03
1.57E\(+\)08
Std
1.93E\(+\)03
8.76E\(+\)04
1.54E\(+\)03
8.49E\(+\)03
9.21E\(+\)03
1.18E\(+\)04
2.77E\(+\)03
9.24E\(+\)07
F13
Mean
1.92E\(+\)03
1.93E\(+\)03
1.93E\(+\)03
1.92E\(+\)03
1.94E\(+\)03
1.92E\(+\)03
1.92E\(+\)03
2.07E\(+\)03
Std
2.75E\(+\)01
1.52E\(+\)01
3.08E\(+\)01
1.23E\(+\)01
3.97E\(+\)01
2.30E\(+\)01
2.88E\(+\)01
3.91E\(+\)01
F14
Mean
8.64E\(+\)03
3.51E\(+\)04
5.26E\(+\)03
6.24E\(+\)03
3.01E\(+\)04
2.14E\(+\)04
1.68E\(+\)04
7.00E\(+\)04
Std
5.02E\(+\)03
1.61E\(+\)04
2.39E\(+\)03
3.43E\(+\)03
1.81E\(+\)04
1.64E\(+\)04
5.49E\(+\)03
3.42E\(+\)04
F15
Mean
8.40E\(+\)04
8.01E\(+\)05
5.69E\(+\)04
6.91E\(+\)04
1.31E\(+\)06
7.64E\(+\)05
6.93E\(+\)05
7.89E\(+\)06
Std
6.98E\(+\)04
8.35E\(+\)05
4.31E\(+\)04
5.58E\(+\)04
1.03E\(+\)06
5.36E\(+\)05
4.24E\(+\)05
4.98E\(+\)06
F16
Mean
4.56E\(+\)03
4.89E\(+\)06
1.04E\(+\)04
5.55E\(+\)03
2.24E\(+\)07
1.26E\(+\)04
1.91E\(+\)07
1.40E\(+\)10
Std
3.26E\(+\)03
9.83E\(+\)06
8.15E\(+\)03
4.35E\(+\)03
4.01E\(+\)07
7.30E\(+\)03
8.00E\(+\)07
3.74E\(+\)09
F17
Mean
6.45E\(+\)03
4.85E\(+\)04
1.71E\(+\)03
6.53E\(+\)03
2.73E\(+\)04
5.01E\(+\)03
3.77E\(+\)04
1.57E\(+\)05
Std
3.68E\(+\)03
9.08E\(+\)03
9.60E\(+\)02
3.03E\(+\)03
1.10E\(+\)04
3.08E\(+\)03
6.78E\(+\)03
4.39E\(+\)04
F18
Mean
6.97E\(+\)02
6.73E\(+\)02
7.05E\(+\)02
6.62E\(+\)02
6.45E\(+\)02
6.16E\(+\)02
7.10E\(+\)02
8.05E\(+\)02
Std
4.75E\(+\)01
2.98E\(+\)01
4.31E\(+\)01
4.14E\(+\)01
2.65E\(+\)01
2.21E\(+\)01
4.01E\(+\)01
2.40E\(+\)01
F19
Mean
1.20E\(+\)03
1.33E\(+\)03
1.20E\(+\)03
1.26E\(+\)03
1.25E\(+\)03
1.28E\(+\)03
1.24E\(+\)03
5.36E\(+\)03
Std
1.88E\(+\)01
8.56E\(+\)01
3.33E\(+\)01
7.98E\(+\)01
7.56E\(+\)01
5.41E\(+\)01
4.05E\(+\)01
1.39E\(+\)03
F20
Mean
2.06E\(+\)05
5.44E\(+\)06
3.91E\(+\)06
3.28E\(+\)05
3.89E\(+\)06
4.07E\(+\)06
3.15E\(+\)06
1.33E\(+\)09
Std
2.04E\(+\)05
7.45E\(+\)06
2.28E\(+\)06
3.55E\(+\)05
2.08E\(+\)06
2.24E\(+\)06
1.70E\(+\)06
5.67E\(+\)08
F21
Mean
1.36E\(+\)04
7.10E\(+\)05
3.05E\(+\)04
1.92E\(+\)04
4.24E\(+\)04
4.06E\(+\)04
1.83E\(+\)04
3.42E\(+\)08
Std
1.18E\(+\)04
1.49E\(+\)06
1.48E\(+\)04
2.12E\(+\)04
2.90E\(+\)04
2.42E\(+\)04
1.87E\(+\)04
1.88E\(+\)08
F22
Mean
8.98E\(+\)03
2.78E\(+\)05
8.39E\(+\)03
6.22E\(+\)03
2.22E\(+\)05
1.35E\(+\)05
3.30E\(+\)05
1.12E\(+\)06
Std
1.32E\(+\)04
3.40E\(+\)05
1.01E\(+\)04
6.05E\(+\)03
2.28E\(+\)05
8.99E\(+\)04
2.21E\(+\)05
8.38E\(+\)05
F23
Mean
7.20E\(+\)03
2.87E\(+\)04
1.55E\(+\)04
9.98E\(+\)03
1.42E\(+\)04
2.19E\(+\)04
5.35E\(+\)03
7.22E\(+\)07
Std
8.84E\(+\)03
2.97E\(+\)04
2.07E\(+\)03
9.96E\(+\)03
1.30E\(+\)04
1.43E\(+\)04
1.72E\(+\)03
4.92E\(+\)07
F24
Mean
2.72E\(+\)03
3.00E\(+\)03
2.82E\(+\)03
2.62E\(+\)03
2.92E\(+\)03
2.54E\(+\)03
2.87E\(+\)03
4.26E\(+\)03
Std
2.77E\(+\)02
4.17E\(+\)02
3.13E\(+\)02
3.16E\(+\)02
2.85E\(+\)02
2.36E\(+\)02
3.23E\(+\)02
2.97E\(+\)02
F25
Mean
2.20E\(+\)03
2.44E\(+\)03
2.24E\(+\)03
2.34E\(+\)03
2.37E\(+\)03
2.35E\(+\)03
2.26E\(+\)03
2.94E\(+\)03
Std
1.87E\(+\)02
2.38E\(+\)02
2.18E\(+\)02
2.17E\(+\)02
2.34E\(+\)02
2.19E\(+\)02
2.40E\(+\)02
1.88E\(+\)02
F26
Mean
2.04E\(+\)05
2.60E\(+\)06
8.41E\(+\)04
1.19E\(+\)05
2.54E\(+\)06
1.54E\(+\)06
8.67E\(+\)05
1.59E\(+\)07
Std
1.76E\(+\)05
2.78E\(+\)06
3.89E\(+\)04
9.70E\(+\)04
3.42E\(+\)06
1.65E\(+\)06
9.13E\(+\)05
8.38E\(+\)06
F27
Mean
8.10E\(+\)03
2.00E\(+\)05
1.03E\(+\)04
8.54E\(+\)03
2.53E\(+\)04
2.73E\(+\)04
5.47E\(+\)03
9.19E\(+\)07
Std
6.78E\(+\)03
6.43E\(+\)05
1.12E\(+\)04
1.04E\(+\)04
2.21E\(+\)04
2.11E\(+\)04
3.65E\(+\)03
8.42E\(+\)07
F28
Mean
2.57E\(+\)03
2.64E\(+\)03
2.48E\(+\)03
2.57E\(+\)03
2.55E\(+\)03
2.61E\(+\)03
2.56E\(+\)03
3.22E\(+\)03
Std
1.56E\(+\)02
2.18E\(+\)02
1.15E\(+\)02
1.48E\(+\)02
2.43E\(+\)02
1.64E\(+\)02
1.74E\(+\)02
2.16E\(+\)02
F29
Mean
3.03E\(+\)03
6.65E\(+\)03
3.45E\(+\)03
3.75E\(+\)03
5.20E\(+\)03
5.91E\(+\)03
3.95E\(+\)03
9.62E\(+\)03
Std
7.05E\(+\)02
2.62E\(+\)03
1.84E\(+\)03
2.11E\(+\)03
1.74E\(+\)03
8.30E\(+\)02
1.93E\(+\)03
1.86E\(+\)03
F30
Mean
2.90E\(+\)03
2.93E\(+\)03
2.91E\(+\)03
2.90E\(+\)03
2.90E\(+\)03
2.89E\(+\)03
2.93E\(+\)03
3.78E\(+\)03
Std
1.99E\(+\)01
2.49E\(+\)01
1.86E\(+\)01
1.83E\(+\)01
1.31E\(+\)01
1.28E\(+\)01
2.33E\(+\)01
1.60E\(+\)02
F31
Mean
3.20E\(+\)03
3.30E\(+\)03
3.23E\(+\)03
3.23E\(+\)03
3.28E\(+\)03
3.26E\(+\)03
3.28E\(+\)03
4.92E\(+\)03
Std
1.03E\(+\)01
1.34E\(+\)01
2.27E\(+\)01
2.28E\(+\)01
5.74E\(+\)01
4.51E\(+\)01
2.35E\(+\)01
4.20E\(+\)02
F32
Mean
8.67E\(+\)03
4.29E\(+\)05
3.00E\(+\)05
1.06E\(+\)04
1.39E\(+\)05
4.96E\(+\)04
9.18E\(+\)04
7.03E\(+\)07
Std
2.15E\(+\)03
7.22E+05
3.39E+05
4.49E+03
1.70E+05
2.86E+04
9.25E+04
3.53E+07
Figure 5 shows the convergence curves of the algorithm on the three functions. Where the horizontal axis represents the number of function iterations and the vertical axis represents the average results obtained by the algorithm in 30 independent runs. It is noted that MGEHO has better accuracy in finding the best results than any of the algorithms, while the convergence rate can be maintained at a faster level. The experiments show that MGEHO has the advantage of handling complex problems.
The Wilcoxon signed-rank test with a significance level of \(5\%\) was adopted to analyze the comparison results of different algorithms, as shown in Table 20. Where “\(+/=/-\)” refers to the number of MGEHO wins on the function, ties on the function, and loses on the function. From observing the results in the table, it can be seen that the number of times MGEHO outperforms other metaheuristics is 31, 27, 22, 29, 29, 27, 32, respectively. Most of the data in the table are less than the significance level of 0.05, so it can be determined that MGEHO significantly outperforms one of the other seven algorithms.
Another statistical procedure of Holms was also invoked in this research [52]. Holm’s test is a post hoc testing procedure that is typically used to correct for experimental precision in multiple comparisons. MGEHO was chosen as the control algorithm. p value provides information regarding whether the statistical hypothesis test is significant or not, and it also indicates the significance of the results. A smaller p value indicates stronger evidence against the null hypothesis. All algorithms are ranked according to the p value and compared to \(a/(k-i)\), where the significance level is 0.05, k is the degree of freedom, and i is the algorithm number. The test starts with the most significant p value and rejects the null hypothesis sequentially, and whenever a null hypothesis cannot be rejected, the test stops and considers all remaining hypotheses accepted. As can be seen in Table 19, the p values of MTDE, EESHHO, HGS, EHOI, SMA, RUN, and GBO are smaller than their \(a/(k-i)\) values, which means that the equality hypothesis is rejected and there are significant differences between these five algorithms and the control algorithm.
Table 19
Consequence of Holm’s test (MGEHO is the control algorithm)
Functions
p value
Rank
MGEHO vs.
\(a/(k-i)\)
Significant differences?
Unimodal functions
5.36E\(-\)23
7.8455
MTDE
0.007143
Yes
3.63E\(-\)10
5.7273
EESHHO
0.008333
Yes
9.14E\(-\)08
4.9818
HGS
0.01
Yes
1.20E\(-\)07
4.7000
EHOI
0.0125
Yes
8.01E\(-\)07
4.3727
SMA
0.016667
Yes
8.41E\(-\)05
3.1455
RUN
0.025
Yes
9.69E\(-\)03
2.4909
GBO
0.05
Yes
Simple Multimodal functions
1.25E\(-\)10
7.9455
MTDE
0.007143
Yes
1.91E\(-\)08
5.4364
EHOI
0.008333
Yes
8.79E\(-\)06
4.2227
RUN
0.01
Yes
5.42E\(-\)05
3.5773
GBO
0.0125
Yes
7.56E\(-\)05
3.5818
HGS
0.016667
Yes
8.23E\(-\)05
3.9364
SMA
0.025
Yes
9.60E\(-\)04
3.3500
EESHHO
0.05
Yes
Hybrid functions
9.29E\(-\)24
7.8880
MTDE
0.007143
Yes
3.14E\(-\)07
4.9960
SMA
0.008333
Yes
8.34E\(-\)07
5.0080
HGS
0.01
Yes
4.14E\(-\)06
4.4200
EHOI
0.0125
Yes
2.93E\(-\)05
4.1160
EESHHO
0.016667
Yes
1.53E\(-\)03
3.2000
RUN
0.025
Yes
2.72E\(-\)02
2.7200
GBO
0.05
Yes
Composition functions
3.49E\(-\)07
7.9750
MTDE
0.007143
Yes
3.60E\(-\)05
5.5750
EHOI
0.008333
Yes
1.32E\(-\)04
4.0500
HGS
0.01
Yes
5.34E\(-\)04
4.4250
SMA
0.0125
Yes
2.79E\(-\)03
3.2750
GBO
0.016667
Yes
5.99E\(-\)03
4.4250
EESHHO
0.025
Yes
3.63E-02
3.9250
RUN
0.05
Yes
Table 20
The p values of the Wilcoxon signed rank test obtained on 32 CEC benchmark functions
Function
EHOI
RUN
GBO
HGS
SMA
EESHHO
MTDE
F1
8.15E\(-\)11
3.03E\(-\)06
9.00E\(-\)01
4.08E\(-\)11
6.07E\(-\)11
3.02E\(-\)11
3.02E\(-\)11
F2
3.47E\(-\)10
3.02E\(-\)11
2.27E\(-\)05
6.12E\(-\)10
1.21E\(-\)10
3.02E\(-\)11
3.02E\(-\)11
F3
1.70E\(-\)08
1.03E\(-\)05
4.64E\(-\)05
2.57E\(-\)07
1.30E\(-\)03
1.47E\(-\)07
3.02E\(-\)11
F4
6.28E\(-\)06
1.37E\(-\)07
7.28E\(-\)05
8.12E\(-\)04
1.17E\(-\)04
2.13E\(-\)05
3.02E\(-\)11
F5
1.44E\(-\)04
1.33E\(-\)07
3.50E\(-\)07
3.02E\(-\)11
3.18E\(-\)05
3.02E\(-\)11
1.17E\(-\)09
F6
3.59E\(-\)05
3.26E\(-\)05
5.75E\(-\)02
5.83E\(-\)03
4.04E\(-\)01
5.40E\(-\)01
3.02E\(-\)11
F7
2.06E\(-\)05
2.12E\(-\)08
1.07E\(-\)09
3.02E\(-\)11
5.49E\(-\)11
3.02E\(-\)11
2.92E\(-\)09
F8
8.68E\(-\)03
1.17E\(-\)04
1.87E\(-\)05
1.25E\(-\)05
1.12E\(-\)02
9.07E\(-\)03
3.02E\(-\)11
F9
6.74E\(-\)06
3.63E\(-\)06
7.28E\(-\)05
1.11E\(-\)06
1.68E\(-\)03
5.30E\(-\)04
3.02E\(-\)11
F10
3.02E\(-\)11
8.99E\(-\)11
4.71E\(-\)04
1.99E\(-\)04
1.50E\(-\)04
1.56E\(-\)08
3.02E\(-\)11
F11
8.10E\(-\)10
7.06E\(-\)04
3.51E\(-\)03
2.67E\(-\)09
4.98E\(-\)11
2.37E\(-\)10
3.02E\(-\)11
F12
3.34E\(-\)03
8.30E\(-\)03
7.24E\(-\)04
1.78E\(-\)04
1.33E\(-\)10
3.63E\(-\)01
3.02E\(-\)11
F13
9.83E\(-\)08
6.01E\(-\)08
1.76E\(-\)02
1.11E\(-\)04
6.97E\(-\)03
8.66E\(-\)05
7.39E\(-\)11
F14
8.99E\(-\)11
5.87E\(-\)04
1.91E\(-\)03
9.06E\(-\)08
1.17E\(-\)04
3.52E\(-\)07
3.02E\(-\)11
F15
1.69E\(-\)09
2.06E\(-\)07
3.40E\(-\)08
8.89E\(-\)10
3.20E\(-\)09
2.61E\(-\)10
3.02E\(-\)11
F16
3.30E\(-\)04
2.83E\(-\)04
3.85E\(-\)03
2.46E\(-\)04
4.40E\(-\)04
3.30E\(-\)04
1.83E\(-\)04
F17
1.83E\(-\)04
5.87E\(-\)02
6.23E\(-\)01
3.30E\(-\)04
2.12E\(-\)01
1.83E\(-\)04
1.83E\(-\)04
F18
6.23E\(-\)02
3.45E\(-\)04
5.21E\(-\)01
5.21E\(-\)01
5.83E\(-\)04
9.11E\(-\)03
1.83E\(-\)04
F19
2.83E\(-\)06
7.34E\(-\)04
3.07E\(-\)04
1.21E\(-\)04
7.28E\(-\)03
4.51E\(-\)04
1.83E\(-\)04
F20
4.40E\(-\)04
3.30E\(-\)04
1.13E\(-\)02
2.46E\(-\)04
4.40E\(-\)04
2.46E\(-\)04
1.83E\(-\)04
F21
5.80E\(-\)06
2.41E\(-\)04
5.71E\(-\)04
8.90E\(-\)04
2.57E\(-\)04
6.23E\(-\)04
1.83E\(-\)04
F22
1.83E\(-\)04
2.41E\(-\)01
4.27E\(-\)01
1.83E\(-\)04
2.46E\(-\)04
2.46E\(-\)04
1.83E\(-\)04
F23
5.39E\(-\)05
2.83E\(-\)04
1.04E\(-\)04
1.73E\(-\)02
5.80E\(-\)03
7.34E\(-\)05
1.83E\(-\)04
F24
1.04E\(-\)04
5.71E\(-\)04
6.78E\(-\)02
8.13E\(-\)04
2.70E\(-\)01
4.73E\(-\)05
1.83E\(-\)04
F25
1.40E\(-\)08
9.70E\(-\)06
3.07E\(-\)07
8.50E\(-\)07
6.78E\(-\)07
2.73E\(-\)09
2.46E\(-\)04
F26
2.46E\(-\)04
1.40E\(-\)02
1.04E\(-\)01
4.40E\(-\)04
3.30E\(-\)04
3.76E\(-\)02
1.83E\(-\)04
F27
2.57E\(-\)05
4.25E\(-\)02
1.86E\(-\)01
1.04E\(-\)03
8.40E\(-\)04
5.39E\(-\)02
1.83E\(-\)04
F28
2.73E\(-\)05
2.11E\(-\)02
7.34E\(-\)04
1.86E\(-\)01
9.70E\(-\)05
9.70E\(-\)01
1.83E\(-\)04
F29
3.81E\(-\)07
6.47E\(-\)06
2.06E\(-\)07
2.01E\(-\)06
1.25E\(-\)07
1.68E\(-\)08
2.61E\(-\)10
F30
6.95E\(-\)07
1.08E\(-\)07
3.01E\(-\)04
1.19E\(-\)06
5.09E\(-\)08
3.51E\(-\)08
3.02E\(-\)11
F31
3.02E\(-\)11
9.12E\(-\)09
5.76E\(-\)09
1.17E\(-\)07
3.03E\(-\)10
5.00E\(-\)09
3.02E\(-\)11
F32
4.20E\(-\)10
3.02E\(-\)11
2.06E\(-\)11
4.08E\(-\)11
3.02E\(-\)11
2.61E\(-\)10
3.02E\(-\)11
+/=/\(-\)
31/0/1
27/3/2
22/2/8
29/1/2
29/0/3
27/1/4
32/0/0
4.2.4 Running time analysis
To analyze the differences between MGEHO and other metaheuristics, the running time of the algorithm on 32 functions is presented in Table 21. The results of time in the table for each algorithm are based on 30 independent runs. As can be seen from the table, SMA takes the most computation time. While MTDE is the fastest in runtime, and this advantage is absolute. The proposed MGEHO ranks six among the compared algorithms but is faster than RUN and SMA. The comparison reveals that the embedding of different operators implements MGEHO require more computation time in exchange for performance improvement. But overall, this trade-off is acceptable because their time complexity is in the same order of magnitude.
Table 21
Time comparison between MGEHO and other selected algorithms on 32 benchmark functions
Functions
MGEHO
EHOI
RUN
GBO
HGS
SMA
EESHHO
MTDE
F1
2.32
1.30
2.80
1.63
0.39
3.49
0.75
0.06
F2
2.31
1.05
2.42
1.44
0.27
3.20
0.63
0.04
F3
1.80
0.97
2.21
1.37
0.25
2.94
0.61
0.04
F4
1.73
0.77
1.67
0.95
0.20
2.29
0.47
0.03
F5
1.13
0.52
1.30
0.65
0.15
1.43
0.30
0.02
F6
3.19
1.59
3.32
1.94
0.53
3.97
0.90
0.08
F7
2.90
2.39
3.61
2.07
2.65
3.83
1.33
0.08
F8
2.49
1.13
2.56
1.51
0.33
3.31
0.68
0.04
F9
2.56
1.13
2.67
1.50
0.32
3.36
0.70
0.04
F10
2.21
1.13
2.48
1.46
0.33
3.18
0.64
0.04
F11
2.09
0.92
1.93
1.18
0.28
2.42
0.53
0.04
F12
1.35
0.59
1.29
0.75
0.17
1.62
0.34
0.02
F13
1.74
1.35
1.95
1.09
0.51
2.03
0.70
0.03
F14
1.27
0.57
1.26
0.75
0.16
1.59
0.34
0.02
F15
1.41
0.64
1.27
0.75
0.18
1.67
0.37
0.02
F16
1.33
0.59
1.44
0.94
0.23
1.75
0.37
0.04
F17
1.27
0.53
1.20
0.68
0.15
1.56
0.31
0.03
F18
1.39
0.58
1.20
0.71
0.19
1.61
0.34
0.02
F19
1.13
0.55
1.18
0.73
0.17
1.53
0.34
0.03
F20
1.40
0.60
1.22
0.72
0.18
1.56
0.34
0.02
F21
1.44
0.62
1.30
0.77
0.18
1.66
0.35
0.02
F22
1.59
0.65
1.28
0.76
0.21
1.62
0.41
0.02
F23
1.63
0.71
1.52
0.89
0.19
2.00
0.42
0.02
F24
1.39
0.59
1.24
0.73
0.19
1.58
0.35
0.02
F25
1.53
0.81
1.70
0.87
0.28
1.73
0.46
0.02
F26
1.28
0.59
1.22
0.70
0.17
1.56
0.34
0.02
F27
1.53
1.80
2.19
1.32
0.76
2.13
0.95
0.04
F28
2.02
0.87
1.84
0.85
0.30
1.72
0.50
0.03
F29
2.44
1.03
1.65
0.97
0.37
1.80
0.58
0.04
F30
1.63
0.98
1.89
0.94
0.36
1.84
0.56
0.03
F31
1.80
1.26
1.83
1.12
0.46
2.00
0.68
0.03
F32
1.71
1.98
2.30
1.37
0.81
2.05
0.97
0.04
Avg
1.78
0.96
1.84
1.07
0.37
2.19
0.55
0.03
Rank
6
4
7
5
2
8
3
1
5 Applications of MGEHO on engineering problems
There are many cases in real-world engineering that are essentially optimization problems [40]. In this subsection, MGEHO is used to determine solutions to real engineering problems and compared with other algorithms to check the applicability of the proposed algorithm. Three practical engineering problems are selected, namely, a design problem regarding a gear train, a design problem with a three-bar truss, and a design problem with respect to a pressure vessel. To handle optimization constraints, constraint methods need to be integrated into the optimizer. Hybrid methods, repair algorithms, special operators, and penalty functions are applicable [53]. Among them, penalty functions are the simplest and easiest to implement. The death penalty is the most suitable type of penalty function. It can automatically eliminate feasible solutions by constructing the main objective value of a mathematical model and applying heuristic algorithms [54]. Therefore, the three abovementioned mathematical model problems are treated in this way. The population size N is uniformly set to 30, and the maximum number of iterations \({t_{\max }}\) is 500.
5.1 Design problem of a gear train
Proposed by Sandgren [55], this problem is a discrete unconstrained optimization problem. Four variables \({x_1},{x_2},{x_3},{x_4}\) are involved, which represent the number of teeth in each of the four gears of the train. All variables are positive integers and lie in the interval [12, 60]. The main objective of the problem is to minimize the gear ratio cost, and the mathematical model is expressed as follows:
The comparison results for the gear train design problem
Algorithm
Optimal values of variables
Optimal value
\({x_1}\)
\({x_2}\)
\({x_3}\)
\({x_4}\)
MGEHO
36.83
12.04
12.26
27.76
2.37E\(-\)16
EHO
53
19.89
16.61
35.05
1.13E\(-\)03
EEHO15
47.99
29.08
13.82
58.06
1.64E\(-\)12
OEHO
52.7
15.37
14.83
25.72
5.67E\(-\)04
LFEHO
40.57
12
29.75
60
5.62E\(-\)06
IGWO
56.82
14.54
25.53
45.27
4.54E\(-\)14
EO
55.62
16.71
21.27
44.31
3.09E\(-\)13
HHO
50.04
23.32
14.82
47.88
1.37E\(-\)15
SFO
42.9
17.03
17.03
46.83
6.00E\(-\)13
WOA
51.15
22.45
17.97
55.90
4.83E\(-\)11
MGEHO is applied to the problem and compared with EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA. The results are presented in Table 22. The best solution is in bold. Compared with other algorithms, MGEHO achieves the best solution. The best cost obtained is 2.3694E–16, which is achieved when \({x_1},{x_2},{x_3},{x_4}\) are set to 36.83, 12.04, 12.26, and 27.76, respectively.
5.2 Design problem with a three-bar truss
This problem is a common minimization constraint problem. The objective is to minimize the total weight of three diffeomorphic frames, where the constraints are mainly stress, deflection, and buckling constraints. The structure of the combined diffeomorphic frames is shown in Fig. 6. The mathematical model of this problem is represented as follows:
The comparison results for the three-bar truss design problem
Algorithm
Optimal values of variables
Optimal value
\({x_1}\)
\({x_2}\)
MGEHO
0.78157
0.42853
263.89578
EHO
0.70718
0.70718
270.73939
EEHO15
0.77259
0.45579
264.10141
OEHO
0.77329
0.5203
270.74917
LFEHO
0.73578
0.58431
266.54055
IGWO
0.91039
0.31279
288.77498
EO
0.79006
0.40433
263.89726
HHO
0.81915
0.36956
263.89728
SFO
0.80382
0.36702
264.05596
WOA
0.81425
0.34033
264.33731
After running MGEHO and EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA independently 30 times each, the best results are obtained, as shown in Table 23. MGEHO obtains the best solution to the problem in bold (263.89578) when \({x_1},{x_2}\) are 0.78157 and 0.42853, respectively. The results show that MGEHO can handle the problem effectively and has superior performance to that of other algorithms.
5.3 Design problem with respect to a pressure vessel
This is a common engineering design problem proposed by Kannan and Kramer [56]. Four variables are involved, namely, the thickness of the shell \({T_s}\), the thickness of the head \({T_h}\), the inner radius of the cylinder R, and the length of the cylinder L. The objective of the problem is to minimize the manufacturing cost of the pressure vessel, as shown in Fig. 7. The problem contains four constraints, three of which are linear, and one is nonlinear. The mathematical form of the problem is as follows:
The comparison results for the pressure vessel design problem
Algorithm
Optimal values of variables
Optimal cost
\({T_s}\)
\({T_h}\)
R
L
MGEGO
0.7938
0.3879
40.6303
195.7186
5966.3796
EEHO15
0.8125
0.4375
42.0984
176.6366
6059.7143
OEHO
0.8125
0.4375
42.0968
176.6580
6059.9382
LFEHO
0.8125
0.4375
42.0912
176.7465
6061.0777
IGWO
0.7782
0.3847
40.3211
199.9802
5988.0275
EO
1.1049
0.5461
57.2487
49.5383
6711.1442
HHO
1.0130
0.5007
52.4910
79.7333
6420.2814
SFO
1.1357
0.5548
58.1465
44.4844
6833.9158
WOA
1.1740
0.5619
58.7656
41.1017
7001.7002
MGEHO is applied to optimize the problem and compared with nine other algorithms. They are EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, the HHO, the SFO, and the WOA, respectively. The obtained experimental results are shown in Table 24. The best solution is in bold. According to the data, MGEHO achieves the best solution among those output by the 10 algorithms, with 0.79386, 0.38794, 40.63038, 195.71868 as the four variable values and 5966.37963 as the best cost.
6 Conclusion
In this paper, a new hybrid metaheuristic algorithm is proposed, named MGEHO. MGEHO is adapted to address the weaknesses of the original EHO algorithm, such as its low accuracy and poor convergence rate. The clan updating operator in EHO is optimized by incorporating the somersault foraging strategy from MRFO, and the patriarch position updating method is optimized. On this basis, the ability of MGEHO to jump out of local optima and achieve global optimization is effectively improved. Furthermore, a nonlinear convergence factor, which can vary with the operating state, is set. This strategy contributes to coordinating global searches and achieving a balance between exploration and exploitation. To improve the diversity of the given population, Gaussian mutation is also introduced to EHO to adjust the position of the worst individual in each clan. In addition, MGEHO is tested on 33 classical benchmark functions and compared with EHO, EEHO15, OEHO, LFEHO, the IGWO, the EO, HHO, the SFO, and the WOA. The experiment on 32 modern CEC2014 and CEC2017 benchmark functions is also performed among MGEHO, EHOI, RUN, GBO, HGS, SMA, EESHHO, and MTDE. The results show that MGEHO is more effective and stable than the competing algorithms. Additionally, MGEHO is applied to three specific engineering practice problems. According to the results, MGEHO obtains better solutions than those of the compared algorithms, verifying the ability of MGEHO to solve engineering optimization problems effectively.
In future work, MGEHO will be applied to multilevel threshold image segmentation and will transformed into a binary version of a feature selection method. Moreover, MGEHO is also intended to be combined with additional models and methods in machine learning, such as support vector machines, unsupervised learning, artificial neural networks, and convolutional neural networks.
Declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Singh P, Choudhary SK (2021) Introduction: optimization and metaheuristics algorithms. In: Metaheuristic and evolutionary computation: algorithms and applications, vol 916. Springer, pp 3–33
2.
Khanduja N, Bhushan B (2021) Recent advances and application of metaheuristic algorithms: a survey (2014–2020). Metaheuristic Evolut Comput 916:207–228
3.
Jin Q, Xu Z, Cai W (2021) An improved whale optimization algorithm with random evolution and special reinforcement dual-operation strategy collaboration. Symmetry 13(2):238
4.
Huang Y, Shen XN, You X (2021) A discrete shuffled frog-leaping algorithm based on heuristic information for traveling salesman problem. Appl Soft Comput 102(2):107085
5.
Hussain K, Salleh MNM, Cheng S, Shi Y (2019) Metaheuristic research: a comprehensive survey. Artif Intell Rev 52(4):2191–2233
6.
Rizk-Allah RM (2019) An improved sine-cosine algorithm based on orthogonal parallel information for global optimization. Soft Comput 23(16):7135–7161
7.
Tang C, Zhou Y, Tang Z, Luo Q (2021) Teaching-learning-based pathfinder algorithm for function and engineering optimization problems. Applied Intelligence 51:5040–5066
8.
Holland JH (1992) Genetic algorithms. Sci Am 267(1):66–73
9.
Simon D (2008) Biogeography-based optimization. IEEE Trans Evolut Comput 12(6):702–713
10.
Storn R, Price K (1997) Differential evolution—a simple and efficient heuristic for global optimization over continuous spaces. J Glob Optimiz 11(4):341–359MathSciNetMATH
11.
Bozorg-Haddad O, Janbaz M, Loáiciga HA (2016) Application of the gravity search algorithm to multi-reservoir operation optimization. Adv Water Resour 98:173–185
12.
Faramarzi A, Heidarinejad M, Stephens B, Mirjalili S (2020) Equilibrium optimizer: a novel optimization algorithm. Knowl-Based Syst 191:105190
13.
Karaboga D, Basturk B (2007) A powerful and efficient algorithm for numerical function optimization: artificial bee colony (abc) algorithm. J Glob Optimiz 39(3):459–471MathSciNetMATH
14.
Shadravan S, Naji H, Bardsiri VK (2019) The sailfish optimizer: a novel nature-inspired metaheuristic algorithm for solving constrained engineering optimization problems. Eng Appl Artif Intell 80:20–34
15.
Mirjalili S, Lewis A (2016) The whale optimization algorithm. Adv Eng Softw 95:51–67
16.
Heidari AA, Mirjalili S, Faris H, Aljarah I, Mafarja M, Chen H (2019) Harris hawks optimization: algorithm and applications. Future Gener Comput Syst 97:849–872
17.
Geem ZW, Kim JH, Loganathan GV (2001) A new heuristic optimization algorithm: harmony search. Simulation 76(2):60–68
18.
Rao RV, Savsani VJ, Vakharia D (2011) Teaching-learning-based optimization: a novel method for constrained mechanical design optimization problems. Comput-Aided Design 43(3):303–315
19.
Wang GG, Deb S, Gao XZ, Coelho LDS (2016) A new metaheuristic optimisation algorithm motivated by elephant herding behaviour. Int J Bio-Inspired Comput 8(6):394–409
20.
Li W, Wang GG (2021) Elephant herding optimization using dynamic topology and biogeography-based optimization based on learning for numerical optimization. Eng Comput 1:1–29
21.
Muthusamy H, Ravindran S, Yaacob S, Polat K (2021) An improved elephant herding optimization using sine-cosine mechanism and opposition based learning for global optimization problems. Expert Syst Appl 172:114607
Elhosseini MA, El Sehiemy RA, Rashwan YI, Gao X (2019) On the performance improvement of elephant herding optimization algorithm. Knowl-Based Syst 166:58–70
24.
Balamurugan K, Mahalakshmi R (2020) Anfis—fractional order pid with inspired oppositional optimization based speed controller for brushless dc motor. Int J Wavelets Multiresolut Inf Process 18(01):1941004MATH
25.
Xu H, Cao Q, Fang C, Fu Y, Su J, Wei S, Bykovyy P (2018) Application of elephant herd optimization algorithm based on levy flight strategy in intrusion detection. In: 2018 IEEE 4th International Symposium on Wireless Systems within the International Conferences on Intelligent Data Acquisition and Advanced Computing Systems (IDAACS-SWS), IEEE, pp 16–20
26.
Li W, Wang GG, Alavi AH (2020) Learning-based elephant herding optimization algorithm for solving numerical optimization problems. Knowl-Based Syst 195:105675
27.
Li J, Guo L, Li Y, Liu C (2019) Enhancing elephant herding optimization with novel individual updating strategies for large-scale optimization problems. Mathematics 7(5):395
28.
Zhao W, Zhang Z, Wang L (2020) Manta ray foraging optimization: an effective bio-inspired optimizer for engineering applications. Eng Appl Artif Intell 87:103300
29.
Ekinci S, Izci D, Hekimoğlu B (2021) Optimal fopid speed control of dc motor via opposition-based hybrid manta ray foraging optimization and simulated annealing algorithm. Arab J Sci Eng 46(2):1395–1409
30.
Shaheen AM, Ginidi AR, El-Sehiemy RA, Ghoneim SS (2020) Economic power and heat dispatch in cogeneration energy systems using manta ray foraging optimizer. IEEE Access 8:208281–208295
31.
Ghosh KK, Guha R, Bera SK, Kumar N, Sarkar R (2021) S-shaped versus v-shaped transfer functions for binary manta ray foraging optimization in feature selection problem. Neural Comp Appl 33:11027–11041
32.
Liu B, Wang Z, Feng L, Jermsittiparsert K (2021) Optimal operation of photovoltaic/diesel generator/pumped water reservoir power system using modified manta ray optimization. J Clean Prod 289:125733
33.
Sheng B, Pan T, Luo Y, Jermsittiparsert K (2020) System identification of the pemfcs based on balanced manta-ray foraging optimization algorithm. Energy Rep 6:2887–2896
34.
Alturki FA, Omotoso HO, Al-Shamma’a AA, Farh HM, Alsharabi K (2020) Novel manta rays foraging optimization algorithm based optimal control for grid-connected pv energy system. IEEE Access 8:187276–187290
35.
Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evolut Comput 1(1):67–82
36.
Barshandeh S, Haghzadeh M (2020) A new hybrid chaotic atom search optimization based on tree-seed algorithm and levy flight for solving optimization problems. Eng Comput 5:1–44
37.
Bäck T, Schwefel HP (1993) An overview of evolutionary algorithms for parameter optimization. Evolut Comput 1(1):1–23
38.
Nautiyal B, Prakash R, Vimal V, Liang G, Chen H (2021) Improved salp swarm algorithm with mutation schemes for solving global optimization and engineering problems. Eng Comput 4:1–23
39.
Yu C, Cai Z, Ye X, Wang M, Zhao X, Liang G, Chen H, Li C (2020) Quantum-like mutation-induced dragonfly-inspired optimization approach. Math Comput Simul 178:259–289MathSciNetMATH
40.
Gupta S, Deep K, Mirjalili S (2020) An efficient equilibrium optimizer with mutation strategy for numerical optimization. Appl Soft Comput 96:106542
41.
Nadimi-Shahraki MH, Taghian S, Mirjalili S (2021) An improved grey wolf optimizer for solving engineering problems. Expert Syst Appl 166:113917
42.
García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf Sci 180(10):2044–2064
43.
Derrac J, García S, Molina D, Herrera F (2011) A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evolut Comput 1(1):3–18
44.
García S, Molina D, Lozano M, Herrera F (2009) A study on the use of non-parametric tests for analyzing the evolutionary algorithms’ behaviour: a case study on the cec’2005 special session on real parameter optimization. J Heuristics 15(6):617–644MATH
45.
Ahmadianfar I, Heidari AA, Gandomi AH, Chu X, Chen H (2021) Run beyond the metaphor: an efficient optimization algorithm based on runge kutta method. Expert Syst Appl 181:115079
46.
Ahmadianfar I, Bozorg-Haddad O, Chu X (2020) Gradient-based optimizer: a new metaheuristic optimization algorithm. Inf Sci 540:131–159MathSciNetMATH
47.
Yang Y, Chen H, Heidari AA, Gandomi AH (2021) Hunger games search: visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Syst Appl 177:114864
48.
Li S, Chen H, Wang M, Heidari AA, Mirjalili S (2020) Slime mould algorithm: a new method for stochastic optimization. Future Gener Comput Syst 111:300–323
49.
Li C, Li J, Chen H, Heidari AA (2021) Memetic harris hawks optimization: developments and perspectives on project scheduling and qos-aware web service composition. Expert Syst Appl 171:114529
50.
Nadimi-Shahraki MH, Taghian S, Mirjalili S, Faris H (2020) Mtde: an effective multi-trial vector-based differential evolution algorithm and its applications for engineering design problems. Appl Soft Comput 97:106761
51.
Tu J, Chen H, Liu J, Heidari AA, Zhang X, Wang M, Ruby R, Pham QV (2021) Evolutionary biogeography-based whale optimization methods with communication structure: towards measuring the balance. Knowl-Based Syst 212:106642
52.
Holm S (1979) A simple sequentially rejective multiple test procedure. Scand J Stat 6(2):65–70
53.
Coello CAC (2002) Theoretical and numerical constraint-handling techniques used with evolutionary algorithms: a survey of the state of the art. Comput Methods Appl Mech Eng 191(11–12):1245–1287MathSciNetMATH
54.
Chen C, Wang X, Yu H, Zhao N, Wang M, Chen H (2020) An enhanced comprehensive learning particle swarm optimizer with the elite-based dominance scheme. Complexity 2020(4968063):1–24
55.
Sandgren E (1990) Nonlinear integer and discrete programming in mechanical design optimization. J Mech Des 112(2):223–229
56.
Kannan B, Kramer SN (1994) An augmented Lagrange multiplier based method for mixed integer discrete continuous optimization and its applications to mechanical design. J Mech Des 116(2):405–411