Equalized Grey Wolf Optimizer with Refraction Opposite Learning

Grey wolf optimizer (GWO) is a global search algorithm based on grey wolf hunting activity. However, the traditional GWO is prone to fall into local optimum, affecting the performance of the algorithm. Therefore, to solve this problem, an equalized grey wolf optimizer with refraction opposite learning (REGWO) is proposed in this study. In REGWO, the issue about the low swarm population variety of GWO in the late iteration is well overcome by the opposing learning of refraction. In addition, the equilibrium pool strategy reduces the likelihood of wolves going to the local extremum. To investigate the effectiveness of REGWO, it is evaluated on 21 widely used benchmark functions and IEEE CEC 2019 test functions. Experimental results show/ that REGWO performs better than the other competitors on most benchmarks.


Introduction
Complex optimization problems in practice are often discontinuous, nondifferentiable, and nonconvex. Prior to the advent of metaheuristic optimization technology [1], the most widely used optimization methods were the gradient descent algorithm and the Gauss-Newton method [2]. However, the gradientbased optimization approach is prone to obtaining the local extremum interference, resulting in a reduction in optimization precision. On the contrary, metaheuristic optimization algorithms are able to identify optimum or near-optimum solutions within an acceptable time.
GWO is a swarm intelligence stochastic optimization algorithm introduced in 2014, as a result of information interaction between the social level and hunting behavior of grey wolves in nature [9]. GWO is an effective metaheuristic, and it attracted the interest of academics when it was initially introduced. It has been widely used in many fields such as feature selection [11,12], image processing [13,14], path planning [15], weld shop inverse scheduling [16], and so on. In GWO, the search process is guided by the leading wolves in each iteration, which shows great convergence toward leading wolves. e leading wolves sometimes fall into the local extremum, especially in multimodal problems. However, when leading wolves get trapped at local optima, other individuals in the population are also vulnerable to local extremes. is is the cause of the decrease in population diversity. erefore, the standard GWO suffers from the same issues as most swarm intelligence algorithms, such as lack of the population diversity and ease of falling into local optimum [17]. e motivation of this paper is to solve the above problem; in order to overcome these weaknesses, an equalized grey wolf optimizer with refraction opposite learning (REGWO) is proposed in this paper. In REGWO, two search strategies with different features are introduced to generate candidate solutions. Among them, the opposite learning of refraction strategy is inspired by the principle of light refraction in nature.
is strategy is introduced to improve population diversity during the search and expand the scope of solution space. At the same time, the fuzzy theory is used to adjust the parameter so that the refraction solution is more random and the algorithm can find more potential solutions. Moreover, the equilibrium pool strategy is designed to impair the leadership of the leading wolves.
is method can make wolves update their position following nonoptimal solution with a certain probability, so wolves have the ability to jump out of the local extremum even when the optimal solution falls into the local optimum. REGWO can achieve better performance by combining the aforementioned tactics. e remainder of this paper is organized as follows. e related work is discussed in Section 2. Section 3 presents the original GWO algorithm. e proposed REGWO algorithm is introduced in Section 4. In Section 5, the performance of the proposed REGWO is evaluated on different benchmark functions; furthermore, the significance of the results is proved by statistical analysis. Finally, we end the paper with conclusions and future work in Section 6.

Related Work
e metaheuristic optimization algorithms have been widely used to solve optimization problems. ese algorithms are divided into three categories: physics-based algorithms, evolutionary algorithms, and swarm intelligence algorithms. Physics-based algorithms mimic the physical rules in nature in which the individuals communicate around the search space by the physical concepts, such as inertia force, light refraction law, gravitational force, and so on. ere are some popular algorithms in this category, such as atom search optimization (ASO) [18] and Henry gas solubility optimization [19]. Evolutionary algorithms are a kind of iterative optimization algorithms simulating natural evolutionary processes. e best individuals are combined to form a new generation, which is the main advantage of EAs because it promotes population improvement during iteration, such as genetic algorithm (GA) [20] and differential evolution (DE) [21]. Swarm intelligence algorithms are inspired by the collective behavior of swarm organisms, such as bird flocking, animal grazing, and so on. Individuals in a population with cooperation and interaction move collectively to the promising areas in the search space. Some recently proposed swarm-based intelligence algorithms are grey wolf optimizer (GWO) [9], monarch butterfly optimization (MBO) [22], moth search algorithm (MSA) [23], Harris hawks optimization (HHO) [24], colony predation algorithm (CPA) [25], and so on. Swarm intelligence algorithms have been shown to be effective at solving optimization problems, but they may fall into local optimum and loss of diversity. As a result, some scholars have proposed modified variations to tackle the flaws. DEWCO algorithm has improved the initial population through a hyperheuristic to increase its convergence speed [26]. EFSABC algorithm is proposed by a search strategy for group escape and foraging based on Levy flight to exit from local optima [27].
GWO is a kind of swarm intelligence algorithm, which imitates the social level of wolves and the group hunting behavior. It has fewer parameters and is easy to implement. erefore, this algorithm has been widely used to solve different optimization problems, such as multidimensional knapsack problem [28], path planning [29], parameter estimation [30], economic dispatch [31], feature selection [32], large scale unit commitment problem [33], wind speed forecasting [34], and so on.
In recent years, numerous scholars have developed variants of the basic GWO to address the weaknesses of GWO and provide better performance. ree different position update methods are proposed [35], which are weighted average, fitness-based, and fuzzy logic. Further experimental analysis reveals that the GWO improved using the fuzzy logic method has better performance. In order to improve the search ability of grey wolf, a modified algorithm RW-GWO based on a random walk has been imported [36]. A cellular grey wolf optimizer with a topological structure (CGWO) is introduced. In CGWO, each wolf has its own topological neighbors, and interactions among wolves are restricted to their neighbors, which favors exploitation. Furthermore, the information diffusion mechanism by overlap among neighbors can allow maintaining the population diversity for longer, usually contributing to exploration [37]. Grey wolf optimizer with crossover and opposition-based learning (GWO-XOBL) is presented to the jump out local optima [38]. An improved grey wolf optimizer is proposed using the explorative equation and opposition-based learning (OBL) [39]. To get a more stable sense of balance between exploitation and exploration, a new modified GWO called memorybased grey wolf optimizer (mGWO) is introduced [40]. Randomized balanced grey wolf optimizer (RBGWO), which improves the overall efficiency of the search process by establishing a balance between its exploitation and exploration capability incorporating three successive enhancement strategies equipped with a social hierarchy mechanism and random walk with student's t-distributed random numbers [41]. By dividing the search process into three stages and using different population updating strategies at each stage, an improved GWO called multistage grey wolf optimizer (MGWO) is proposed; the MGWO is improved while maintaining a certain convergence speed [42].
Another area of interest for researchers is to combine other evolutionary algorithms or operators to improve the performance of GWO. PSO-GWO algorithm merged with PSO [43], the idea of PSO was introduced into the GWO to update the position information of each individual using the optimal value of the individual and the optimal value of the group, which enhanced the diversity of the population and improved the global search ability. e crossover operator is introduced into GWO to promote population diversity [44]. e purpose of the crossover operator is to enhance information sharing among individuals in the population. At the same time, the search accuracy and convergence speed of the algorithm are improved. Grey wolf optimizer has been hybridized with differential evolution (DE) mutation, and two versions, namely DE-GWO and gDE-GWO, have been proposed to avoid the stagnation of the solution [45]. To improve the performance of the GWO, a new variant of the GWO called a mutation-driven modified grey wolf optimizer and denoted by MDM-GWO is proposed.
e MDM-GWO combines a new update search mechanism, modified control parameter, mutation-driven scheme, and greedy approach of selection in the search procedure of the GWO [46]. SCGWO algorithm combines GWO with an improved spread strategy and a chaotic local search mechanism to accelerate the convergence rate of the evolving agents [47]. GWO variant enhanced with a covariance matrix adaptation evolution strategy, Levy fight mechanism, and orthogonal learning strategy named GWOCMALOL is proposed. GWOCMALOL algorithm uses these strategies to bring more effective exploratory inclinations [48].
According to the various improvement strategies mentioned above, the main aim of the GWO variants is to improve search accuracy and convergence speed. Although the above GWOs can overcome some drawbacks of the original GWO, GWO still faces the problem of poor global exploration ability in the late iteration. erefore, an equalized grey wolf optimizer with refraction opposite learning (REGWO) is presented in this paper.

Grey Wolf Optimizer
GWO is a typical swarm intelligence optimization algorithm. e model of GWO originates from the leading class and hunting behavior of grey wolves. ere is a clear division of labor and cooperation among grey wolf individuals. As shown in Figure 1, the grey wolf population is separated into four levels, namely α, β, δ, and ω wolves. e first layer is the α wolf, and the next layer is called the β wolf. e δ wolf is located in the third layer. e grey wolf in the population is called the ω wolf (search wolf), which is located in the bottom layer. e α, β, and δ wolves are called leading wolves, and their number is set to 1. In GWO, the ω wolf must update its position to obtain the optimal solution, and α, β, and δ wolves represent the optimal value, suboptimal value, and third optimal value, respectively. e hunting process of grey wolves is mainly guided by α, β, and δ wolves, and ω wolves update iteratively according to the position of leading wolves. e formula of the grey wolf around prey can be expressed as follows [41]: (1) Here, t is the current number of iterations; ∘ is the Hadamard product operation; X p and X denote the position vector of the prey and a grey wolf, respectively; and the calculation formulas of random vectors A and C are as follows: where r 1 and r 2 are random vectors between [0, 1] and the vector a is linearly decreased from 2 to 0 over the course of iterations.
To better understand the optimization rules of GWO, the possible areas are shown in Figure 2 when the position of the grey wolf is updated. It can be seen from Figure 2 that ω wolf can reach different positions around the prey by adjusting the values of parameters A and C. Furthermore, random variables r 1 and r 2 can assist search wolves to reach any of the points depicted in Figure 2. Parameters A and C are responsible for exploration and exploitation behavior in GWO. A is a random value between [− a, a]. When A > 1 and C > 1, the population is inclined to exploration. In addition, when A < 1 and C < 1, the population is prone to exploitation. e formulas of grey wolf tracking target prey are as follows [41]: Here, D α , D β , and D δ denote the distance between α, β, and δ wolves and other individuals, respectively; X α , X β , and X δ represent the positions of α, β, and δ, severally, respectively; C 1 , C 2 , and C 3 are random vectors; X 1 , X 2 , and X 3 represent the step length and direction of ω wolf toward α, β, and δ, respectively; and ω wolf determines the final position according to equation (3).

Proposed Algorithm REGWO
To improve the global search ability of GWO, the opposite learning of refraction and equilibrium pool technique are introduced in this work. Opposite learning of refraction is chosen in the proposed algorithm to generate more potential solutions. Besides, the equilibrium pool strategy can achieve a better exploration by weakening the leadership of the leading wolf. e two strategies are introduced in Sections 4.1 and 4.2, respectively. e proposed algorithm is named REGWO and described in Section 4.3.

Opposite Learning of Refraction.
e opposition-based learning (OBL) technique is proposed in 2005 [49]. e fundamental idea is to expand the search space of the population by calculating the opposite solution of the current solution, so as to select the candidate solution that is more suitable for the optimization problem. Applying this method to the optimization algorithm can effectively improve the search accuracy of the algorithm [50]. However, the standard OBL has certain shortcomings. OBL only speeds up the convergence of GWO and obtains only one opposite solution of fixed position. erefore, the opposite solution may fall near the local optimal solution, causing the algorithm to fall into the local optimum with the iteration [51]. In order to tackle this problem, this paper introduces the refraction principle to improve the traditional opposite learning process. Refracted opposition-based learning (ROBL) strategy is based on the OBL, combined with the principle of light refraction to identify a better solution. ROBL strategy not only considers the opposite direction of individuals but also considers other directions of individuals. e schematic is shown in Figure 3.
In the one-dimensional space where the individual of the population is located, the x-axis is separated into the upper and lower parts. Above the x-axis is the natural vacuum part of the refraction model, and below the x-axis is the other propagation medium of the refraction model. In Figure 3, the search range of individuals on the x-axis is [a, b], that is, x∈ [a, b], and the y-axis is normal. x′ is the incident point of the light source, and the length of the incident light segment x′o is denoted by h. e incident light refracts at the intersection o, and x′ * is the refraction point; the length of refraction light segment ox′ * is denoted by h * ; and θ and φ are the incidence angle and the refraction angle, respectively. e x coordinate (position) of intersection o is (a + b)/2, which is the midpoint of the individual search range [a, b]. From the geometric relationship in Figure 3, we can obtain e refractive index (n � sin θ/sin φ) of light obtained by equations (4) and (5) is  Computational Intelligence and Neuroscience Let k � h/h * ; equation (6) can be transformed into According to equation (7), we can obtain When n and k are both 1, equation (8) is the standard opposite learning formula: Obviously, the OBL strategy is a special case of the ROBL strategy. In order to improve the ability of GWO to jump out of the local extremum, the above ROBL model is applied to GWO. Since the individuals in GWO are multidimensional, equation (8) can be extended to the D-dimensional space as follows: Here, x i,j represents the value of the j-th dimension of the i-th individual; x * i,j is the opposite solution obtained by the ROBL model; and a j and b j are the j-th dimension upper bound and lower bound, respectively.
As shown in Figure 3 and above, as the k value changes, the position of the opposite solution generated by (10) will be changed. at is to say, the adjustment of k improves the randomness of the solution. e k is calculated by where k max � 1, k min � 0, t denotes the current iteration number, and T is the total number of iterations. Meanwhile, in order to make the decline rate of k well match the convergence rate of the fitness value, this paper proposes a method to adjust the k value as shown in Figure 4. e fuzzy membership degree μ k (t) of k is where μ k (t)∈[0,1], μ k (t) increases with the increase of k(t), μ k (t) decrease with the decrease of k(t), and the optimal fitness relative change rate η is Here, f(t)is the objective function, namely the optimal fitness value of the t-th iteration of the population; f(t − 10) represents the optimal fitness value of the (t − 10)-th iteration of the population; then η denotes the relative change rate of the optimal fitness value in 10 iterations of evolution; and when η value is large, it indicates that the change rate of the optimal fitness value is large, and at this time, the k value should be larger to improve global search capability; otherwise, when η value is small, it indicates that the change rate of the optimal fitness value is small, and at this time, the k value should be smaller. In the ROBL model, firstly, the current η and μ k (t) are judged. en, the k of the next iteration is adaptively adjusted by fuzzy rules, which can accelerate the convergence speed of the algorithm.
Fuzzy rules to adjust (k): Here, c is the threshold, and its value is 0.05; τ 1 and τ 2 are the parameters between [0, 1] in the rules.
On further observation of Figure 3, the purpose of k is to adjust the population, improve population diversity, expand the search space, and improve the global search ability of the algorithm. erefore, the adjustment of parameter k based on fuzzy rules in the ROBL model can effectively improve the diversity of individual distribution in the search space. It makes up for the weak exploration ability of GWO in the late iteration. e advantage of the ROBL model over the OBL model is that the candidate solution can be obtained dynamically by parameter adjustment, which enhances the chance of the algorithm jumping out of the local optimum to a great extent. However, OBL can only obtain a fixed candidate solution. at is, parameter k has the ability to extend search space.

Equilibrium Pool Strategy.
In GWO, the search is primarily guided by α, β, and δ wolves. If the leading wolves fall into the local optimum, the entire population will update their position in the direction of the local optimum. To address this issue, this paper introduces the equilibrium pool strategy to enhance population diversity [52]. e fundamental idea of the strategy is to calculate the fitness value of each individual after population initialization and choose three candidate solutions (X 1 , X 2 , and X 3 ) based on the fitness value. Among them, X 1 , X 2 , and X 3 represent α, β, and δ wolves, respectively. In addition, the average value of three candidate solutions is calculated as the average candidate solution X avg , and then the equalization pool X_ pool is constructed. e mathematical model is as follows: Computational Intelligence and Neuroscience Here, three candidate solutions (X 1 , X 2 , X 3 ) are contributed to exploitation and the average candidate solution (X avg ) is contributed to exploration. ω wolf randomly selects candidate individuals from the candidate pool with equal probability for location updating. Besides, parameter F is used to balance exploration and exploitation, and the mathematical description is as follows: where λ is the random vector between [0,1], r denotes a random number between [0,1], sign(r-0.5) is used to control the direction of exploration and exploitation, t represents the current number of iterations, and T notes the maximum number of iterations. In addition, the generation rate G is used to improve exploitation capability. e mathematical expression is as follows: where X eq is a candidate solution randomly chosen from the equilibrium pool with equal probability.
In summary, when GWO applies equilibrium pool strategy, the position update formula of the individual is as follows:

REGWO Algorithm.
It is well known that exploration and exploitation are necessary for population-based optimization algorithms, such as PSO, ABC, ACO, and so on. In standard GWO, the issue is that since all of the other individuals are attracted toward leader wolves, they may converge prematurely without enough exploration of search space, that is, standard GWO is prone to premature convergence.
To improve the performance of GWO, each grey wolf obtains the opposite solution via the ROBL strategy, which enhances individual randomness. Refraction opposite learning strategy makes up for the shortcomings of traditional opposite learning, expands the search space, and effectively enhances population diversity. In addition, an equilibrium pool strategy is introduced to reduce the likelihood of the algorithm falling into the local extremum. e equilibrium pool retains four individuals, namely α, β, and δ wolves as well as their mean values. e ability of exploration is properly improved by randomly selecting an individual from the equalization pool to lead the position update of ω wolf. e process of REGWO is described in Algorithm 1, and the flowchart of the proposed REGWO algorithm is shown in Figure 5. In REGWO, the α, β, and δ wolves are chosen by population initialization and fitness calculations. en, the position is updated with equal probability by equation (3) or (17). Finally, the refraction opposite solution and its fitness value are calculated by equation (10) Select the former three-excellent solution as α, ß, and d wolves t � t+1 end while return α ALGORITHM 1: Equalized grey wolf optimizer with refraction opposite learning (REGWO). 6 Computational Intelligence and Neuroscience solution is better than the original solution; otherwise, the original solution is retained. Until the end of the iteration, α wolf is the ultimate optimization result.

Experimental Settings.
To fairly compare the performance of different algorithms, the function test set is needed. e numerical efficiency of REGWO developed in this paper was tested by solving 31 mathematical optimization problems. e first 21 benchmark functions are the classical functions utilized in literature [9,53]; they (f 1 ∼f 21 ) are composed of unimodal, multimodal, fixed-dimensional multipeak, and shifted functions. e specific expressions and search intervals of these functions are shown in Table 1.
e unimodal function (f 1 ∼f 9 ) with just one local and also global optimal solution is commonly used to evaluate the local exploitation ability of the algorithm; f 10 ∼f 14 are multimodal function and often used to test the ability of the algorithm to explore. e f 15 and f 16 are fixed-dimensional multipeak functions with many extreme points but low dimensions, so it is easy to optimize and can be used to assess the stability of the algorithm. e last 5 functions in Table 1 are shifted functions, which are mainly to avoid the situation that some algorithms copy one parameter to another to generate a neighbor solution [53]. e other 10 test problems (f 22 ∼f 31 ) considered in this paper (see Table 2) regard composite benchmark functions considered in the IEEE CEC 2019 special session [54]. ese benchmark functions are more complex than the first 21 benchmark functions, and f 22 ∼f 31 are designed to have a minimum value of 1. e optimization performance of each algorithm can be further verified by solving complex problems.
Two sets of experiments are conducted in this paper. In the first experiment, REGWO is compared with some popular algorithms and novel algorithms to evaluate convergence speed and optimization accuracy. Furthermore, the comparisons of GWO, RGWOL (GWO improved only by refraction principle with linear control parameter k), RGWOF (GWO improved only by refraction principle with fuzzy control parameter k), EGWO (GWO improved only by equilibrium pool strategy), and REGWO are executed. e influences of two strategies and dynamic changes of parameter k on the optimization results can be observed through the experiment. In the second experiment, the comparisons of mean and standard deviation are performed between REGWO and the other six different GWO variants. Meanwhile, the population size is set to 30; the maximum number of iterations is set to 5000, and experimental results are based on 30 independent experiments.

Comparison with Swarm Intelligence Algorithms and Strategies Analysis.
To validate the performance of the proposed REGWO algorithm in this paper, it is used to solve the functions f 1 ∼f 31 , and the performance is compared with RGWOL, RGWOF, EGWO, and other swarm intelligence algorithms, including standard GWO [9], sparrow search algorithm (SSA) [55], Archimedes optimization algorithm (AOA) [56], particle swarm optimization (PSO) [57], firefly algorithm (FA) [58], and artificial bee colony (ABC) [4]. Among them, SSA and AOA are novel intelligent algorithms, and PSO, FA, and ABC are popular intelligent algorithms. For better comparison, the other parameters of algorithms are shown in Table 3.
e results on each benchmark function of the algorithms are shown in Table 4. It can be seen from Table 4 that the effectiveness of each strategy has been verified. RGWOL outperforms six test functions. RGWOF outperforms nine test functions. EGWO outperforms six test functions. Among them, the results of the RGWOF algorithm on most functions are better than RGWOL. e experimental results in Table 4 show that the effect of fuzzy control parameter k is better than the linear decrease of parameter k. erefore, the REGWO algorithm proposed in this paper combines the equilibrium pool strategy (EGWO) with the refraction opposite learning strategy (RGWOF). At the same time, the fuzzy theory control parameter k is used in the refraction opposite learning. Although the RGWOF and EGWO can find the solutions, combining the two strategies has more benefits since the solutions obtained by REGWO are always better than theirs.
On the other hand, it can be seen from Update the positions of wolves by Equation 17 Termination criteria satisfied ?

Calculate refraction solutionby Equation 10
Preserve Computational Intelligence and Neuroscience global optimal solution in solving functions f 1 ∼f 3 , f 7 , f 8 , and f 11 ∼f 19 , indicating that REGWO has the potential to converge to the global optimal value. On further observation in Table 4, the standard GWO is still competitive compared with SSA and AOA for unimodal functions. It is proved that the standard GWO has good exploitation ability when solving unimodal functions. However, the optimization performance of standard GWO is relatively weak for multimodal functions, while REGWO shows better performance both in unimodal functions and multimodal functions. Moreover, it can be seen from the Friedman test average rank in Table 4 that the order from low to high is REGWO, RGWOF, EGWO, RGWOL, SSA, FA, GWO, AOA, PSO, and ABC. Obviously, REGWO has preferable competitiveness compared with novel algorithms (SSA, AOA) and the classical algorithms (PSO, FA, ABC).
e superior performances of REGWO should be attributed to the improved strategies. Individuals maintain high diversity during optimization due to refraction opposite solution strategy, and the equalization pool strategy weakens the leadership of   [58] Light absorption coefficient ζ � 1, step size s � 0.2

ABC [4]
Control parameter limit � 0.6 * population size * dim SSA [55] Discoverers n � 0.2 * population size AOA [56] Constants Control parameter a decrease linearly from 2 to 0 No. Functions Computational Intelligence and Neuroscience      the optimal solution. erefore, the combination of the two strategies can effectively improve the performance of standard GWO in solving multimodal functions. at is to say, the ability of the algorithm to jump out of the local extremum is enhanced. Due to the stochastic nature of these algorithms, the statistical test is necessary for providing confidential comparisons [44]. erefore, the Wilcoxon sign rank test is conducted in this paper. e test results of the REGWO algorithm and the other 9 selected algorithms on 31 test functions are shown in Table 5. e sign + (-) denotes that the REGWO algorithm is better (worse) than its compared algorithms. e symbol � indicates that the REGWO algorithm gets the same results as its competitors. It can be seen from Table 5 that REGWO provides higher R + values than R − values in all cases. Moreover, the p values of 9 algorithms are less than 0.05, indicating that they are significantly different from REGWO, and REGWO is superior to other algorithms.
ere are some reasons for REGWO has such a good performance. Firstly, the refraction principle and fuzzy control parameter are introduced in GWO; individual diversity is enhanced. RGWO improves the optimization precision and convergence speed of the standard GWO algorithm by retaining better (original solution and refraction inverse solution). Secondly, the equilibrium pool strategy enhances the global search ability of the original GWO algorithm by reducing the leadership of the leading wolves. e advantage of this strategy is particularly obvious when solving the multimodal function. en, REGWO, as an improved GWO algorithm, combines the advantages of the two strategies. It not only improves the convergence speed of the GWO but also improves the optimization precision of the original GWO algorithm. e convergence histories of the compared algorithms are shown in Figure 6. rough the convergence histories in Figure 6, we can find that the convergence speed of REGWO is faster than other swarm intelligence optimization algorithms on unimodal functions except f 5 . Although the convergence speed of the REGWO algorithm on multimodal function is not as fast as that on unimodal function, REGWO has high search precision than other algorithms. Especially, the optimization performance of REGWO is remarkable when solving more complex functions (IEEE CEC 2019 test suite). It demonstrates that REGWO algorithm not only improves the convergence speed of the standard GWO algorithm on unimodal function but also enhances the optimization precision on complex functions.

Comparison with GWOs.
To further validate the effectiveness of the REGWO. e performances of different GWO variants are compared, and the benchmark functions (f 1 ∼f 31 ) are solved by REGWO, WGWO [35], DGWO [59], AGWO [60], IGWO [61], RLGWO [62], and GNHGWO [63]. To make a fair comparison, the 6 algorithms use the same parameter settings as their original literature. en, the results are analyzed by Friedman test average rank and Wilcoxon signed-rank test, and statistical results (mean cost and standard deviation) for 30 independent experiments are reported in Table 6.
It can be seen from Table 6 that the average fitness of the REGWO is superior except for the functions f 5 , f 10 , and f 20 . In addition, the corresponding standard deviation is much smaller than other algorithms for most functions. e average fitness of the REGWO outperforms the other 6 enhanced GWOs on 14 benchmark functions. It can be seen that the combination of the refraction opposite learning approach and equilibrium pool strategy effectively improves the optimization accuracy of the standard GWO. e above algorithms can achieve the theoretical optimal value for the function f 16 . Because f 16 is a fixed-dimensional multimodal function with low dimension, it is simple to solve. However, REGWO shows better stability in terms of standard deviation.
e f 22 ∼f 31 are more complicated than the test functions listed in Table 1. It can better test the algorithm exploration and exploitation ability. Especially, REGWO converges to the theoretical optimal value 1 on function f 22 . For functions f 26 and f 30 , the iterative optimization results of REGWO are also close to the theoretical optimal value 1. It demonstrates that REGWO still has the ability to converge to the global optimum for more complex mathematical optimization problems. Moreover, REGWO also achieves better performance in most functions in terms of standard deviation, indicating that REGWO has a better stability. From the Friedman test average rank in Table 6, the order from low to high is REGWO, IGWO, DGWO, RLGWO, AGWO, WGWO, and GNHGWO (IGWO is equal to DGWO). It shows that the performance of REGWO is much superior to other GWOs in accuracy. From the results of the Wilcoxon sign rank test in Table 7, REGWO provides higher R + values than R − values in all cases. Moreover, it can be seen that the p values of WGWO, DGWO, AGWO, IGWO, RLGWO, and GNHGWO are less than 0.05, indicating that they are significantly different from REGWO, and REGWO is far superior to the other six algorithms.
In summary, REGWO has such good performance because of the contribution of the two strategies. Firstly, the diversity of solutions is increased by refractive opposite learning. Secondly, the equilibrium pool strategy weakens the leadership of the leading wolves to increase the probability of individuals jumping out of the local optimum. erefore, the REGWO algorithm combining two strategies    The best fitness so far optimal curve (c) Figure 6: Convergence diagrams.  is able to achieve competitive optimization results on the test problems compared with other GWOs.

Conclusion and Future Work
In order to further improve the optimization performance of GWO, this paper proposes an equalized grey wolf optimizer with refraction opposite learning (REGWO). e main idea of the algorithm is to improve the opposite learning process of OBL based on the refraction principle of light. is strategy further expands the search space of the population, increases population variety, and enhances the ability of individuals to jump out of the local extremum. At the same time, the equilibrium pool strategy is combined to weaken the leadership of the leading wolves, which effectively avoids the situation that the rest of the individuals move to the leading wolves when the leading wolves fall into the local optimum. erefore, the combination of the two strategies effectively enhances the exploration of GWO in the late iteration. In addition, REGWO is tested on 31 benchmark functions. e experimental results show that REGWO has higher convergence speed, search accuracy, and stability compared with standard GWO, other state-of-art GWOs, and other swarm intelligence algorithms. On the whole, REGWO is more effective in solving complex optimization problem.
In our future work, the selection of search strategies still needs to be further investigated. Furthermore, the REGWO algorithm can be extended to solve multiobjective optimization, binary optimization, and application-designed problems in the future.

Data Availability
e data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest
e authors declare that they have no conflicts of interest regarding the publication of this paper.