An Improved Grasshopper Optimizer for Global Tasks

.e grasshopper optimization algorithm (GOA) is a metaheuristic algorithm that mathematically models and simulates the behavior of the grasshopper swarm. Based on its flexible, adaptive search system, the innovative algorithm has an excellent potential to resolve optimization problems. .is paper introduces an enhanced GOA, which overcomes the deficiencies in convergence speed and precision of the initial GOA. .e improved algorithm is named MOLGOA, which combines various optimization strategies. Firstly, a probabilistic mutation mechanism is introduced into the basic GOA, which makes full use of the strong searchability of Cauchymutation and the diversity of genetic mutation..en, the effective factors of grasshopper swarm are strengthened by an orthogonal learningmechanism to improve the convergence speed of the algorithm.Moreover, the application of probability in this paper greatly balances the advantages of each strategy and improves the comprehensive ability of the original GOA. Note that several representative benchmark functions are used to evaluate and validate the proposed MOLGOA. Experimental results demonstrate the superiority ofMOLGOA over other well-knownmethods both on the unconstrained problems and constrained engineering design problems.

Since its introduction, GOA has been applied to tackling many problems. Aljarah et al. [26] introduced GOA to find the variables in the design of the support vector machine (SVM) and to evaluate the feature functions of the ideal subsets.
is technique minimizes the amount of chosen characteristics and is superior to other techniques in classification precision. e optimized GOA was introduced to the electrical field by Barman et al. [27]. A regional hybrid short-term load forecast (STLF) model based on GOA with SVM was suggested to assess the relevant parameters in order to enhance precision. Hamad et al. [28] applied enhanced GOA combined with SVM for automatic detection of EEG epilepsy. e results showed that this method has better classification accuracy. Ibrahim et al. [29] combined SVM with improved GOA for feature selection and applied the method to Iraqi cancer patients and biomedical datasets (California Irvine datasets) to achieve high-performance accuracy in real-world datasets. In order to accelerate its global convergence rate, Arora and Anand. [30] proposed improved GOA by using chaos theory to balance the optimization process of GOA. Saxena et al. [31] used chaotic mapping to enhance GOA and applied it to the parameter estimation of three bar truss layout and FM acoustic synthesis. El-Fergany [] [32] took the modeling of the proton exchange membrane fuel cell reactor as an example and the findings highlighted the feasibility and validity of the GOAbased program in steady-state and dynamic simulation. Fathy [33] used the objective function to maximize the set of slaves and matched GOA with the complete cross-connection (TCT), the Su Do Ku connection, and the GA set up to boost the output capacity of the photovoltaic array. Ewees et al. [34] proposed the modified GOA by utilizing a learning strategy, named opposition-based learning strategy (OBLGOA). Heidari et al. [35] used hybrid random training to improve GOA in terms of avoiding falling into the local optimal solution and balancing the trends of exploration and exploitation and applied this enhanced method to the medical field. Liu et al. [36] suggested a detailed approach mixing linear weighted sum (LWS) with GOA to address the energy management issue, improve exhaustive energy efficiency, and ensure regional coordination and optimization. Luo et al. [37] put forward an enhanced GOA system combining three approaches to achieve a more suitable development-exploration equilibrium. Mafarja et al. [38] chose the enhanced binary GOA to subject it to the selection of the ideal subset of features and that of the wrapper-based classification field, which made the algorithm's output superior. Mirjalili et al. [39] introduced a selection strategy into GOA to make it have the ability to deal with multiobjective issues, which made the algorithm archive a good maintenance in terms of high convergence and adaptability.
arwat et al. [40] integrated an external archive with GOA to preserve the optimal solution; then, this archive was used in multiobjective search space to define the social behavior of GOA and apply it to the field of multiobjective problems. Wu et al. [41] combined comprehensive empirical mode decomposition (EMD) with extreme learning machines and GOA and applied them to STLF in power management. e improvement in the algorithm can find the most appropriate parameters of the extreme learning machine. Wu et al. [42] improved the GOA by developing a natural selection approach, a democratic selection-making mechanism, and a vibrant 1/5-based feedback mechanism. Zakeri and Hokmabadi [43] used statistical measures to complement the iteration process that replaced repetitive features with the most promising ones, and they were applied to the field of feature selection in a final. Zhang et al. [44] suggested a GOA-based parameter-adaptive variational mode decomposition (VMD) technique for analyzing vibration signals of rotating machinery. Zhao et al. [45] merged the technique of decomposition of Beveridge-Nelson (B-N) with the least squares support vector machine (LSSVM) and GOA and applied it to the prediction model of wind power.
is paper introduces two approaches to GOA and proposes an enhanced version called MOLGOA. First, taking into account the benefits and disadvantages of Cauchy mutation and genetic mutation, we incorporated the probabilistic combination mutation system into GOA to enhance the algorithm's worldwide search capability. en we applied the orthogonal learning mechnism (OL) to GOA, which significantly enhanced the algorithm's convergence velocity. In order to check the effectiveness of the enhanced MOLGOA, the algorithm was evaluated on some representative benchmark functions, including single-mode, multimode, and fixed-dimensional models, and the test results were compared with other competitors including mature metaheuristic algorithms and variations of several sophisticated algorithms. Experiments show that the mechanism proposed in this paper effectively alleviates the problems of local optimization and slow convergence of the original GOA and solves the balance the diversification and intensification. e remainder of this paper is structured as follows. e original GOA is presented in Section 2. Section 3 shows the details of the enhanced MOLGOA. e fourth section presents and discusses the experimental outcomes. MOL-GOA is defined for engineering issues in Section 5. Finally, conclusions and future works are summarized.

An Overview of GOA
GOA was proposed by Saremi et al. [25] in 2017, which simulates the population behavior of grasshopper. Grasshoppers are pests whose life cycle consists of eggs, nymphs, and adults that affect crop production and agriculture. GOA simulates grasshopper's behavior in nymph and adult stages. In the nymph stage, grasshopper's primary motion features are jumping and moving in the manner of rolling cylinders (tiny steps, slow motion) and eat crops discovered along with the manner, while in the phase of adulthood, grasshoppers migrate in groups over lengthy distances (quick steps).
GOA simulates grasshopper population behavior according to the following formula: where X i is the position of the ith grasshopper and Si means social interaction. G i denotes the gravity of the i grasshopper and A i denotes the gravity wind advection of the i grasshopper.
Firstly, the S component in equation (1) is calculated according to the following equation: (2) dij represents the distance between grasshopper i and grasshopper j, which is defined as d ij � |x j − x i |, and d ij represents the distance unit vector from grasshopper i to grasshopper j. e population power function s(r) is described as follows: where f reflects the intensity of the appeal and the duration of the attraction.

Complexity
Secondly, the G component in equation (1) is obtained as where g denotes the gravitational constant, and e g is a unit vector pointing to the origin. irdly, the component A in equation (3) is determined by the following equation: where u and e w represent a unit vector of constant drift and wind direction, respectively. en combining S, G, and A into one formula, the expression is changed as follows: where N represents the number of grasshoppers. e coefficient c of each iteration is determined according to the following equation: where c max and c min represent the upper bound and lower bound of c, respectively, l is the current iteration, and L represents the maximum iteration number.

Improved MOLGOA Method
e improved MOLGOA has two effective strategies. Firstly, the orthogonal mechanism is successfully integrated with GOA to accelerate the convergence speed of it. en, for the advantages of Gaussian mutation and nonlinear programming mutation, these two mutation strategies are adopted in the improved MOLGOA.

Orthogonal Learning Mechanism.
e orthogonal learning mechanism has been proved to be effective in lifting the optimization capability of the intelligent algorithms [46][47][48][49][50][51][52][53] According to the data structure of the orthogonal table, it can be seen that its distribution is a "crisscross" table, where "vertical" refers to the horizontal distribution of the factors selected in the experiment, and "horizontal" refers to the number of factors examined in the implementation. e total number of "longitudinal" is the exact amount of experimental schemes of the orthogonal test. As an orthogonal table of row F and column Q, its characteristics are mainly described as in each column; different numbers appear the same number of times and the numbers in any two columns are arranged in a complete and balanced manner. is fully reflects the two advantages of the orthogonal table, namely, "uniform dispersion, neatly comparable". In general, each level of each factor touches each level of the other factors, and that's orthogonal. Based on the orthogonal combination and factor analysis, the highest combination can be readily discovered, which can be proved in the literature [54]. e orthogonal experimental structure is efficient for solving the [55] multi-factor and multi-level experimental issues. By arranging a few tests with the orthogonal then a j is called the basic column; the other columns are called the nonbasic columns. A previous study [55] demonstrated an algorithm for generating the orthogonal table L M (Q F ) that first generates fundamental columns and then generates nonbasic columns. As described in Table 2, Q is the primary column and M � Q J , J satisfies the formula F ≤ (Q J − 1)/(Q − 1).
rough the orthogonal table, we can find the optimal combination between different levels of various factors with fewer test times. By comparing the characteristics of uniform dispersion and the neatness of the orthogonal table, the algorithm is not only implemented on a smaller scale but also accelerates the convergence rate of itself.

Genetic Mutation.
Genetic algorithm [56] is an evolutionary algorithm whose fundamental concept is to imitate the evolutionary law of the biological world's "survival of the Complexity 3 fittest in natural selection". Genetic algorithms encode issue parameters as chromosomes and use iterative selection, crossover, mutation, and other procedures to exchange chromosome data in the population in order to produce chromosomes in line with optimization goals. Nonlinear programming is a new discipline formed in the 1950s that is used under a set of constraints of equality or inequality to study the extreme problem of the n-element real function. e constraint conditions are where, x, b, b eq , lb, and ub are vectors. A and A eq are matrices. c(x) and c eq (x) return functions of vectors. f(x), c(x), andc eq (x) are nonlinear functions. e mutation procedure appears in the process of using a genetic algorithm to fix this issue, whose primary aim is to maintain the diversity of the population.
is mutation procedure chooses a person from a population randomly and then chooses a point in the person to mutate to create a better person. e mutation operation of the j th gene in the ith individual a ij is formulated as follows: where a max and a min are the maximum and minimum of a ij , respectively: In equation (10), r 2 is a random number, g represents the current iteration number, G max is the maximum evolution number, and r is a randomly generated value between 0 and 1.
By referring to the GA-based mutation system for solving nonlinear programming problems, new mutant individuals can be acquired according to a certain variety of probability; the potential individual can be acquired (1) Choose the smallest J fulfilling ( (3) Construct the basic columns as follows: |mod Q; end for; end for; (4) Construct the following nonbasic columns:  4 Complexity according to a certain variety of probability, which enhances the algorithm's search capability and retains population diversity.

Cauchy Mutation.
e Cauchy mutation is used to enhance the search capabilities of algorithms in the context of searching and also to avoid the grasshopper population from falling into local optimum. e cauchy mutation is to produce a set of random variables that are consistent with the Cauchy distribution to change the population's position to increase population searchability. e graph of Cauchy density function [57] is quite similar to that of the Gaussian variation density function. e difference between the two is that Cauchy density function slides slowly on both sides of the Y-axis. Moreover, the Cauchy distribution with this feature means that the Cauchy distribution can produce a large disturbance to avoid falling into local optimum and have a stronger global search ability.
(11) e one-dimensional density function of Cauchy mutation is determined as So when δ � 1, μ � 0, the standard form of the onedimensional Cauchy density function at the center of the origin is obtained as follows: Correspondingly, the standard Cauchy distribution function is illustrated as follows: e advantage of Cauchy mountain is to generate highprobability disturbance to enhance global search capability. e Cauchy mutation is widely used in various optimization algorithms, such as PSO [15], which uses a Cauchy mutation to mutate the optimal particles obtained and achieved the optimal global solution. Another instance is the krill herd optimization algorithm (KHA) [58] that utilizes the Cauchy mutation to achieve better alternatives by optimizing the present situation.

Combined Mutation Strategy.
In this subsection, the technique of variability in probability choice is implemented to enhance GOA's search precision and worldwide search capability. Because of the benefits of Cauchy mutation in improving worldwide searchability and the features of potential individuals produced by genetic algorithm, Cauchy mutation is first performed, and then the iteration genetic mutation is feasible. e steps are given as follows: Step 1: Judge the times of the present iteration. If it is not a multiple of 5, create a one-dimensional row vector that is compatible with equation (11) and perform Cauchy mutation on each individual's present place in the population. At the same moment, determine whether the position is superior to the original position after Cauchy mutation. If so, there will be a Cauchy mutation; otherwise, the original position will be retained. If the current number of iterations is a multiple of 5, then jump to Step 2 directly.
Step 2: Generate a number n randomly (0 < n < 1) and continue to Step 3 to the current genetic mutation limit.
Step 3: Equation (9) carries out genetic mutation on each individual's present place in the population. (Note that the mutated position does not need to be compared with the original position to join the following iteration.) e proposed MOLGOA steps are illustrated in Algorithm 1, and the procedure of MOLGOA is presented in the form of a flowchart given in Figure 1.
In Figure 1, l is the current number of iterations, T is the maximum number of iterations, X new is the population at the random initial position, X newnut is the population by carrying out mutation operation, X newcause is the population after Cauchy mutation, and F best is the best fitness.
In this paper, the time complexity of this algorithm mainly depends on the algorithm (E) and the total number of problems (d). rough analysis, it can be concluded that the

Experimental Results and Discussions
In this section, various experiments on optimization benchmark problems are implemented, and the results are discussed in detail.  Table 3, which can be split into three classifications: F1-F3 are single combined transports, F4-F10 are multimodal transports, F11-F20 are blended functions, and F21-F30 are combined functions. Since they have only one worldwide best and no local optimum (LO), monomodal features can be implemented to benchmark development capacities. Compared with singlemode functions, multimode functions have a large amount of LO and can boost with volume, which can be used to test the algorithm's capacity of exploration and that of jumping out of LO. Furthermore, the multimodal fixed-dimensional transport functions only have one ideal global solution but a large amount of LOs, as well as multimodal transport functions. However, since the former's solution space is tiny, it is necessary to reduce the search agent's adaptive step size. erefore, in order to comprehensively assess the algorithm's efficiency and tackle with various kinds of issues, thirty benchmark features are used to evaluate the algorithms.

Comparative Results on CEC2017
As mentioned above, the solution space is tiny; therefore, the search agent's adaptive step size is low.
us, to investigate the algorithm's efficiency and for solving various (1) Initialize the population GrasshopperPositions i (i � 1, 2, . . . , N) (2) Calculate the fitness of initial Grasshoppers (3) Evaluate the fitness of each grasshopper and copy the best grasshopper's fitness to FoodFitness and copy the best grasshopper's position to FoodPosition. (4) while (t < maximum number of iterations) do (5) Update c1 by the equation (7)  (6) for (each Grasshopper (GrasshopperPositions i )) do (7) Update the position of the Grasshopper by equation (6)  (8) if (nowiter%5! � 0) (9) Generate mut_GrasshopperPositions_C using Cauchy mutation by equation (11)  (10) Evaluate the fitness of V i and copy it to FitnessV (11) if (FitnessV < FoodFitness) (12) Update the best position and copy it to GrasshopperPosition (13) Randomize n (14) else (15) Perform nonlinear mutation using equation (9) to generate mut_GrasshopperPositions_G (16) Perform OL strategy to generate GrasshopperPosition (17) Return FoodPosition and FoodFitness 6 Complexity types of problems, and evaluate the optimization algorithm's performance comprehensively, thirty functions are adopted for the experimental tests. For the fairness of the trial, all tests were carried out by 30 search agents for 1,000 iterations under the same conditions. Furthermore, it should be observed that to mitigate the impacts of randomness, each benchmark feature is performed 30 times separately.

4.2.
e Impact of Orthogonal Learning and Combined Mutation. As can be seen from the above sections of this paper, two strategies, namely, combinatorial mutation and orthogonal mechanism, have been implemented into the initial GOA. We created three distinct GOA algorithms to study the impact of each mechanism and mixture, as shown in Table 4. "M" and "OL" mean "combinatorial mutation" and "orthogonal mechanism," respectively. In Table 4, "1" implies that this operation is integrated into GOA, and "0" implies that there is no corresponding operation. For instance, OLGOA implies that GOA combines the orthogonal mechanism, excluding combinatorial mutations. e efficiency of various enhanced GOA variants was evaluated based on these functions, as shown in Table 4. Because nonparametric statistical tests were usually adopted in optimization algorithms comparison, the Wilcoxon sign rank test [59] and the Friedman test [60] have been conducted. Table 5 shows the statistical outcomes of different enhanced GOA variants. First, to study the statistical significance distinction between pairing algorithms, the nonparametric Wilcoxon sign rank test with a meaning rate of 5 percent was used. It is shown in Table 5, the p values of MOLGOA and other competitors in 30 functions (F1-F30) are given in rows 2 through 5. e symbols of "+," "− ," and " � " show that MOLGOA is higher than, lower than, and equal to other algorithms, respectively. MOLGOA showed statistically significant distinctions in the situation of "+/− / � " compared to these techniques of comparison. MOLGOA is usually better than OLGOA, MGOA, and GOA among the 30 functions, as you can see from the figure, so MOLGOA is the best performing variant. In addition, to further explore the distinctions between different techniques, in order to evaluate the average classification value of all these algorithms, the Friedman test is used in the experiments. Obviously, the combination of the two Complexity 7 approaches has the smallest value according to these sorting outcomes, which verifies that MOLGOA performs better than other combinations in solving these test functions. Lastly, based on the above assessment, MOLGOA is chosen as GOA's best technique of enhancement.

Comparison with Other Well-Established Methods.
In this section, the performance of the MOLGOA will be validated by comparing with the original GOA and some more successful heuristic methods such as gray wolf optimizer (GWO), whale optimization algorithm (WOA) [61], sine cosine algorithm (SCA) [62], differential evolution algorithm (DE), dragonfly algorithm (DA), bat algorithm (BA), moth-flame optimization algorithm (MFO), particle swarm optimization (PSO), Harris hawks optimization (HHO) [23], and GOA [25]. e dimension of each algorithm to be tested is set to 30. Table 6 shows the detailed comparison results. Table 6 shows that MOLGOA considerably increases the initial GOA's performance. Furthermore, in Table 6, all values indicate that MOLGOA has the lowest deviation compared to other techniques. MOLGOA has better stability compared to other techniques and is able to locate the ideal solution in a narrower spectrum. As shown in Table 7, it can be seen from the "+/− /�" line that MOLGOA accomplished the outstanding outcomes.
e findings indicate that MOLGOA has evident benefits and statistical significance in different kinds of functions compared to other rivals. We performed the Friedman test on the median ranking value shown in Table 8 further to investigate the distinctions between MOLGOA and the other rivals. From the outcomes, it can be evidently seen that MOLGOA has reached the    Complexity 9 highest quality of the solutions with the average ranking value of 1.16667. Figure 2 demonstrates the convergence characteristics of MOLGOA and other counterparts on various 30-dimensional benchmark issues. As shown, the MOLGOA has the greatest convergence precision and efficiency when it comes to F3, F14, F18, F22, and F26 issues, while other algorithms fall prematurely into local optimization.
However, they all have comparatively decent convergence precision in F1, F5, and F10. In sum, MOLGOA can achieve better convergence curve and shows a quicker convergence rate and can find the best solution to the issue compared to other rivals. From the above assessment, it can be seen that MOLGOA displays great global search capability and high search precision for low-dimensional tasks as well as high-dimensional tasks, so we can  conclude that MOLGOA is superior to those of algorithms in this domain.

Comparison with the State-of-e-Art Algorithms.
To further evaluate the performance of MOLGOA, this impressive method was compared with other sophisticated algorithms in this chapter, including improved GOA (IGOA) [37], CSSA [63], ESSA [64], CGSCA [65], m_SCA [66], LWOA [67], BWOA [4], CCMWOA [5], CLPSO [68], and GL25 [69]. e detailed illustration of the comparison functions and the parameters involved in the experiment are listed in Table 3. In this experiment, Table 9 gives the mean and standard deviation values achieved by these algorithms, while the Wilcoxon test demonstrates the final outcomes of the comparison of distinct algorithms. Table 10 demonstrates the outcomes of the Friedman test, which can demonstrate the benefits of each algorithm more obviously.
According to the information shown in Table 9, it can be found that MOLGOA is superior to other sophisticated algorithms in most cases. As shown in Table 11, we can find that MOLGOA has strong competitiveness among the selected functions in the Wilcoxon test, and the last row of Table 10 shows that MOLGOA performs better in most problems than other advanced methods. Table 11 lists the final ranking outcomes acquired through the Friedman test by distinct algorithms. e convergence rate of MOLGOA is presented in Figure 3; it can be seen that the convergence speed of the proposed algorithm is the fastest among all the methods.
According to the convergence curves in Figure 3, it can be seen that MOLGOA has the greatest convergence precision in all these functions. In F1, although MOLGOA's convergence rate is slow at an early point, in the later phase, MOLGOA can jump from the location optimum and have high convergence precision. As shown, it is evident that MOLGOA has better convergence precision and quicker convergence rate on F2 ,  F6, F14, F18, F22, and F26 functions.
MOLGOA has improved efficiency in the single modal functions, multimodal functions, and fixed-dimensional multimodal functions compared to other advanced methods. is is primarily because MOLGOA adopts the orthogonal mechanism and the joint mutation mechanism, which significantly enhances the algorithm's global search capability, while the orthogonal mechanism offers efficient variables with the improvement feature. Overall, the benchmark issues can be efficiently solved by MOLGOA.

MOLGOA for the Engineering Benchmarks
MOLGOA is implemented in this chapter to deal with two engineering benchmark issues, namely, pressure vessel (PV) design problem and tension-compression string (TCS) design problem. e flowchart of MOLGOA for constraint optimization is shown in Figure 4.

PV Problem.
is mathematical modeling is aimed at minimizing the total cost of the cylindrical pressure vessel, which is closely related to material, forming, and welding. In this optimization task, the thickness of the shell, the thickness of the head, the inner radius, and the length of the cylindrical section, without considering the head, are the variables to be optimized. e formulation can be described as bellow: where 0 ≤ x 1 ≤ 99, 0 ≤ x 2 ≤ 99, 10 ≤ x 1 ≤ 200, 10 ≤ x 1 ≤ 200. MOLGOA solves this issue and is compared to initial algorithms such as GOA, IGOA, and others. Table 12 shows variable range We adopted MOLGOA to solve the TCS problem and compared the results with other studies in the literature [70][71][72][73][74][75][76][77]. Mathematical approaches play an important role in engineering problems. However, there are still some problems in the mathematical approaches, such as the difficulty in solving the generalized geometric programming, the nondifferentiability of the dual objective function, and the divergence of the solutions of the sequential programming formed by the union terms. e proposed algorithm has the advantages of strong randomness and fewer parameters, which can avoid the above problems. e results recorded in Table 13 indicated that OLGOA found the minimum optimal cost. Compared with other existing methods, the effectiveness of this method is verified.
To conclude, our studies on these two classical engineering design issues demonstrate that the suggested MOLGOA can well optimize the practical issues. e algorithm's efficacy in addressing constraint issues is further confirmed. e reason why MOLGOA is superior to other techniques in constraint issues is that the powerful exploration capability of MOLGOA can efficiently help the GOA in coordinating the inclination for exploration.

Discussion.
e main reason why the proposed MOLGOA can be applied to engineering optimization problems is the solution strategy formed by the mutual promotion of GOA, orthogonal mechanism, and joint mutation mechanism. Firstly, the basic equation of GOA is used to construct the population to maintain the diversity of the population. Secondly, Cauchy mutation has a significant ability to break through new areas in the exploration space, which means that the exploration ability of the population may be more powerful. A genetic algorithm is another mechanism for exploring new regions in solving nonlinear programming. e joint mutation mechanism, combined with these two different exploration methods, can help GOA to achieve the best solution faster and improve the convergence performance of the proposed algorithm. e orthogonal mechanism in the algorithm can reduce the search scale and make GOA find the global optimal value faster. Although the proposed MOLGOA has advantages in solving optimization problems, it may require considerable computational cost in the addition of orthogonal steps in MOLGOA. erefore, although the orthogonal mechanism is very effective in some applications, a new mechanism should be studied in future research to dynamically select the most reasonable Q and F, so that MOLGOA can take into account the computational    cost and get the best solution, to reduce the problem of high computational cost. At this stage, we mainly conducted a large number of experiments, continuously adjusted to select reasonable values for Q and F. However, in the existing experiments and engineering applications, we can see that the introduction of optimization mechanism in MOLGOA can greatly enhance the performance of GOA. We hope this method can be applied to other methods in the optimization field.

Conclusions and Future Works
In the presented study, MOLGOA is developed to alleviate the limits of the original GOA, such as insufficient exploration ability and slow convergence speed. is approach mainly introduces the orthogonal learning mechanism to make MOLGOA converge faster. Moreover, a combinatorial mutation strategy is proposed to realize the complementary advantages of Cauchy mutation and genetic mutation. e numerical results of typical benchmark functions show that the proposed MOLGOA can effectively tackle different types of optimization tasks. For the constrained optimization issues, the simulation results also illustrate that MOLGOA can enhance the calculation results and have practical value when solving two engineering design problems.
Of course, as a relatively new algorithm, there are still many problems worthy of further study because of the tremendous potential of GOA. Firstly, classical metaheuristic algorithms can be integrated into GOA to achieve a better balance between exploration and exploitation and improve the global search ability of GOA. en, how to use MOLGOA to solve multiobjective tasks and dynamic optimization is scheduled as the next work. Besides, considering the time and space complexity of the algorithm when solving relatively complex problems, for further improving computational capacity and efficiency of the algorithm, GPU parallel algorithm and multithread parallel processing technology can be adopted to explore more possibilities. Finally, the exploration of the proposed method to other promising problems will be also interesting such as the parameter tuning for deep learning models [78][79][80], feature selection problems [81][82][83][84], social evolution modelling [85], and video coding optimization [86].

Data Availability
e data involved in this study are all public data, which can be downloaded through public channels.

Conflicts of Interest
e authors declare that there are no conflicts of interest regarding the publication of the article.