An Improved Whale Optimization Algorithm Based on Aggregation Potential Energy for QoS-Driven Web Service Composition

With more complex user needs, the web service composition (WSC) has become a key research area in the current circumstance. *e swarm intelligence algorithms are proved to solve this problem well. However, no researchers have applied the whale optimization algorithm (WOA) to theWSC problem. In this work, we propose a logarithmic energy whale optimization algorithm (LEWOA) based on aggregation potential energy and logarithmic convergence factor to solve this problem. Firstly, the improved algorithm uses a chaotic strategy to enhance the initial swarm diversity. After that, a logarithmic convergence factor is applied to obtain the nonlinear search step. Furthermore, aggregation potential energy as the spatial evaluation is employed in the swarm intelligence algorithms for the first time. Finally, the aggregation potential energy is used to dynamically adjust the nonlinear weight, which improves the search efficiency and prevents the algorithm from falling into local optimization. *e experimental results of the benchmark functions show that the LEWOA has better optimization ability and convergence speed than other swarm intelligence algorithms. In the second experiment of the WSC optimization, the effectiveness and superiority of the LEWOA are verified.


Introduction
e internet of things (IoT) has become the hot spot of information technology reform. And the research on IoT mainly focuses on privacy protection [1,2], edge computing and data processing [3], web service composition optimization [4], and so on. In this work, we mainly study the advanced strategies of WOA and the application of the improved whale optimization algorithm in web service composition optimization problems.
Presently under the complex requirements of user needs and scene scheduling, it is far from enough only with a single web service module to provide the solution.
erefore, combining web services to solve the problem has become an inevitable choice for different task requirements that specific workflows can represent. e workflow is a combination of various web services, and each web service has quantity choices of subservices. Obviously, WSC is an NP-hard problem. In reference [5], Strunk explained the related issues and applications of service composition. Quality of service (QoS) is currently used as the criterion better to evaluate the pros and cons of web services. QoS aims to evaluate web services quantitatively. Although a large number of web services have similar functions, the QoS with different attributes of each web service is different. So the quantity of WSC tends to explode exponentially in the face of specific problems. Furthermore, the time consumption cannot be ignored if the exhaustive algorithm is selected in a large order of magnitude. e swarm intelligence algorithms can be regarded as an outstanding solution to trade off efficiency and effectiveness. In reference [6], Ouarda proposed that the swarm intelligence algorithm effectively solves such largescale and NP-hard problems.
Here are some typical research and application of various swarm intelligence algorithms in face of WSC problem. In reference [7,8], the authors proposed IDPSO and IDIPSO, respectively, aiming to improve the speed and performance of finding the optimal solution of WSC problem. In reference [9], the author proposes an improved ant colony algorithm; EFACO applies a pheromone-driven scheme composed of QoS multiple weights to improve the efficiency of ant colony search. In reference [10], the genetic algorithm (GA) is used to solve the WSC problem based on elitism and an elite-based learning mechanism.
It should be noted that the traditional swarm intelligence algorithm is oriented to a continuous field, but the WSC is a discrete problem. is article will separately integer coding the candidate service set of the workflow and the multidimensional coordinates of the swarm intelligence algorithm, making them correspond to each other. en, the traditional continuous algorithm has been transformed into a discrete optimization algorithm. Because it only converts the definition of position coordinates without changing the algorithm's optimization principles, the discrete swarm intelligence algorithm can still ensure optimization effectiveness. Nevertheless, the algorithms are more likely to fall into a local optimum in a discrete environment. Aiming to prevent such circumstances and ensure continuous optimization, we must reduce the possibility of search agents gathering. For example, swarm mutation strategy or splitting swarm are to keep the diversity of search agents in the optimization process.
However, not all swarm intelligence algorithms can be applied to WSC. e complexity and the optimization model of an algorithm are the key considerations. WOA is a new swarm intelligence algorithm proposed by Mirjalili and Lewis in reference [11] that is derived from one of the special hunting behaviours of humpback whales called the bubblenet hunting technique [12]. In reference [13], Gharehchopogh made a comprehensive survey of WOA. At present, ant colony system (AG) and particle swarm algorithm (PSO) are widely applied. e optimization principles of WOA are close to PSO in essence. WOA shows its advantages in simpler structure, fewer adjustment parameters, and more excellent global searchability. ese essential factors determine WOA can better apply to optimize this problem with minor changes, which ensures the optimization properties of the algorithm will not be significantly affected.
ere are still some problems with the WOA algorithm. When facing complex multimodal problems, it shows slow convergence speed and low convergence accuracy and easily falls into the local optimal demerits. Currently, three main ways are able to enhance these defects. First, combining with traditional mathematical principles: in reference [14], Chu incorporates WOA with simulated annealing algorithm to improve the searchability of the algorithm. In references [15,16], chaotic strategy has been used to initialize the swarm to enhance the diversity of the initial state. Second, combining with the metaheuristic algorithm: in reference [17], Jadhav integrates GWO into WOA and proposes the WGC algorithm. In reference [18], Trivedi leads into the PSO model and presents PSO-WOA to improve the local searchability of WOA.
ird, combining with ANN: in reference [19], WOA is combined with ANN to improve the accuracy of the image segmentation algorithm. In reference [20], ANN is used to find the optimal weight to hasten the convergence speed of WOA and enhance its capacity of jumping out of the local optima, while in reference [21], it increases the recognition accuracy of the algorithm with SVM using WOA to optimize the parameters.
Although there are a large number of improved WOA algorithms, most of them have ignored the influence of the distance relationship between search agents on searchability. In these areas of improving the convergence speed, helping the algorithm jump out of the local optimum, and enhancing the search accuracy, WOA still needs further research. In this work, we first try to improve WOA's optimization ability through three new optimization strategies, focusing on improving the original algorithm based on the aggregation potential of the search agent. In the simulation experiment, this article uses the public test function set to verify the effectiveness of the optimization strategies. Finally, the improved algorithm is transformed into a discrete algorithm DLEWOA through several methods to optimize the WSC problem, and the superiority of the improved algorithm is verified through the QWS public data set.

Standard WOA Algorithm.
In WOA, the search agent will optimize through three strategies: encircling prey, bubble-net attacking, and searching for prey. Moreover, WOA will use the coefficient vector A (in Section 1.2) and the random probability P (in Section 1.3) to control the strategy of the next generation. In encircling strategy, the search agent will select the current optimal individual as the target direction; in search for prey, a random search agent will be selected as the target direction; while in bubble-net attacking, the search agents approach the target along shrinking encircling and spiral updating methods simultaneously.

Encircling Prey.
Before encircling prey, the search agent will first select the current optimal candidate solution as the target value. is value is assumed to be the optimal value or close to the optimal value that will be updated with evolutionary iterations. e remaining search agents will approach the target search agent. Equation (1) allows any search agent to update its position for encircling prey. Equation (2) represents the distance between the remaining search agents and the target. e equations are as follows: where j indicates the current iteration, A and C are coefficient vectors, and X * is the position vector of the best solution obtained so far and should be updated in each iteration if there is a better solution. e equations of A and C are as follows:

Wireless Communications and Mobile Computing
where r a and r c are two random vectors distributed uniformly within [0,1]. a is the convergence factor that is defined as follows: e convergence factor a controls the value of A. Encircling prey and hunting method are chosen when | A | <1, while the searching method is selected when | A |≥1.

Bubble-Net Attacking
Method. At this phase, the search agent has two simultaneous hunting behaviours: Shrinking encircling and spiraling updating.
Shrinking encircling. Search agents will randomly approach the current best agent Spiraling updating. Search agents will spirally approach the current best agent e mathematical model is as follows: where b is a logarithmic helix shape constant and l is a random number in [−1, 1]. Equation (7) represents the distance of the j-th whale to the current best agent. Equation (8) represents the search agent updating position along a spiral-shaped path. Due to a decreased linearly, the position updating range of the search agent will be decreased so as to improve the search accuracy. WOA sets a random number P in [0, 1] to model this simultaneous behaviour. e equation is as follows:

Search for Prey.
is phase emphasizes guaranteeing the WOA algorithm to perform a global search. In this phase, | A | ≥ 1. A random whale will be chosen from the current group. X rand is the position vector of the random whale. en, the remaining search agents move towards it, reflecting that the whale will follow the whale group as a whole, which can help the WOA algorithm get rid of the local optimum to a certain extent and perform a global search. e mathematical model is as follows:

Improved LEWOA Algorithm
e original WOA performs well in the low-dimensional unimodal optimization field. However, the nonlinear optimization process of the original WOA does not match with linear decreasing search step size, which will weaken the ability of global optimization in the exploration stage and reduce the convergence accuracy in the exploitation stage. For WOA, although it has the probability to escape through the search for prey strategy after falling into the local optimum, the algorithm itself does not have a specific method of avoiding the local optimum.
is paper presents an improved whale optimization algorithm based on logarithmic convergence factor and aggregation potential energy, aiming to optimize the defects from three aspects.

Chaotic Map.
e searchability of the swarm intelligence algorithm is greatly affected by the initial diversity of swarm conditions. In the WOA algorithm, the initial population state is generated randomly, which is unable to guarantee the swarm diversity in the search space. In this sense, a chaotic strategy can solve the problem well for its ergodicity, high randomness, and regularity. Reference [22] shows the effectiveness of combining a chaotic strategy with WOA. In general, the more homogeneous the chaotic sequence is, the more diverse the swarm states will be; hence, we choose the cubic map with better uniformity performance to initialize the swarm. e cubic map expression is as follows: where y 0 cannot be 0 as the initial value of the iteration; otherwise, the chaotic map cannot be established. e steps through the cubic map to initialize the swarm are as follows: (1) First, a d-dimensional vector coordinate y 1d � (y 11 , y 12 , y 13 , . . ., y 1d ) is randomly generated, substituting it as the first search agent position into equation (13) and iterating it to obtain n 1 + n 2 d-dimensional vectors. (2) Due to the limitations of the chaotic model, the coordinate values obtained through iteration are all between [−1, 1]. And it is necessary to map the chaotic sequence into the search space through the map function. e function is as follows: where d represents the space dimension of the solution, ub is the upper limit of the d-th dimension of the limited space, while lb is the lower limit one. y id is the premapping coordinate of the ith search agent in the d dimension obtained according to equation (13), and x i d is the postmapping agent.

Wireless Communications and Mobile Computing
(3) Calculate the fitness value of each coordinate and choose the points with the minimum fitness as the initial position of the search agent.

e Aggregation Potential Energy.
In the process of searching from dispersion to aggregation and finally converging to the optimal point, the state of motion between search agents will affect each other. Hence, their position relationship cannot be simply summarized by the motor pattern of a single search agent. In the field of crowd panic detection [23], the state of the crowd motion can be well illustrated by the increase or decrease of the crowd potential energy [24]. Since there are some similarities in the moving between the swarm and the crowd, this paper introduces aggregation potential energy to describe the aggregation level of the swarm. Aggregation potential energy can well represent the characteristics of individual distribution during the optimization process, so as to better understand the position transformation relationship of search agents in the search space. According to the base definition of potential energy, the larger the aggregation potential energy is, the more scattered the search agents' distribution is. In this state, the algorithm is still in the exploration phase; otherwise, the algorithm is in the exploitation phase or falls into the local optima.
Due to the high sophistication of calculating the distance between each coordinate, this paper defines the aggregation potential energy of swarm as the Euclidean distance between individuals, setting the mean values of coordinates of all search agents in each generation as the population centre. In this way, we replace distance between individuals with Euclidean distance between individuals, and the final population aggregation potential energy is obtained by equation (15). e expression is as follows: where ϕ is the search range correction factor and is generally 1 here, c i is the position of the i-th search agent, and n is the total number of search agents. By comparing the convergence curve with the change of the aggregation potential energy during experiments, it can be shown that the search agents will aggregate to the optimal point with the increase of the iteration. At the same time, the aggregation potential energy will generally decrease and approach 0. For the problem of judging whether the group is conducting the local search or global search, we cannot confirm to a fixed point. In the optimization process, we can only estimate whether the group's motion strategy tends to global or local search, which is the same for aggregation potential energy. e boundary of aggregation potential energy between global and local is an empirical value obtained through a large number of experiments, generally limited to about 10 −2 ∼ 10 −4 . And in the face of different search environments, the optimal boundary value of aggregation potential energy needs to be obtained by testing and adjusting.
With the optimization of the aggregation potential energy, the global searchability of the algorithm can be improved. When the algorithm falls into the local optimum, the search agents converge, and the aggregation potential energy decreases. At this time, the chaotic map disturbance being added to the swarm, the algorithm can get rid of the current state by regenerating half search agents. e steps are as follows: (1) Evaluating whether the algorithm falls into local optimum. Different aggregation potential energy thresholds should be set for different problem models to improve the estimation accuracy. If the aggregation potential energy reaches the set threshold value but the current optimal value does not reach the theoretical optimal value, the algorithm will be judged to fall into the local optimum.
(2) Jumping out of the local optimization through cubic chaotic map. A d-dimensional vector quantity is randomly generated using the cubic chaotic model, iterating it into equation (13) for times. After which particle points are randomly selected and mapped to the search area through equation (14). en the original half of the search agents are substituted with the worst fitness to enter the iteration again. e algorithm can be well restarted under the preferable uniformity of the cubic chaotic model by rescattering search agents and replacing them with poor fitness. In the new search process, the search agents that previously fell into the local optimum will reoptimize driven by the new search agents, thus improving the optimization ability of WOA.

Nonlinear Inertia Weight and Logarithmic Convergence
Factor. In WOA, the step sizes of each generation are controlled by the coefficient A. e convergence factor a controls the change of A. However, as a linear diminishing factor, a leads to the mismatch between the linear diminish of the search step size and the nonlinear convergence of the algorithm. erefore, this paper converts a in logarithmic convergence form, which improves the optimization ability of the algorithm while guaranteeing that the convergence factor a decreases. e optimization process can generally be divided into two stages: global optimization and local optimization. e former generally lasts a short time, and the latter lasts a long time. erefore, the linearly reduced convergence factor cannot match the convergence characteristics of this algorithm. e convergence factor obtained by logarithmic type can quickly reduce the convergence step in the early stage and change the search step in the later stage, so it is appropriate for a to be advanced in this new type. e new expression of a is as follows: Swarm intelligence algorithm requires a large step size in the exploration phase. While in the exploitation phase, it requires a small one. In the overall search process, it is adverse for algorithm optimization if the linear variation of step size in the search process does not conform to the actual nonlinear search process. Reference [25] corroborates the effectiveness of the nonlinear adaptive strategy to improve the optimization ability of swarmed-based algorithms. Hence, this paper introduces nonlinear adaptive inertia weight to improve the step size, particularly, combining the aggregation potential energy based on the mathematical model proposed in reference [25]. In the light of the aggregation potential energy, the search state of each search agent can be judged, controlling the algorithm to distribute different inertia weight strategies, aiming to enhance the rate of convergence and the accuracy of convergence. e improved nonlinear inertia weight equation is as follows: where w(j) i represents the inertia weight value of the i-th search agent of the j-th generation; w 1 and w 2 are the initial minimum and maximum inertia weight values, respectively; and T max is the maximum number of iterations. E(j) is the jth value of the aggregation potential energy. f(j) i is the i-th fitness of the search agents of the j-th generation, and f(i) avg is the average fitness value of all the search agents of the j-th generation. As for f(j) max and f(j) min , they are the maximum fitness value and the minimum one of the swarm of the j-th generation. c is the search range correction factor, which is inversely proportional to ϕ, and their product is 1.
Equation (17) is the inertia weight equation controlled by the aggregation potential energy E. Based on the experiment analysis, we find that when E ≥ 1, the aggregation potential energy can be considered large with the relatively scattered distribution of search agents. e algorithm is basically in the exploration phase. In this case, if the fitness value of an individual is less than the average fitness value, the search agent is close to the current optimal point, and it should be assigned a smaller inertial weight to approach the optimal point; on the contrary, search agents with poor fitness should be assigned larger inertia weights to enlarge the search step size that is able to improve the global search capability. When E < 1, the aggregation potential energy is small, and it is appropriate to search locally with the concentrated distribution of search agents. Each search agent can be distributed with a smaller inertia weight under the protection of jumping-from-local optimal strategy to optimize with high precision around the optimal point, thus ensuring convergence accuracy. e equation of the update of the new position of the search agent is as follows: 2.4. e Flow of the LEWOA. e flowchart of LEWOA is shown in Figure 1, and the detailed steps are described as follows: Step 1: initialize parameters n, n 1 , n 2 , d, and T max , set iteration initial value j as 1, and identify optimization targets and search areas.
Step 2: use equation (13) to generate n 1 + n 2 d-dimensional vectors, map them into the search area by equation (14), calculate their fitness values, and select the n points with the worst fitness values as the initialization search agents.
Step 3: calculate the fitness value of each search agent to update the current optimal fitness search agent as X * .
Step 4: update parameters a, l, P, A, C, E, and w.
Step 5: estimate whether the algorithm is trapped in a local optimum by aggregation potential energy E and the current optimal fitness value. If local optimal is entered, regenerate n/2 points through step 2, replacing the n/2 search agents with the worst current fitness values, and if not, skip this step.
Step 6: update the positions of search agents through equation (18).
Step 7: determine whether the current generation has reached the maximum number of iterations. If not, j � j + 1 and turn to step 3. If reached, exit the loop and output the X * .

QoS-Driven Web Service
Composition. WSC problem generally evaluates service capability through QoS. is work divides QoS attributes into positive attributes and negative attributes, including response time, availability, success ability, reliability, service price, throughput rate, credibility, and so on. e establishment of this model selects the first four kinds of attributes for research. In addition, in order to standardize the attribute value of each QoS, the normalization function of QoS is established as follows:

Wireless Communications and Mobile Computing
where i is the service number, j is the QoS attribute of the j-th service, QoS i,j (S) is the value of the i-th service and the j-th attribute, and QoS j_max and QoS j_min are the maximum and minimum values of the j-th QoS attribute, respectively. QoS ' i,j (S) is the normalized value. WSC is often carried out by the following several common workflow patterns: series, concurrent, select, and circulation, as shown in Table 1. Further to say, all of these four patterns can be transformed into series types. Hence, in this work, we choose a series WSC workflow as an instance to complete the experiment, as shown in Figure 2.

Problem Modeling.
e basic idea of applying the LEWOA algorithm to the WSC problem is as Figure 3 shows: in the WSC problem, m specific candidate services are selected from each abstract service class to form a service composition. erefore, the coordinate dimension of the search agents can be set to m to map the corresponding abstract service class. e coordinate value range of a dimension is mapped to n candidate services available for selection in the corresponding abstract service class. Figure 3 shows the specific mapping process. It should be noted that in practical problems, the number of candidate services in each service class might be different. erefore, a specific search agent represents a specific combination of web services, and its QoS can be calculated through the fitness function. Based on the value of QoS, we can use LEWOA to optimize WSC, and then the appropriate service composition can be obtained.
According to the above selected parameters and workflow patterns, the fitness function can be converted as follows:

f(x) � F(min T(s), min A(s), min S(s), min R(s)), (18)
where T(s) is the response time, A(s) is availability, S(s) is success ability, and R(s) is reliability.

Feasibility Analysis of Improved Strategy.
After the QoSdriven WSC problem is converted into a single objective optimization problem, its essence has been transformed into efficiently selecting the optimal composition from a large    number of permutations. For most swarm intelligence algorithms, the search logic can be regarded as taking randomly from the coding candidate set and feeding back according to the obtained value to optimize the next selected generation, which is a logical and efficient random extraction. According to this extraction principle, we can find two ways to improve the optimization: on the one hand, having as many extraction times as possible in limited generations and, on the other hand, ensuring that the extracted value can better optimize the extraction strategy of the next generation.
Based on the first optimization principle, considering the time and complexity of the WSC problem, it is hard for the swarm intelligence algorithm to guarantee to search the optimal value of the web service composition. erefore, only when the algorithm keeps searching within limited generations will it not cause search waste. First, we need to avoid the algorithm falling into the local optimum. Second, the algorithm needs to ensure the diversity of the search population, so as to obtain the most information in each generation of extraction. For the improved algorithm, first, the cubic chaotic mapping principle is adopted to ensure the diversity of initial search agents; second, the aggregation situation is estimated by the aggregation potential energy of the search agent. As long as the algorithm enters the local search stage, the split mutation strategy is adopted to regenerate a new part of the population to ensure the swarm diversity of the next generation. e second optimization principle needs a swarm intelligence algorithm to efficiently feedback the search information obtained by the previous generation and choose the suitable search strategy. For example, in applying the ant colony algorithm to web service composition, pheromones are efficiently used to transmit the search information of the previous generation. However, in the original WOA algorithm, no such parameter can better feedback information to the next generation. In the face of the solution of most continuity problems, WOA's search logic will not have a great defect. However, it is difficult to have good optimization performance when encountering complex high-dimensional problem models and discrete problems. e new parameter of aggregation potential energy in LEWOA proposed in this paper can be better used as a medium to transmit the search information between each generation. In the aspect of search strategy, the search strategy of the next generation is determined according to the aggregation potential energy of each generation. In an aspect of step size, the improved algorithm changes the linear step size and convergence factor to nonlinear variation, which is to match the actual nonlinear search process.

Encoding Rules for Mapping.
Since the coordinate values of each dimension of the search agent in the LEWOA are continuous, the serial number of the specific candidate service is discrete. erefore, the fuzzy function f d should be used to convert the coordinate values of the search agent into corresponding integer code to form the discrete logarithmic energy whale optimization algorithm (DLEWOA). e fuzzy function f d is as follows: where x t i,d is the coordinate value of the t generation of the search agent i in the d dimension, m d is the number of candidate services of the abstract service class corresponding to this dimension, and z is the integer on [1, m d ]. e random variable Y is the result of a Bernoulli test with a probability of 0.5. e value of the function iff(P, u, v) depends on whether the proposition P is true. If true, it is u; else, it is v.

Parameters Setting and Benchmark Functions.
To evaluate the optimization ability of the LEWOA, several simulation experiments are conducted based on eight benchmark functions (shown in Table 2), whose experiment results are compared to other swarm intelligence algorithms. e experiment parameters are set as follows: the number of the search agents is 30, the average experimental simulation is 30 times, and the maximum number of iterations is 1,500. As for MWOA, the inertia weight is set according to reference [25], whose minimum value w 1 equals 0.01 and the maximum one w 2 equals 0.4. For the reason that the original fitness parameters are weighted by the aggregation potential energy of the population, after a multitude of experiments, the parameter of the inertia weight is w 1 � 0.01, and the maximum weight parameter is w 2 � 0.08. e simulation experiment in this section is conducted in the conditions of Intel Core, CPU i7-7700HQ, 2.80 GHz, 8 GB, and MATLAB 2016a. Table 3 manifests the specific algorithmic parameters. Table 2 is the model of eight benchmark functions. F1, F2, F3, and F4 belong to unimodal function mainly utilized to test the convergence rate and accuracy of the algorithm; F5 and F6 are multimodal functions mainly wielded to test the ability of global search of the algorithm; and F7 and F8 belong to the multimodal function of mixed dimensions, which are applied to test the impact of search dimension on the searchability of the algorithm.

Experimental Results and Analysis.
In this section, a comprehensive analysis of the LEWOA will be presented.
e experiment records the fitness values of the best search agent and calculates the mean value and the standard deviation to evaluate the accuracy and stability of algorithm optimization results. It analyses the number of convergence and depicts the variation curves of the fitness values of the best search agent and convergence generations to evaluate the algorithm's convergence rate. e results are shown in Table 4, in which the black-labeled represents the optimal data.   [11] Convergence factor [a] Linearly decreases from 2 to 0 MWOA [25] Convergence factor [a] Nonlinearly decreases from 2 to 0 LEWOA Convergence factor [a] Nonlinearly decreases from 2 to 0  Table 4, LEWOA has surpassed WOA from the perspectives of each comparison.

Wireless Communications and Mobile Computing
is result shows that LEWOA is obviously superior to WOA in the accuracy, stability, and rate of optimization based on the six benchmark functions. From F1 to F4 and in F6, LEWOA and MWOA all reach the target values, and the former convergence rate is better since the convergence number of iterations of LEWOA is evidently less than those of MWOA, which is more apparent in the unimodal function test environment. In terms of multimodal functions from F5 to F8, except for F8 in which the standard deviation of LEWOA is slightly inferior to those of MWOA, there are no disparities on account of the order of magnitude. For other benchmark functions, the accuracy and stability of LEWOA's convergence are better than those of MWOA, and the convergence rate is also preponderant. In the case of F8, considering that the increase of the search dimensions has the corresponding influence on the effect of optimization, it can be concluded that the test results in the mean value and standard deviation of the optimal fitness of LEWOA are much better than those of WOA, and it also has better accuracy and stability of convergence for MWOA.
To more intuitively reflect the algorithms' convergence in the benchmark function, a convergence curve is drawn due to one experiment that is the closest to the 30 times of the average test results, comparing the characteristics of convergence of WOA, MWOA, and LEWOA. Figure 4 relatively corresponds to functions F1 to F8, in which the bold font is the best result.
From the convergence curve of Figure 4, LEWOA has a qualitative improvement over WOA in search efficiency and local search accuracy. Especially in F3 and F4, it can be seen that compared with the improved algorithms, WOA proves to fall into the local optimum easily when processing complex problems, while the improved algorithms using the nonlinear strategy still have better convergence. For F1, F2, and F6, the WOA appears very weak in the later local search phase due to its linear allocation step size. On the contrary, the convergence rate of the improved algorithms does not decrease at the later stage because the later algorithms can still allocate an appropriate search step. Compared with the MWOA algorithm, LEWOA is better in convergence generations while maintaining astringency because, under the control of aggregation potential energy, LEWOA can more efficiently allocate optimal search strategies for search agents at each generation.
In theory, the optimization ability of the swarm intelligence algorithm increases exponentially with the increase of the swarm population. erefore, in order to demonstrate the optimization ability of LEWOA, Table 5 discusses the influence of the ability of optimization of LEWOA on the population size, in which the black-labeled represents the optimal data. e swarm number in Table 5 takes into account the nonlinear growth optimization ability of the swarm intelligence algorithm. From unimodal test functions f 1 ∼ f 4 , we can find that the change of population has little impact on the excellent optimization ability of LEWOA. In general, small populations have more efficient optimization speed in such a simple search environment. In the multimodal function of f 4 -f 8 , although the large population has relatively good optimization results, careful observation of the data shows that the small population's optimal value and standard deviation have only a decline of one order of magnitude at most. e LEWOA algorithm for the small population, even in a complex environment still maintains good convergence accuracy and stability. Compared with PSO, GSA, ABC, and other algorithms that need a large number of populations to maintain their astringency, the LEWOA algorithm still occupies a dominant position in the field of small population search. Table 6 compares data from LEWOA and the current swarm intelligence algorithms, selecting MFO, PSO, ABC, GSA, and WOA for performance comparison. To ensure the objectivity and accuracy of the experimental data, each algorithm runs 30 times independently, and the test functions are f 1 ∼f 8 . e average value of optimal solutions of six algorithms and the standard deviation of 30 times of independent operation are shown in Table 5, in which the bold font is the best result.
Concluded from the test data of 8 benchmark functions f 1 ∼f 8 in Table 6, the optimal data of LEWOA are not acquired merely in f 7 , which embodies the superiority of optimization of the improved algorithm. For the high-dimensional unimodal function f 1 ∼f 4 , LEWOA is able to converge to the optimal value every time under the 1,500 generations, which other comparison algorithms cannot achieve. e excellent optimization results not only show that LEWOA has higher precision local convergence ability in unimodal problems but also has better algorithm stability. e improvement of the local convergence ability of LEWOA is due to the improved inertia weight and nonlinear convergence factor, which enables the algorithm to allocate appropriate search steps to deal with the complex search environment. In higher dimensional multimodal functions f 5 ∼f 6 , LEWOA excels better in the accuracy of optimization and stability than other current algorithms other than WOA, which indicates WOA itself possesses a good global optimization ability of global optimization. According to the results of f 5 , LEWOA is better in the accuracy and stability of convergence than those of WOA. It proves the progress of the ability of global optimization of the improved algorithm in higher dimensional multimodal problems and the effectiveness of the combination of aggregation potential energy with nonlinear inertia weight strategy to improve the ability of optimization. For mixed low dimensional multimodal functions f 7 and f 8 , the optimal solution and standard deviation of LEWOA are only slightly smaller than that of the ABC algorithm in f 7 . However, from f 7 to f 8 , with the increase of function dimension, it exactly turns out LEWOA performs better than the ABC algorithm both in convergence accuracy and stability. is change shows that with the help of aggregation potential energy, LEWOA can better jump out of local optimal and conduct global optimization, which proves its astringency and dominance in complex problems.   (17) for normalization. In addition to availability, the other three are treated as positive factors, and the weights of the four are set as 0.2, 0.3, 0.2, and 0.3. e parameter settings of the following comparison algorithms mentioned in Table 7 are the same as those in Table 3. In order to adapt to the discrete search process of WSC problem, the coordinates of each algorithm are integer processed by fuzzy function equation (19).

Experimental Simulation of Web
Since, in the WSC problem of this work, the search scope is expanded to two orders of magnitudes, in order to ensure the search efficiency, ϕ and c will be changed to 0.01 and 100, respectively. Besides, the optimization ability of the WSC problem is guaranteed by the global search as mentioned in Section 3.3. erefore, in order to ensure the diversity of the population and enhance the global searchability of the DLEWOA, the aggregation potential energy threshold is set to 1, and 20 independent experiments are conducted under the same conditions. e other experimental parameters are set the same as before.
e experimental results are as follows, among which bold font is the best: To further illustrate DLEWOA, we use the P-value obtained by t-test to test whether the result values of each algorithm belong to the same distribution to prove the uniqueness of the LEWOA algorithm. P-value is a probability of observed samples and more extreme cases on the premise that the original hypothesis is true. We initially assumed that the values obtained by other algorithms belong to the same distribution as those obtained by LEWOA. erefore, the smaller the P-value obtained, the more rejected the original hypothesis. Generally speaking, when the value of P-value is less than 0.001, it can be considered that there is a significant difference. In Table 8, where α is the   significance level and set to 0.05 and then the confidence level is 95%. In this article, we use the t-test function from MATLAB to obtain P value and confidence interval. It can be seen from Table 7 that compared with the other algorithms, the average optimization result and the optimal value of the DLEWOA are much better, and it also has the lowest standard deviation, which shows that the DLEWOA algorithm in the WSC problem guarantees both validity and stability. Obviously, based on the data of Table 8, no matter which algorithm DLEWOA is compared with, the order of magnitude of the p-value obtained is small enough to verify its uniqueness.
At the same time, this paper tries to conduct a comparative experiment on experimental data through a pseudorandom traversal algorithm in the same experimental environment. When the optimization result reaches about 1.50, the pseudorandom algorithm needs to iterate around 9,000 generations, which takes several times as the swarm intelligence algorithm. Although the pseudorandom traversal algorithm can get better optimization results with enough iterations, the swarm intelligence algorithm is obviously a better choice when time cost is considered.

Conclusions
is paper mainly advances WOA into LEWOA combining three strategies aimed at its defects in processing some multimodal functions and studies the application of DLE-WOA in the QoS-driven WSC problem. e DLEWOA is proposed by using integer coding with the fuzzy function, which solves the problems of mismatch between continuity algorithm and discrete problem model. In the first analysis of LEWOA with eight test functions, the improved algorithm demonstrates its strengths in the convergence rate, optimization ability, and convergence accuracy. In the meantime, this paper tests the impact of swarm's quantity on algorithms, showing that the improved algorithm can still ensure a higher convergence rate and search accuracy in the small population. In the second experiment of QoS-driven WSC, this paper tests the DLEWOA through the QWS data set, and the experiment proves the superiority of the improved algorithm in the comprehensive performance of the WSC optimization problem. As a result, the above experiments validate the effectiveness and superiority of the improved algorithm: LEWOA.

Data Availability
e optimization functions used to support the findings of this study are included within the article.

Conflicts of Interest
e authors declare that there are no conflicts of interest regarding the publication of this study.