Prediction of Mechanical Properties of Aluminium Alloy Strip Using the Extreme Learning Machine Model Optimized by the Gray Wolf Algorithm

. Mechanical properties are important indicators for evaluating the quality of strips. Tis paper proposes a mechanical performance prediction model based on the Gray Wolf Optimization (GWO) algorithm and the Extreme Learning Machine (ELM) algorithm. In the modeling process, GWO is used to determine the optimal weights and deviations of ELM and experiments are used to determine the model’s key parameters. Te model efectively avoids manual intervention and signifcantly improves aluminum alloy strips’ mechanical property prediction accuracy. Tis paper uses processed data from the aluminum alloy production plant of Shandong Nanshan Aluminum Co., Ltd. as experimental data. When the prediction deviation is controlled within ± 10%, the GWO-ELM model can achieve a correct rate of 100% for tensile strength, 97.5% for yield strength, and 77.5% for elongation on the test set. Te RMSE of the tensile strength, yield strength, and elongation of the GWO-ELM model was 5.365, 11.881, and 1.268, respectively. Te experimental results show that the GWO-ELM model has higher accuracy and stability in predicting aluminum alloy strips’ tensile strength, yield strength, and elongation. Te GWO-ELM model efectively avoids the defects of the traditional model. It has a special guiding signifcance for producing aluminum alloy strips.


Introduction
Among the many indicators of hot-rolled alloy products, mechanical properties, which are infuenced by the complexity of processing parameters, are one of the important indicators of the quality of hot-rolled alloy products. Terefore, the mechanical properties of alloy products are valued by many manufacturers and researchers. Accurate prediction of the mechanical properties of hot-rolled aluminium alloys has become one of the hot topics in the development and application of aluminium alloys [1,2]. Up to now, there are two main models in the feld of metal property prediction. One of them is the construction of material constitutive models for material property prediction [3]. Jia et al. [4] used experimentally obtained stressstrain data to calculate the material constants involved in the Arrhenius-type constitutive model and the modifed Zerilli-Armstrong (MZA) model. Zhang et al. [5] established the constitutive model for the thermal deformation behavior of high-strength aluminium alloy (Al-Zn-Mg-Cu). Te results show that the constitutive model can accurately predict the fow behavior of the Al-Zn-Mg-Cu alloy. Although the emergence of the constitutive model enables researchers to reasonably predict the product performance according to the processing technology, the construction of the constitutive model still needs experimental support and reasonable assumptions, which leads to the fact that the constitutive model cannot fully meet the requirements of industrialization. To meet the needs of industrial production, another model emerged along with the maturity of computer technology, which uses intelligent algorithms to analyze large amounts of real production data to obtain more desirable predictions of mechanical properties. Along with the development of computer technology, artifcial intelligence algorithms play an increasingly important role in materials' research. Various machine learning algorithms applied to materials' research have solved many challenging problems [6]. Te combination of computer technology and materials' science will likely yield more important achievements in the foreseeable future [7]. Guo et al. [8] proposed a method for multiproperty prediction of materials using the Interior Point Algorithm. Te rationality and reliability of the above theory are verifed successfully by the experiment using the data of the steel production process. Among the many machine learning algorithms on which statistical models rely, neural network-like algorithms are favored and widely used in the feld of material property prediction due to their excellent properties [9,10]. Niklas et al. [11] needed to develop accurate models to ensure efcient control equipment for superelastic shape memory alloys (SMA). After training, the ANN can successfully calculate model parameters from cyclic tensile stress-strain tests, thereby improving the efciency of SMA control devices. Lan et al. [12] experimentally established a dataset of the chemical composition, processing parameters, and mechanical properties of the A380 alloy. Te Fe/Mn content ratio, Sr content, cooling rate, and porosity content were the input variables, and ultimate tensile strength (UTS) and elongation (El) were used as output parameters. Two ANN models were developed, namely, (1) a back propagation artifcial neural network (BP-ANN) model and (2) a back propagation artifcial neural network model for particle swarm optimization (PSO-BP-ANN). Te results show that the established PSO-BP-ANN model has better reliability and prediction accuracy than the BP-ANN model. Wu et al. [13] established an ANN model to predict the bending deformation in welded thin-walled aluminium alloy tube structures. Te model's input variables include the four controllable position parameters of the weld, while the target output is the bending distortion in the x-and y-axis directions. A supervised multilayer feedforward back propagation (BP) neural network is proposed to estimate the bending distortion quickly and accurately. Te predicted values of the designed BP neural network are compared with the fnite element simulation results. Te results show that the proposed BP neural network model can accurately predict the bending distortion caused by welding over a range of welding position parameters. Jaafreh et al. [14] used a multiobjective evolution (MOE) algorithm and machine learning techniques to predict the age-hardening behavior of aluminium alloy under various machining conditions. Te results showed that combining the MOE algorithm and the machine learning process can successfully refne features and construct accurate machine learning prediction models compared to other feature selection and preprocessing methods. Sun et al. [15] used ANN methods to develop a correlation model between the thermal processing parameters and mechanical properties of Ti-6Al-4V alloy based on a series of forging and heat treatment experimental data. Te correlation model was developed using the ANN method. Te results show a reliable correlation between the process parameters and the mechanical properties of the Ti-6Al-4V alloy. Aykut et al. [16] improved the experimental model of surface roughness using ANN and response surface methodology (RSM).
ANN is widely used in various felds of material research [17,18], but ANN is very time-consuming in the training process. To overcome this shortcoming, the ELM algorithm is proposed. ELM, an emerging single hidden layer neural network algorithm [19], has received attention in material property prediction research due to its fast learning speed and high generalization capability. Cao et al. [20] used ELM optimized by the genetic algorithm (GA) with other numerical models (including fnite element model simulation) to determine the critical dimensions of prefabricated gear blanks with complex geometry for comparative analysis. Te results showed that the GA-ELM and R-GPLVM predictions were in good agreement with the experimental results. Sui and Lv [21] combined the ELM algorithm-based model with an attribute reduction method. With the attribute reduction method combining information entropy and Gram-Schmidt orthogonal transformation, the process parameters that can efectively afect quality were selected to form a subset of features. Compared with traditional modelling methods, the model has the advantages of a simple structure, low time consumption, and high prediction accuracy. Te prediction results show that the model has better adaptability to complex hot rolling processes, and the prediction performance is better than that of the traditional ELM model. Liu et al. [22] proposed a rolling force prediction method based on GA, PSO, and multihidden layer extreme learning machine (MELM), namely, the PSO-GA-MELM algorithm, which uses MELM as the primary model for rolling force prediction. In the modelling process, a genetic algorithm is used to determine the optimal number of hidden layers and nodes and a particle swarm algorithm is used to search for the optimal input weights and biases. Experimental results show that the rolling force prediction model trained by the algorithm has an excellent performance in prediction accuracy and can be used for the prediction of rolling force in the hot strip rolling process.
Although machine learning models, especially neural network type models, have been widely used in the study of steel materials, there has been little research in the prediction of mechanical properties of aluminium alloy materials. Te main objective of this study is to develop an optimised ELM model that can accurately predict the mechanical properties of aluminium alloy strips. Firstly, the model is built using a dataset obtained from an aluminium alloy strip production line. Secondly, the ELM model was optimised using the GWO algorithm. Finally, a comparative analysis of the accuracy of the model was carried out. Te experimental results show that the model is feasible to be applied to the prediction of mechanical properties of aluminium alloy strips.  [16]. Te algorithm principle of ELM algorithm is shown in equations (1)-(7) [19]. Te structure of the ELM algorithm is illustrated in Figure 1.

A Brief
As illustrated in Figure 1, we suppose that there exist N samples ( . . , t im , ] T ∈ R m , when the neural network has a hidden layer while that hidden layer has L hidden layer neurons, the output value of the neural network can be expressed as follows: where β i is the output weight between the frst i output weight between the neuron in the hidden layer and the neuron in the output layer, g( • ) is the activation function, and g(W i • X j + b i ) represents the input parameter X j of the output of the frst i output of the hidden layer neuron. W i � [w i1 , w i2 , w i3, . . . , w in ] T is the input weight between the frst i input weight between the frst input neuron and the hidden layer neuron, and b i is the output bias of the frst i output bias of the frst hidden layer neuron, and W i • X j denotes W i as the inner product of X j of the inner product.
Te smaller the error between the predicted and measured values of a single hidden layer neural network, the more accurate the model's prediction results, as shown in the following equation: i.e., parameters β i W i and b i exist as shown in the following equation: In the matrix form, this can be expressed as shown in the following equation: of which H(W 1 , W 2 , · · · , W L , b 1 , b 2 , · · · , b L , X 1 , X 2 , · · · , X L ) ,H represents the output of the nodes in the hidden layer, and β represents the output weights from the hidden layer to the output layer. T represents the desired output value with high prediction accuracy. As shown in equation (5), to obtain a single hidden layer neural network with the desired prediction accuracy, the parameters β * , W * , and b * are desired to be obtained.
where i � 1, 2, 3, · · · , L, is equivalent to making the loss function J obtain its minimum value, as shown in the following equation: Te process of training a single hidden layer neural network can be transformed into solving Hβ � T, which can solve for the matrix H of the Moore-Penrose generalized inverse matrix H † . Due to the characteristics of the ELM itself, i.e., the input weights W i and the hidden layer bias b i are determined randomly. Once the input weights W i and hidden layer bias b i are determined, the output matrix of the hidden layer H is uniquely determined. At this point, H † is the matrix H of the Moore-Penrose generalized inverse, as shown in the following equation: 2.2. GWO Algorithm. Te GWO algorithm is inspired by the hunting behavior of the grey wolf pack and is designed to simulate the group collaboration behavior of the grey wolf pack in the process of predation to achieve the purpose of optimization [23]. Te GWO algorithm has an adaptive convergence factor and a feedback mechanism that allows it to avoid falling into the trap of local optimality in the global search process. It, therefore, has higher accuracy and robustness in the problem-solving process. Te grey wolf is a canine predator that prefers to live in packs, with a clear hierarchical division of labor among its members during the hunt. Te hierarchical distribution of the grey wolf population is similar to the pyramid's structure, as shown in Figure 2. Te pack's leader is located at the frst level of the pyramid and is called α. Te think tank team of the pack at the second level of the pyramid is called β. β is second only to α in terms of leadership in the pack, and when there is a vacancy in α′s position in the pack, β will fll the vacant position and become the new α. Te executors of the pack are located at the third level of the pyramid and are Advances in Materials Science and Engineering called δ. Te wolves at the bottom of the pyramid are called ω and are responsible for executing αβ and δ′s commands. Te strict hierarchical distribution is directly refected in the hunting activities of the grey wolf, which will be led by α. During the hunting process, αβδ and ω will strictly follow their duties and have a clear division of labor and the hunting process can be divided into three main steps, namely, fnding and surrounding the prey; harassing and tracking the prey; and attacking the prey.
Te algorithm principle of the GWO algorithm is shown in equations (8)-(15) [23]. Te grey wolf optimization algorithm abstracts the hunting pattern of a wolf pack into an optimization algorithm. In the process of fnding and encircling prey, the grey wolf's process of encircling prey is defned in equations (8) and (9) as follows. Equation (8) represents the distance between the grey wolf and the prey. Here, D → represents the direction and distance travelled by individual grey wolves. C → is the parameter used to determine the direction of the wolf pack's search for prey, t represents the number of iterations of grey wolf positions, and X → P (t) represents the position that individual grey wolves tend to approach. X → (t) represents the position of the individual grey wolf when the iteration number is t. Equation (9) represents the iterative formula for the position of the grey wolf, and A → is the parameter used to adjust the radius of the prey search. A → and C → calculation formulas are shown in equations (10) and (11). a is the convergence factor, as shown in equation (12), and a decreases from a maximum value of 2 until it becomes 0 as the number of iterations increases. r 1 → with r 2 → is a random vector with the modulus length set in the range [0, 1].
In the practical application of the optimization algorithm, the optimal three solutions are assigned to α, β, and δ according to their ftness, making αβδ the leader of the pack closest to the prey. In turn, the other grey wolves adjust their positions according to the best-positioned grey wolf, calculate their ftness after the position update, and rerank the grey wolves according to their ftness size. Te mechanism for individuals in the pack to update their positions is shown in Figure 3. Te mathematical model of individual grey wolves updating their positions can be shown as equation (13). Here, D → α D → β D → δ represents the distance relationship between α and β δ individuals and other individuals, respectively, represents a random vector, and X → represents the current position of an individual grey wolf. In a wolf pack ω, the distance and direction of the individual to the target are shown in equations (14) and (15).
From equations (10) and (12), the value of a will gradually decrease with the progress of iteration and the fuctuation range of parameter A → will also decrease. Te grey wolf algorithm uses the | A → | and C → to avoid the solution of the optimization algorithm falling into a local optimum. As shown in Figure 4(a), when |A → | > 1, the grey wolf individual moves away from the target, i.e., a global search is performed. As shown in Figure 4(b), when |A → | < 1, the individual grey wolf will launch a fnal attack on the prey, i.e., the optimal solution is obtained. From equation (11), it can be seen that C is a random value taking values in the interval [0, 2]. C represents the size of the hindrance for the grey wolf to approach the prey. C > 1 means that the grey wolf individual is not easy to approach the target; on the contrary, C < 1 means that the grey wolf individual is easy to approach the target and C can maintain randomness during the iterative process and helps the algorithm escape in time when it is about to fall into a local optimum trap. Tis algorithm feature also contributes to the ease with which the grey wolf optimization algorithm can obtain a global optimum.

MGWO Algorithm.
Te MGWO algorithm was proposed by Zhang and Zhou [24] in 2021. Te algorithm considers that a with diferent update strategies can signifcantly afect the algorithm's performance [25]. Te algorithm principle of the MGWO algorithm is shown in equations (16)-(18) [24]. It uses a convergence factor update based on regular exponential changes, as shown in equation (16). In addition, to enhance the balance between global search and local exploitation, the MGWO algorithm proposes a new adaptive shifting strategy [24,26], whose mathematical expression is shown in equation (17). Here, the weights W i of the mathematical expression are shown in equation (18).

Proposed GWO-ELM Prediction Model
ELM is an efective mechanism for pattern classifcation and regression learning. In the ELM algorithm model, the number of hidden layer neurons needs to be determined in advance and the structure of the neural network has a direct impact on the accuracy of the ELM algorithm model, which relies on a large number of hidden layer nodes for good performance. As the number of nodes in the hidden layer increases, the computational cost increases signifcantly [27]. An excessive number of hidden layer neurons will make an ELM model so complex that it will slow down the prediction speed or even overft. However, too few hidden layer neurons will result in an ELM model that does not achieve the desired accuracy. To date, there is no accepted and valid theory to guide the selection of the number of hidden layer neurons. In most cases, the determination of the number of hidden layer neurons relies on the experience of the model trainer to obtain the appropriate neural network parameters. If the essential parameters of the algorithm depend too much on manual selection, then the robustness of the model will be reduced. In addition, the weights and biases of the ELM model are randomly generated, which simplifes the algorithm but also has a signifcant negative impact on the algorithm's accuracy, as the randomness of the parameters makes the model's accuracy highly uncertain. Mean Absolute Percentage Error (MAPE) mean square error (MSE) and root mean square error (RMSE) will be used to evaluate the performance of the model. Te calculation methods of MAPE MSE RMSE and R 2 are shown in equations (19)- (22), respectively [15].

Advances in Materials Science and Engineering
where n represents the number of samples to be substituted into the formula for operation, _ y i represents the predicted value, y i represents the actual value, and y represents the mean.
Te suitable combination of the number of hidden layer neurons with the input weights and biases becomes the focus of improving the prediction accuracy of ELM. Terefore, this paper introduces the GWO algorithm to optimize the ELM algorithm, forming a new hybrid algorithm named GWO-ELM. Te fow of the GWO-ELM algorithm is shown in Figure 5. In the modelling process of the GWO-ELM algorithm, the population size, objective function, and the maximum number of iterations of the GWO algorithm are frst determined. Ten, the optimal number of hidden layer neurons, the optimal input weights, and optimal bias of the ELM network are determined using the GWO algorithm. Te GWO algorithm then determines the optimal number of hidden layer neurons, the optimal input weights, and the ELM network's optimal bias.
Te specifc steps of the GWO-ELM algorithm are as follows: (1) We determine the input data. Te raw data were preprocessed and randomly divided into training and test sets. (2) We initialize the GWO algorithm and determine the number of grey wolves in the grey wolf algorithm population, the maximum number of iterations, and the number of parameters to be optimized. We determine the parameters A and C. We take i � 0.  structure is obtained when the number of neurons in the hidden layer is L. (9) To determine whether L reaches the maximum number of hidden layer neurons. If it does not reach the maximum number of hidden layer neurons, step (3) is repeated; otherwise, the model's prediction accuracy can be compared when L takes diferent values. Finally, the optimal ELM network structure can be obtained.

Model Structure Building and Experimental Analysis
Extensive experiments were conducted to test the prediction accuracy of the proposed GWO-ELM model and compare and analyze the performance of the GWO-ELM model with that of the conventional ELM model and the MGWO-ELM model.

Acquisition and Processing of Training Samples.
Te prediction performance of the extreme learning machine is very dependent on the quality of the input data, and redundant or conficting data will not result in the desired prediction accuracy [28]. High-quality data are the basis for establishing accurate models. Tis paper uses the actual production data of Shandong Nanshan Aluminum Industry Co., LTD, as experimental data. Te data come from the big data platform database of the enterprise. Tis data are available and can be obtained as shown in the Data Availability section. Te test standard number GB/T 16865-2013 was used to obtain data on material properties. In the experiments of this paper, the correlation between the input and output parameters was frst analyzed using correlation analysis algorithms and physical metallurgy principles. Parameters with signifcant infuence on tensile strength, yield strength, and elongation were selected as input parameters for the model, and a total of 22 input parameters were extracted. Te input parameters of the model are shown in Table 1. Tis data are available and can be obtained as shown in the Data Availability section. Te output parameters were set to tensile strength, yield strength, and elongation, and the output parameters are shown in Table 2. Table 1 shows the statistics for the mean, standard deviation, and minimum and maximum values for each input parameter. Table 2 shows each output parameter's mean, standard deviation, and minimum and maximum value statistics. After removing the outlier data for the initially identifed parameters, the mean values were used to fll in the vacant values. Te processed data will be used for model training and testing. A total of 131 production data sets on 6series aluminium alloys and 5-series aluminium alloys were collected and applied. Te main compositions of the aluminium alloys used for the data in this paper are shown in Table 3. Seventy percent of the total data (91 sets) will be assigned to the training set and used to train the model to have the desired prediction accuracy. Te remaining data (40 sets) will be allocated to the test set to test the model's performance.
Data processing needs to avoid the adverse impact of the diference in the order of magnitude between diferent parameters on the prediction results, so the input parameters need to be normalized before input into the model. Te input parameters need to be normalized before being fed into the model. Te data normalization method is shown in equation (23), the input parameters will be normalized to [0, 1] [29]. Here, x represents the normalized data, x max and x min represent the maximum and minimum values of data in the dimension, respectively, and x i represents the normalized result.

Determination of the Parameters of the GWO-ELM
Algorithm. Te network structure of the GWO-ELM model is closely related to the data structure. Te dimensionality of the input determines the number of neurons in the input layer. Te dimensionality of the output determines the number of neurons in the output layer. Te input parameter is 22 dimensions, so the number of neurons in the input layer is set to 22; the output parameter is 3 dimensions, so the number of neurons in the output layer is set to 3. In the following experiment, 70% of the data will be used to train the model and the remaining 30% will be used to test the performance of the model. Te entire simulation process in this experiment was carried out using Python 3.6. Firstly, the activation function was determined by the experiment [30]. Te number of neurons in the hidden layer was tentatively set to 22 to test the model's prediction accuracy with different activation functions. Te experimental results are shown in Table 4. Te mean square error of the test set and the mean absolute percentage error of the test set were minimized when the activation function was "sigmoid." Te mean squared error of the training set for the "sigmoid" activation function is 316.72, the mean squared error of the test set is 594.99, and the total mean squared error of the prediction is 401.68. Te mean squared error of the training set for the "sigmoid" activation function and the mean absolute percentage error for the training set are 8.02, the mean absolute percentage error for the test set is 10.45, and the total mean absolute percentage error is 8.76. Te ELM algorithm activation function is set to "sigmoid." Next, experiments are used to determine the optimal number of hidden layer neurons for the ELM algorithm. Te number of hidden layer neurons has a large impact on the prediction results of the ELM model, and too many or too few hidden layer neurons will decrease the model's prediction accuracy. In this section, the ELM activation function is "sigmoid" and the other parameters remain unchanged. Te number of ELM hidden layer neurons was determined to be in the range [1,38] based on the dimensionality of the input parameters used in this experiment. Te performance of the ELM model with diferent numbers of hidden layer neurons was compared, and the results are shown in Figure 6. When the number of neurons in the hidden layer was less than 23, for tensile strength, yield Advances in Materials Science and Engineering strength, and elongation, the prediction accuracy of the test set increased with the increase of the number of neurons. When the number of neurons in the hidden layer exceeds 23, especially the prediction accuracy of elongation, the prediction accuracy of the test set shows a trend of fuctuation decline. Te reason for the previously mentioned phenomenon is that with the increase of the number of hidden layer neurons, the overftting tendency of the model increases and the overftting leads to the abnormal increase of the prediction error of the test set. As shown in Figure 6, the prediction accuracy reached its peak when the number of hidden layer neurons was 23.
Next, the population size of the grey wolf algorithm was determined by experiments. Te number of iterations was set to 300, and the number of individuals in the population ranged from [10,100]. Te prediction results of the test set     Figure 7. Te prediction results of the test set showed low prediction errors when the population size was 67, 78, and 86. Still, too large population would cause the model training time to be extended, resulting in unnecessary time wastage. Te choice of objective function has a massive impact on the performance of the GWO algorithm. When the model is a multioutput algorithm, the infuence of the weights between diferent objectives on the prediction accuracy of the fnal algorithm is of interest. To investigate the efect of the objective function weights on the mechanical properties of aluminium alloys, the objective function used in this paper is shown in equation (24). Here, P T , P Y , and P E are the predicted values of tensile strength, yield strength, and elongation, respectively; E T E Y E E are the measured values of tensile strength, yield strength, and elongation, respectively; N is the number of measured values; j i is the penalty coefcient of yield strength and elongation, respectively. Te number of iterations of the GWO algorithm is set to 100, and the trends of tensile strength, yield strength, and elongation with the change of penalty coefcients are shown in Figure 8. As an example, Figure 8 Te trend of variance of MAPE for the three output variables is shown in Figure 9, which indicates that when the penalty coefcients j i are 80 and 20, respectively, the corresponding MAPE variances for tensile strength, yield strength, and elongation are smaller. When the penalty coefcients are 80 and 20, respectively, the predicted values of tensile strength, yield strength, and elongation can maintain a balance of the predicted values of the three mechanical properties while ensuring a small accuracy of their own. Te penalty coefcients in the objective function equation (23) j i are chosen as 80, 20. Te comparison between the optimal penalty coefcient and the original coefcient is shown in Table 5. Te optimal penalty coefcient can not only reduce the MAPE of the test set but also signifcantly reduce the variance of the prediction errors of the three variables. A lower variance means that the prediction error distribution of the three variables is more balanced.

Advances in Materials Science and Engineering
Te number of iterations of the GWO algorithm signifcantly impacts the prediction results, and each iteration will obtain a better value of the ELM algorithm parameters than the previous iteration. Te GWO algorithm population size was set to 30, the ftness function was set to equation (24), and i � 20, j � 80. Te relationship between the prediction accuracy of the GWO-ELM model test set and the number of iterations, as well as the model training time and the number of iterations, is shown in Figure 10. As the iterations proceeded, the R 2 of the test set increases gradually and the modelling time increased linearly with the number of iterations. When the number of iterations reaches 3150, the prediction accuracy of tensile strength, yield strength, and elongation converges.
In summary, the parameters of the GWO-ELM algorithm have been determined and are shown in Table 6. When the values of parameters are shown in Table 6, the iterative optimization of weight ω and bias b by the GWO algorithm can be realized. Tus, when the maximum number of iterations is reached, the optimal input weight ω and bias b of the ELM network in the GWO-ELM model are determined.

Experimentation and Evaluation of Model Performance.
In this section, mechanical property prediction experiments demonstrate the excellent prediction accuracy of the proposed GWO-ELM model. Te model comparison experiment of the GWO-ELM model, MGWO-ELM model, and traditional ELM model is designed to verify the advantages of the proposed GWO-ELM model. Each model was run through twenty replicates, and the results were averaged to avoid misleading results from model chance.
To ensure a fair comparison, the corresponding parameters for the GWO-ELM model and the ELM model in this experiment are shown in Table 6. Te parameters used by the MGWO algorithm are consistent with the GWO algorithm. MAPE and RMSE were error measures for the mechanical properties' prediction models. Te prediction errors of the three models after twenty replicate experiments for the three mechanical properties are shown in Table 7. Table 7 Figure 9: Trend of MAPE variance with the penalty coefcient for the test set of the three output variables.   Te abovementioned results show that GWO-ELM and MGWO-ELM achieve an ideal ftting efect in the prediction of tensile strength, yield strength, and elongation, and the performance of the two models is far better than that of the ELM model. In addition, GWO-ELM was better than MGWO-ELM in the prediction of tensile strength and elongation, so the GWO-ELM model was slightly better than MGWO-ELM.
Te predicted results of the GWO-ELM model, MGWO-ELM model, and ELM model compared to the actual values after twenty replicate experiments for the three mechanical properties are shown in Figure 11. Both in the training set and the test set, the prediction accuracy of the yield strength and elongation of the tensile strength is higher in the GWO-ELM model and the ft degree of the yield strength and elongation of the tensile strength is higher in the GWO-ELM model. Te error distributions of the three model test sets are shown in Table 8. For the three mechanical properties of tensile strength, yield strength, and elongation, the GWO-ELM model has a more obvious advantage in the prediction results of the test set. Tis is especially true for the tensile and yield strengths. When the prediction deviation is controlled within ±10%, the prediction accuracy of tensile strength in the GWO-ELM model, MGWO-ELM model, and ELM model is 100%, 97.5%, and 80%, respectively. Te prediction accuracy of yield strength of the GWO-ELM model, MGWO-ELM model, and ELM model is 97.5%, 92.5%, and 62.5%, respectively. Te prediction accuracy of elongation in the GWO-ELM model, MGWO-ELM model, and ELM model is 77.5%, 77.5%, and 37.5%, respectively. Although the prediction accuracy of elongation of GWO-ELM and MGWO-ELM is the same when the error is controlled within ±10%, the prediction accuracy of the GWO-ELM model is better than that of MGWO-ELM when the prediction error is 5%. Te abovementioned results indicate that the prediction accuracy of GWO-ELM is more stable than MGWO-ELM and ELM models.
In the prediction results of the test set, the absolute error distribution of the prediction results of tensile strength, yield strength, and elongation in diferent models is shown in Figure 12. As shown in Figure 12(a), the prediction error range of tensile strength in the GWO-ELM model is smaller than that of the other two models. As shown in Figure 12(b), the prediction error range of yield strength on the GWO-ELM model is smaller than that of the other two models. As shown in     14 Advances in Materials Science and Engineering Figure 12(b), the prediction error range of yield strength in the GWO-ELM model is signifcantly smaller than that of the other two models. Te GWO-ELM model performs better than the MGWO-ELM model in predicting tensile strength and yield strength but not in predicting elongation. As shown in Figure 12(c), the MGWO-ELM model is slightly better than the GWO-ELM model in the prediction of elongation. In general, compared with the other two models, the prediction accuracy of tensile strength, yield strength, and elongation can be higher in the GWO-ELM model.

Conclusion
Tis paper proposes an intelligent optimization algorithm, which has the advantages of a simple structure, short training time, and high prediction accuracy. Tis algorithm is successfully applied to the prediction of mechanical properties of aluminium alloy strips in this paper and eventually achieves high prediction accuracy, providing a new method for predicting the mechanical properties of aluminium alloy strips, which has broad application prospects.
(1) Preprocessed industrial production data were used to train and test the GWO-ELM model. Trough several experiments, the optimal network structure of the ELM algorithm and the optimal parameters of the Grey Wolf algorithm were determined. Te weights and biases of the ELM were determined by the GWO algorithm. Te optimal GWO-ELM model structure was successfully determined.

Data Availability
Te data used to support the fndings of this study are available from the corresponding author upon request.

Conflicts of Interest
All authors declare that they have no conficts of interest.