Abstract

Mechanical properties are important indicators for evaluating the quality of strips. This paper proposes a mechanical performance prediction model based on the Gray Wolf Optimization (GWO) algorithm and the Extreme Learning Machine (ELM) algorithm. In the modeling process, GWO is used to determine the optimal weights and deviations of ELM and experiments are used to determine the model’s key parameters. The model effectively avoids manual intervention and significantly improves aluminum alloy strips’ mechanical property prediction accuracy. This paper uses processed data from the aluminum alloy production plant of Shandong Nanshan Aluminum Co., Ltd. as experimental data. When the prediction deviation is controlled within ±10%, the GWO-ELM model can achieve a correct rate of 100% for tensile strength, 97.5% for yield strength, and 77.5% for elongation on the test set. The RMSE of the tensile strength, yield strength, and elongation of the GWO-ELM model was 5.365, 11.881, and 1.268, respectively. The experimental results show that the GWO-ELM model has higher accuracy and stability in predicting aluminum alloy strips’ tensile strength, yield strength, and elongation. The GWO-ELM model effectively avoids the defects of the traditional model. It has a special guiding significance for producing aluminum alloy strips.

1. Introduction

Among the many indicators of hot-rolled alloy products, mechanical properties, which are influenced by the complexity of processing parameters, are one of the important indicators of the quality of hot-rolled alloy products. Therefore, the mechanical properties of alloy products are valued by many manufacturers and researchers. Accurate prediction of the mechanical properties of hot-rolled aluminium alloys has become one of the hot topics in the development and application of aluminium alloys [1, 2]. Up to now, there are two main models in the field of metal property prediction. One of them is the construction of material constitutive models for material property prediction [3]. Jia et al. [4] used experimentally obtained stress-strain data to calculate the material constants involved in the Arrhenius-type constitutive model and the modified Zerilli–Armstrong (MZA) model. Zhang et al. [5] established the constitutive model for the thermal deformation behavior of high-strength aluminium alloy (Al-Zn-Mg-Cu). The results show that the constitutive model can accurately predict the flow behavior of the Al-Zn-Mg-Cu alloy. Although the emergence of the constitutive model enables researchers to reasonably predict the product performance according to the processing technology, the construction of the constitutive model still needs experimental support and reasonable assumptions, which leads to the fact that the constitutive model cannot fully meet the requirements of industrialization. To meet the needs of industrial production, another model emerged along with the maturity of computer technology, which uses intelligent algorithms to analyze large amounts of real production data to obtain more desirable predictions of mechanical properties. Along with the development of computer technology, artificial intelligence algorithms play an increasingly important role in materials’ research. Various machine learning algorithms applied to materials’ research have solved many challenging problems [6]. The combination of computer technology and materials’ science will likely yield more important achievements in the foreseeable future [7]. Guo et al. [8] proposed a method for multiproperty prediction of materials using the Interior Point Algorithm. The rationality and reliability of the above theory are verified successfully by the experiment using the data of the steel production process. Among the many machine learning algorithms on which statistical models rely, neural network-like algorithms are favored and widely used in the field of material property prediction due to their excellent properties [9, 10]. Niklas et al. [11] needed to develop accurate models to ensure efficient control equipment for superelastic shape memory alloys (SMA). After training, the ANN can successfully calculate model parameters from cyclic tensile stress-strain tests, thereby improving the efficiency of SMA control devices. Lan et al. [12] experimentally established a dataset of the chemical composition, processing parameters, and mechanical properties of the A380 alloy. The Fe/Mn content ratio, Sr content, cooling rate, and porosity content were the input variables, and ultimate tensile strength (UTS) and elongation (El) were used as output parameters. Two ANN models were developed, namely, (1) a back propagation artificial neural network (BP-ANN) model and (2) a back propagation artificial neural network model for particle swarm optimization (PSO-BP-ANN). The results show that the established PSO-BP-ANN model has better reliability and prediction accuracy than the BP-ANN model. Wu et al. [13] established an ANN model to predict the bending deformation in welded thin-walled aluminium alloy tube structures. The model’s input variables include the four controllable position parameters of the weld, while the target output is the bending distortion in the x- and y-axis directions. A supervised multilayer feedforward back propagation (BP) neural network is proposed to estimate the bending distortion quickly and accurately. The predicted values of the designed BP neural network are compared with the finite element simulation results. The results show that the proposed BP neural network model can accurately predict the bending distortion caused by welding over a range of welding position parameters. Jaafreh et al. [14] used a multiobjective evolution (MOE) algorithm and machine learning techniques to predict the age-hardening behavior of aluminium alloy under various machining conditions. The results showed that combining the MOE algorithm and the machine learning process can successfully refine features and construct accurate machine learning prediction models compared to other feature selection and preprocessing methods. Sun et al. [15] used ANN methods to develop a correlation model between the thermal processing parameters and mechanical properties of Ti-6Al-4V alloy based on a series of forging and heat treatment experimental data. The correlation model was developed using the ANN method. The results show a reliable correlation between the process parameters and the mechanical properties of the Ti-6Al-4V alloy. Aykut et al. [16] improved the experimental model of surface roughness using ANN and response surface methodology (RSM).

ANN is widely used in various fields of material research [17, 18], but ANN is very time-consuming in the training process. To overcome this shortcoming, the ELM algorithm is proposed. ELM, an emerging single hidden layer neural network algorithm [19], has received attention in material property prediction research due to its fast learning speed and high generalization capability. Cao et al. [20] used ELM optimized by the genetic algorithm (GA) with other numerical models (including finite element model simulation) to determine the critical dimensions of prefabricated gear blanks with complex geometry for comparative analysis. The results showed that the GA-ELM and R-GPLVM predictions were in good agreement with the experimental results. Sui and Lv [21] combined the ELM algorithm-based model with an attribute reduction method. With the attribute reduction method combining information entropy and Gram–Schmidt orthogonal transformation, the process parameters that can effectively affect quality were selected to form a subset of features. Compared with traditional modelling methods, the model has the advantages of a simple structure, low time consumption, and high prediction accuracy. The prediction results show that the model has better adaptability to complex hot rolling processes, and the prediction performance is better than that of the traditional ELM model. Liu et al. [22] proposed a rolling force prediction method based on GA, PSO, and multihidden layer extreme learning machine (MELM), namely, the PSO-GA-MELM algorithm, which uses MELM as the primary model for rolling force prediction. In the modelling process, a genetic algorithm is used to determine the optimal number of hidden layers and nodes and a particle swarm algorithm is used to search for the optimal input weights and biases. Experimental results show that the rolling force prediction model trained by the algorithm has an excellent performance in prediction accuracy and can be used for the prediction of rolling force in the hot strip rolling process.

Although machine learning models, especially neural network type models, have been widely used in the study of steel materials, there has been little research in the prediction of mechanical properties of aluminium alloy materials. The main objective of this study is to develop an optimised ELM model that can accurately predict the mechanical properties of aluminium alloy strips. Firstly, the model is built using a dataset obtained from an aluminium alloy strip production line. Secondly, the ELM model was optimised using the GWO algorithm. Finally, a comparative analysis of the accuracy of the model was carried out. The experimental results show that the model is feasible to be applied to the prediction of mechanical properties of aluminium alloy strips.

2. A Brief Introduction to MGWO, GWO, and ELM

2.1. ELM Algorithm

ELM was born in 2004 to overcome the limitation of the complex and slow iterative structure of traditional neural networks. ELM’s single hidden layer feedforward neural network randomly selects the weights and biases of the number of hidden nodes. ELM has received a lot of attention from scholars since its introduction because, compared with traditional neural networks, ELM as single hidden layer feedforward neural networks (SLFNs) can have faster iteration speed and better generalization while ensuring accuracy [16]. The algorithm principle of ELM algorithm is shown in equations (1)–(7) [19]. The structure of the ELM algorithm is illustrated in Figure 1.

As illustrated in Figure 1, we suppose that there exist N samples (, ), where  =  ∈ ,  =  ∈ , when the neural network has a hidden layer while that hidden layer has L hidden layer neurons, the output value of the neural network can be expressed as follows:where is the output weight between the first output weight between the neuron in the hidden layer and the neuron in the output layer, is the activation function, and represents the input parameter of the output of the first output of the hidden layer neuron.  =  is the input weight between the first input weight between the first input neuron and the hidden layer neuron, and is the output bias of the first output bias of the first hidden layer neuron, and denotes as the inner product of of the inner product.

The smaller the error between the predicted and measured values of a single hidden layer neural network, the more accurate the model’s prediction results, as shown in the following equation:i.e., parameters and exist as shown in the following equation:

In the matrix form, this can be expressed as shown in the following equation:of which ,,, represents the output of the nodes in the hidden layer, and represents the output weights from the hidden layer to the output layer. represents the desired output value with high prediction accuracy. As shown in equation (5), to obtain a single hidden layer neural network with the desired prediction accuracy, the parameters , , and are desired to be obtained.where L, is equivalent to making the loss function obtain its minimum value, as shown in the following equation:

The process of training a single hidden layer neural network can be transformed into solving , which can solve for the matrix of the Moore–Penrose generalized inverse matrix . Due to the characteristics of the ELM itself, i.e., the input weights and the hidden layer bias are determined randomly. Once the input weights and hidden layer bias are determined, the output matrix of the hidden layer is uniquely determined. At this point, is the matrix of the Moore–Penrose generalized inverse, as shown in the following equation:

2.2. GWO Algorithm

The GWO algorithm is inspired by the hunting behavior of the grey wolf pack and is designed to simulate the group collaboration behavior of the grey wolf pack in the process of predation to achieve the purpose of optimization [23]. The GWO algorithm has an adaptive convergence factor and a feedback mechanism that allows it to avoid falling into the trap of local optimality in the global search process. It, therefore, has higher accuracy and robustness in the problem-solving process.

The grey wolf is a canine predator that prefers to live in packs, with a clear hierarchical division of labor among its members during the hunt. The hierarchical distribution of the grey wolf population is similar to the pyramid’s structure, as shown in Figure 2. The pack’s leader is located at the first level of the pyramid and is called α. The think tank team of the pack at the second level of the pyramid is called β. β is second only to α in terms of leadership in the pack, and when there is a vacancy in α′s position in the pack, β will fill the vacant position and become the new α. The executors of the pack are located at the third level of the pyramid and are called δ. The wolves at the bottom of the pyramid are called ω and are responsible for executing αβ and δ′s commands. The strict hierarchical distribution is directly reflected in the hunting activities of the grey wolf, which will be led by α. During the hunting process, αβδ and ω will strictly follow their duties and have a clear division of labor and the hunting process can be divided into three main steps, namely, finding and surrounding the prey; harassing and tracking the prey; and attacking the prey.

The algorithm principle of the GWO algorithm is shown in equations (8)–(15) [23]. The grey wolf optimization algorithm abstracts the hunting pattern of a wolf pack into an optimization algorithm. In the process of finding and encircling prey, the grey wolf’s process of encircling prey is defined in equations (8) and (9) as follows. Equation (8) represents the distance between the grey wolf and the prey. Here, represents the direction and distance travelled by individual grey wolves. is the parameter used to determine the direction of the wolf pack’s search for prey, represents the number of iterations of grey wolf positions, and represents the position that individual grey wolves tend to approach. represents the position of the individual grey wolf when the iteration number is t. Equation (9) represents the iterative formula for the position of the grey wolf, and is the parameter used to adjust the radius of the prey search. and calculation formulas are shown in equations (10) and (11). is the convergence factor, as shown in equation (12), and decreases from a maximum value of 2 until it becomes 0 as the number of iterations increases. with is a random vector with the modulus length set in the range [0, 1].

In the practical application of the optimization algorithm, the optimal three solutions are assigned to α, β, and δ according to their fitness, making αβδ the leader of the pack closest to the prey. In turn, the other grey wolves adjust their positions according to the best-positioned grey wolf, calculate their fitness after the position update, and rerank the grey wolves according to their fitness size. The mechanism for individuals in the pack to update their positions is shown in Figure 3. The mathematical model of individual grey wolves updating their positions can be shown as equation (13). Here, represents the distance relationship between α and β δ individuals and other individuals, respectively, represents, respectively, the the current position of the individual, represents a random vector, and represents the current position of an individual grey wolf. In a wolf pack ω, the distance and direction of the individual to the target are shown in equations (14) and (15).

From equations (10) and (12), the value of will gradually decrease with the progress of iteration and the fluctuation range of parameter will also decrease. The grey wolf algorithm uses the and to avoid the solution of the optimization algorithm falling into a local optimum. As shown in Figure 4(a), when || > 1, the grey wolf individual moves away from the target, i.e., a global search is performed. As shown in Figure 4(b), when || < 1, the individual grey wolf will launch a final attack on the prey, i.e., the optimal solution is obtained. From equation (11), it can be seen that is a random value taking values in the interval [0, 2]. C represents the size of the hindrance for the grey wolf to approach the prey. C > 1 means that the grey wolf individual is not easy to approach the target; on the contrary, C < 1 means that the grey wolf individual is easy to approach the target and C can maintain randomness during the iterative process and helps the algorithm escape in time when it is about to fall into a local optimum trap. This algorithm feature also contributes to the ease with which the grey wolf optimization algorithm can obtain a global optimum.

2.3. MGWO Algorithm

The MGWO algorithm was proposed by Zhang and Zhou [24] in 2021. The algorithm considers that with different update strategies can significantly affect the algorithm’s performance [25]. The algorithm principle of the MGWO algorithm is shown in equations (16)–(18) [24]. It uses a convergence factor update based on regular exponential changes, as shown in equation (16). In addition, to enhance the balance between global search and local exploitation, the MGWO algorithm proposes a new adaptive shifting strategy [24, 26], whose mathematical expression is shown in equation (17). Here, the weights of the mathematical expression are shown in equation (18).

3. Proposed GWO-ELM Prediction Model

ELM is an effective mechanism for pattern classification and regression learning. In the ELM algorithm model, the number of hidden layer neurons needs to be determined in advance and the structure of the neural network has a direct impact on the accuracy of the ELM algorithm model, which relies on a large number of hidden layer nodes for good performance. As the number of nodes in the hidden layer increases, the computational cost increases significantly [27]. An excessive number of hidden layer neurons will make an ELM model so complex that it will slow down the prediction speed or even overfit. However, too few hidden layer neurons will result in an ELM model that does not achieve the desired accuracy. To date, there is no accepted and valid theory to guide the selection of the number of hidden layer neurons. In most cases, the determination of the number of hidden layer neurons relies on the experience of the model trainer to obtain the appropriate neural network parameters. If the essential parameters of the algorithm depend too much on manual selection, then the robustness of the model will be reduced. In addition, the weights and biases of the ELM model are randomly generated, which simplifies the algorithm but also has a significant negative impact on the algorithm’s accuracy, as the randomness of the parameters makes the model’s accuracy highly uncertain.

Mean Absolute Percentage Error (MAPE) mean square error (MSE) and root mean square error (RMSE) will be used to evaluate the performance of the model. The calculation methods of MAPE MSE RMSE and are shown in equations (19)–(22), respectively [15].where represents the number of samples to be substituted into the formula for operation, represents the predicted value, represents the actual value, and represents the mean.

The suitable combination of the number of hidden layer neurons with the input weights and biases becomes the focus of improving the prediction accuracy of ELM. Therefore, this paper introduces the GWO algorithm to optimize the ELM algorithm, forming a new hybrid algorithm named GWO-ELM. The flow of the GWO-ELM algorithm is shown in Figure 5. In the modelling process of the GWO-ELM algorithm, the population size, objective function, and the maximum number of iterations of the GWO algorithm are first determined. Then, the optimal number of hidden layer neurons, the optimal input weights, and optimal bias of the ELM network are determined using the GWO algorithm. The GWO algorithm then determines the optimal number of hidden layer neurons, the optimal input weights, and the ELM network’s optimal bias.

The specific steps of the GWO-ELM algorithm are as follows:(1)We determine the input data. The raw data were preprocessed and randomly divided into training and test sets.(2)We initialize the GWO algorithm and determine the number of grey wolves in the grey wolf algorithm population, the maximum number of iterations, and the number of parameters to be optimized. We determine the parameters A and C. We take i = 0.(3)We initialize the weights and biases of the ELM, at which point the number of hidden layer neurons is no longer determined empirically but by experimental comparison. At this point, the number of hidden layer neurons is L = i + 1.(4)We calculate the fitness of each individual in the pack to determine the best individual, in which MAPE is used as the individual’s fitness. The smaller the error of an individual, the better it is. We determine α, β, δ, and ω in the wolf pack based on the individual error in the pack.(5)We update the position and parameters of each grey wolf A and C.(6)We update the input weight ω and the bias b of the ELM model from the results of step (5) so that the parameters of the ELM model take values closer to the optimal values.(7)We calculate the fitness of all grey wolves and update αβδ and its fitness. The individual prediction error size determines that individual’s leadership position in the wolf pack. That is, the αβδ of wolves is determined according to the error of individual prediction values from small to large.(8)We determine whether the maximum number of iterations is reached; if not, we return to step (5); otherwise, the optimal limit learning machine structure is obtained when the number of neurons in the hidden layer is L.(9)To determine whether L reaches the maximum number of hidden layer neurons. If it does not reach the maximum number of hidden layer neurons, step (3) is repeated; otherwise, the model’s prediction accuracy can be compared when L takes different values. Finally, the optimal ELM network structure can be obtained.

4. Model Structure Building and Experimental Analysis

Extensive experiments were conducted to test the prediction accuracy of the proposed GWO-ELM model and compare and analyze the performance of the GWO-ELM model with that of the conventional ELM model and the MGWO-ELM model.

4.1. Acquisition and Processing of Training Samples

The prediction performance of the extreme learning machine is very dependent on the quality of the input data, and redundant or conflicting data will not result in the desired prediction accuracy [28]. High-quality data are the basis for establishing accurate models. This paper uses the actual production data of Shandong Nanshan Aluminum Industry Co., LTD, as experimental data. The data come from the big data platform database of the enterprise. This data are available and can be obtained as shown in the Data Availability section. The test standard number GB/T 16865-2013 was used to obtain data on material properties. In the experiments of this paper, the correlation between the input and output parameters was first analyzed using correlation analysis algorithms and physical metallurgy principles. Parameters with significant influence on tensile strength, yield strength, and elongation were selected as input parameters for the model, and a total of 22 input parameters were extracted. The input parameters of the model are shown in Table 1. This data are available and can be obtained as shown in the Data Availability section. The output parameters were set to tensile strength, yield strength, and elongation, and the output parameters are shown in Table 2. Table 1 shows the statistics for the mean, standard deviation, and minimum and maximum values for each input parameter. Table 2 shows each output parameter’s mean, standard deviation, and minimum and maximum value statistics. After removing the outlier data for the initially identified parameters, the mean values were used to fill in the vacant values. The processed data will be used for model training and testing. A total of 131 production data sets on 6-series aluminium alloys and 5-series aluminium alloys were collected and applied. The main compositions of the aluminium alloys used for the data in this paper are shown in Table 3. Seventy percent of the total data (91 sets) will be assigned to the training set and used to train the model to have the desired prediction accuracy. The remaining data (40 sets) will be allocated to the test set to test the model’s performance.

Data processing needs to avoid the adverse impact of the difference in the order of magnitude between different parameters on the prediction results, so the input parameters need to be normalized before input into the model. The input parameters need to be normalized before being fed into the model. The data normalization method is shown in equation (23), the input parameters will be normalized to [0, 1] [29]. Here, represents the normalized data, and represent the maximum and minimum values of data in the dimension, respectively, and represents the normalized result.

4.2. Determination of the Parameters of the GWO-ELM Algorithm

The network structure of the GWO-ELM model is closely related to the data structure. The dimensionality of the input determines the number of neurons in the input layer. The dimensionality of the output determines the number of neurons in the output layer. The input parameter is 22 dimensions, so the number of neurons in the input layer is set to 22; the output parameter is 3 dimensions, so the number of neurons in the output layer is set to 3. In the following experiment, 70% of the data will be used to train the model and the remaining 30% will be used to test the performance of the model. The entire simulation process in this experiment was carried out using Python 3.6. Firstly, the activation function was determined by the experiment [30]. The number of neurons in the hidden layer was tentatively set to 22 to test the model’s prediction accuracy with different activation functions. The experimental results are shown in Table 4. The mean square error of the test set and the mean absolute percentage error of the test set were minimized when the activation function was “sigmoid.” The mean squared error of the training set for the “sigmoid” activation function is 316.72, the mean squared error of the test set is 594.99, and the total mean squared error of the prediction is 401.68. The mean squared error of the training set for the “sigmoid” activation function and the mean absolute percentage error for the training set are 8.02, the mean absolute percentage error for the test set is 10.45, and the total mean absolute percentage error is 8.76. The ELM algorithm activation function is set to “sigmoid.”

Next, experiments are used to determine the optimal number of hidden layer neurons for the ELM algorithm. The number of hidden layer neurons has a large impact on the prediction results of the ELM model, and too many or too few hidden layer neurons will decrease the model’s prediction accuracy. In this section, the ELM activation function is “sigmoid” and the other parameters remain unchanged. The number of ELM hidden layer neurons was determined to be in the range [1, 38] based on the dimensionality of the input parameters used in this experiment. The performance of the ELM model with different numbers of hidden layer neurons was compared, and the results are shown in Figure 6. When the number of neurons in the hidden layer was less than 23, for tensile strength, yield strength, and elongation, the prediction accuracy of the test set increased with the increase of the number of neurons. When the number of neurons in the hidden layer exceeds 23, especially the prediction accuracy of elongation, the prediction accuracy of the test set shows a trend of fluctuation decline. The reason for the previously mentioned phenomenon is that with the increase of the number of hidden layer neurons, the overfitting tendency of the model increases and the overfitting leads to the abnormal increase of the prediction error of the test set. As shown in Figure 6, the prediction accuracy reached its peak when the number of hidden layer neurons was 23.

Next, the population size of the grey wolf algorithm was determined by experiments. The number of iterations was set to 300, and the number of individuals in the population ranged from [10, 100]. The prediction results of the test set with the number of iterations are shown in Figure 7. The prediction results of the test set showed low prediction errors when the population size was 67, 78, and 86. Still, too large population would cause the model training time to be extended, resulting in unnecessary time wastage.

The choice of objective function has a massive impact on the performance of the GWO algorithm. When the model is a multioutput algorithm, the influence of the weights between different objectives on the prediction accuracy of the final algorithm is of interest. To investigate the effect of the objective function weights on the mechanical properties of aluminium alloys, the objective function used in this paper is shown in equation (24). Here, , , and are the predicted values of tensile strength, yield strength, and elongation, respectively; are the measured values of tensile strength, yield strength, and elongation, respectively; N is the number of measured values; is the penalty coefficient of yield strength and elongation, respectively. The number of iterations of the GWO algorithm is set to 100, and the trends of tensile strength, yield strength, and elongation with the change of penalty coefficients are shown in Figure 8. As an example, Figure 8(a) shows the relationship between the MAPE error of tensile strength and the penalty coefficient. The closer the color was to dark, the smaller the MAPE error and the better the prediction accuracy. Different combinations of the penalty coefficient correspond to different prediction accuracies of mechanical properties. When the combination of penalty coefficients (i, j) is (20, 80), (90, 50), (60, 30), and (90, 80), the prediction error of tensile strength is small. Similarly, after comparing Figures 8(b) and 8(c), the penalty coefficient corresponding to the minimum prediction error of yield strength and elongation can be obtained. The prediction errors of yield strength and elongation of tensile strength are summarized. When the penalty coefficient (i, j) is (20, 80), the prediction errors of three mechanical properties are smaller at the same time.

The trend of variance of MAPE for the three output variables is shown in Figure 9, which indicates that when the penalty coefficients are 80 and 20, respectively, the corresponding MAPE variances for tensile strength, yield strength, and elongation are smaller. When the penalty coefficients are 80 and 20, respectively, the predicted values of tensile strength, yield strength, and elongation can maintain a balance of the predicted values of the three mechanical properties while ensuring a small accuracy of their own. The penalty coefficients in the objective function equation (23) are chosen as 80, 20. The comparison between the optimal penalty coefficient and the original coefficient is shown in Table 5. The optimal penalty coefficient can not only reduce the MAPE of the test set but also significantly reduce the variance of the prediction errors of the three variables. A lower variance means that the prediction error distribution of the three variables is more balanced.

The number of iterations of the GWO algorithm significantly impacts the prediction results, and each iteration will obtain a better value of the ELM algorithm parameters than the previous iteration. The GWO algorithm population size was set to 30, the fitness function was set to equation (24), and i = 20, j = 80. The relationship between the prediction accuracy of the GWO-ELM model test set and the number of iterations, as well as the model training time and the number of iterations, is shown in Figure 10. As the iterations proceeded, the R2 of the test set increases gradually and the modelling time increased linearly with the number of iterations. When the number of iterations reaches 3150, the prediction accuracy of tensile strength, yield strength, and elongation converges.

In summary, the parameters of the GWO-ELM algorithm have been determined and are shown in Table 6. When the values of parameters are shown in Table 6, the iterative optimization of weight ω and bias b by the GWO algorithm can be realized. Thus, when the maximum number of iterations is reached, the optimal input weight ω and bias b of the ELM network in the GWO-ELM model are determined.

4.3. Experimentation and Evaluation of Model Performance

In this section, mechanical property prediction experiments demonstrate the excellent prediction accuracy of the proposed GWO-ELM model. The model comparison experiment of the GWO-ELM model, MGWO-ELM model, and traditional ELM model is designed to verify the advantages of the proposed GWO-ELM model. Each model was run through twenty replicates, and the results were averaged to avoid misleading results from model chance.

To ensure a fair comparison, the corresponding parameters for the GWO-ELM model and the ELM model in this experiment are shown in Table 6. The parameters used by the MGWO algorithm are consistent with the GWO algorithm. MAPE and RMSE were error measures for the mechanical properties’ prediction models. The prediction errors of the three models after twenty replicate experiments for the three mechanical properties are shown in Table 7. Table 7 provides an excellent visual representation of the prediction results of the GWO-ELM model, MGWO-ELM model, and ELM model for the three mechanical properties. In the test set, RMSE errors of tensile strength of the GWO-ELM model, MGWO-ELM model, and ELM model are 5.365, 11.557, and 21.489, respectively. The RMSE errors of yield strength in the GWO-ELM model, MGWO-ELM model, and ELM model are 11.881, 20.465, and 43.313, respectively. The RMSE errors of the elongation in the GWO-ELM model, MGWO-ELM model, and ELM model are 1.268, 1.848, and 5.360, respectively. The abovementioned result indicates that the GWO-ELM model mentioned in this paper has a significantly higher prediction accuracy than the conventional ELM model for the three mechanical properties of tensile strength, yield strength, and elongation. The MGWO-ELM model outperforms the ELM model but is weaker than the GWO-ELM model. In the test set, for tensile strength, yield strength, and elongation, the MAPE of the GWO-ELM model is 1.703, 2.703, and 5.994 respectively, while the corresponding MAPE of the MGWO-ELM model is 2.316, 5.185, and 9.244 respectively, and that of the ELM model is 5.317, 16.165, and 19.630, respectively. The abovementioned result indicates that the average accuracy of the GWO-ELM model is significantly higher than that of the MGWO-ELM model for the three mechanical properties of tensile strength, yield strength, and elongation. R2 was used as the evaluation index to evaluate the prediction results of the test set. The R2 values of tensile strength in GWO-ELM, MGWO-ELM, and ELM models were 0.9857, 0.9805, and 0.7102, respectively. The R2 values of yield strength in GWO-ELM, MGWO-ELM, and ELM models were 0.9662, 0.9804, and 0.3958, respectively. The R2 values of elongation in GWO-ELM, MGWO-ELM, and ELM models were 0.9141, 0.9074, and −0.8312, respectively. The abovementioned results show that GWO-ELM and MGWO-ELM achieve an ideal fitting effect in the prediction of tensile strength, yield strength, and elongation, and the performance of the two models is far better than that of the ELM model. In addition, GWO-ELM was better than MGWO-ELM in the prediction of tensile strength and elongation, so the GWO-ELM model was slightly better than MGWO-ELM.

The predicted results of the GWO-ELM model, MGWO-ELM model, and ELM model compared to the actual values after twenty replicate experiments for the three mechanical properties are shown in Figure 11. Both in the training set and the test set, the prediction accuracy of the yield strength and elongation of the tensile strength is higher in the GWO-ELM model and the fit degree of the yield strength and elongation of the tensile strength is higher in the GWO-ELM model. The error distributions of the three model test sets are shown in Table 8. For the three mechanical properties of tensile strength, yield strength, and elongation, the GWO-ELM model has a more obvious advantage in the prediction results of the test set. This is especially true for the tensile and yield strengths. When the prediction deviation is controlled within ±10%, the prediction accuracy of tensile strength in the GWO-ELM model, MGWO-ELM model, and ELM model is 100%, 97.5%, and 80%, respectively. The prediction accuracy of yield strength of the GWO-ELM model, MGWO-ELM model, and ELM model is 97.5%, 92.5%, and 62.5%, respectively. The prediction accuracy of elongation in the GWO-ELM model, MGWO-ELM model, and ELM model is 77.5%, 77.5%, and 37.5%, respectively. Although the prediction accuracy of elongation of GWO-ELM and MGWO-ELM is the same when the error is controlled within ±10%, the prediction accuracy of the GWO-ELM model is better than that of MGWO-ELM when the prediction error is 5%. The abovementioned results indicate that the prediction accuracy of GWO-ELM is more stable than MGWO-ELM and ELM models.

In the prediction results of the test set, the absolute error distribution of the prediction results of tensile strength, yield strength, and elongation in different models is shown in Figure 12. As shown in Figure 12(a), the prediction error range of tensile strength in the GWO-ELM model is smaller than that of the other two models. As shown in Figure 12(b), the prediction error range of yield strength on the GWO-ELM model is smaller than that of the other two models. As shown in Figure 12(b), the prediction error range of yield strength in the GWO-ELM model is significantly smaller than that of the other two models. The GWO-ELM model performs better than the MGWO-ELM model in predicting tensile strength and yield strength but not in predicting elongation. As shown in Figure 12(c), the MGWO-ELM model is slightly better than the GWO-ELM model in the prediction of elongation. In general, compared with the other two models, the prediction accuracy of tensile strength, yield strength, and elongation can be higher in the GWO-ELM model.

5. Conclusion

This paper proposes an intelligent optimization algorithm, which has the advantages of a simple structure, short training time, and high prediction accuracy. This algorithm is successfully applied to the prediction of mechanical properties of aluminium alloy strips in this paper and eventually achieves high prediction accuracy, providing a new method for predicting the mechanical properties of aluminium alloy strips, which has broad application prospects.(1)Preprocessed industrial production data were used to train and test the GWO-ELM model. Through several experiments, the optimal network structure of the ELM algorithm and the optimal parameters of the Grey Wolf algorithm were determined. The weights and biases of the ELM were determined by the GWO algorithm. The optimal GWO-ELM model structure was successfully determined.(2)After repeated training, the RMSEs of the GWO-ELM model for tensile strength, yield strength, and elongation were 5.365, 11.881, and 1.268, respectively. While the corresponding RMSEs of the MGWO-ELM model were 11.557, 20.465, and 1.845, and the corresponding RMSEs of the ELM model were 21.489 and 43.313. This indicates that the ELM model optimized by the GWO algorithm has higher prediction accuracy.(3)When the prediction deviation is controlled within ±10%, the prediction accuracy of tensile strength in the GWO-ELM model, MGWO-ELM model, and ELM model is 100%, 97.5%, and 80%, respectively. The prediction accuracy of yield strength of the GWO-ELM model, MGWO-ELM model, and ELM model is 97.5%, 92.5%, and 62.5%, respectively. The prediction accuracy of elongation in the GWO-ELM model, MGWO-ELM model, and ELM model is 77.5%, 77.5%, and 37.5%, respectively. The comprehensive performance of the GWO-ELM model in predicting tensile strength, yield strength, and elongation is better than that of the other two models. The GWO algorithm successfully improves the stability and robustness of the ELM model.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

All authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the Major Project of Science and Technology in Nanning (No. 20191002).