Abstract
Polyethylene as a thermoplastic has received the uppermost popularity in a vast variety of applied contexts. Polyethylene is produced by several commercially obtainable technologies. Since Ziegler–Natta catalysts generate polyolefin with broad molecular weight and copolymer composition distributions, this type of model was utilized to simulate the polymerization procedure. The EIX (ethylene index) is the critical controlling variable that indicates product characteristics. Since it is difficult to measure the EIX, estimation is a problem causing the greatest challenges in the applicability of production. To resolve such problems, ANNs (artificial neural networks) are utilized in the present paper to predict the EIX from some simply computed variables of the system. In fact, the EIX is calculated as a function of pressure, ethylene flow, hydrogen flow, 1-butane flow, catalyst flow, and TEA (triethylaluminium) flow. The estimation was accomplished via the Multi-Layer Perceptron, Radial Basis, Cascade Feed-forward, and Generalized Regression Neural Networks. According to the results, the superior performance of the Multi-Layer Perceptron model than other ANN models was clearly demonstrated. Based on our findings, this model can predict production levels with R2 (regression coefficient), MSE (mean square error), AARD% (average absolute relative deviation percent), and RMSE (root mean square error) of, respectively, 0.89413, 0.02217, 0.4213, and 0.1489.
Keywords: polyethylene, ethylene index, intelligence approaches, prediction
1. Introduction
Polyethylene as a thermoplastic has received the uppermost popularity in a vast variety of applied contexts. In spite of the need for considerable financial investment for producing polyethylene, the end consumer might receive a very inexpensive and throwaway product. Therefore, improvements in polymer fabrication procedures aiming at reducing manufacturing expenses remain as an investigational, developmental, and process expansion topic [1].
High-quality, low-cost petrochemical products are increasingly in demand for many applications including modernistic HDPE (high-density polyethylene) marketplaces [2]. Polyethylene is manufactured using common process technologies including high-pressure autoclave, high-pressure tubular, slurry (suspension), gas phase, and solution. The gas phase process in particular is one of the typical and frequently used technologies in the production of HDPE and other polyolefins. Various reactors are applied in ethylene polymerizations ranging from simple autoclaves and steel piping to CSTR (continuously stirred tank reactors) and vertical fluidized beds [3]. The main incentives for this innovation include the elimination of the necessity for removing the catalyst following the reaction and making the product in a style appropriate for manipulation and reposition. In the gas phase reactors, polymerization occurs at the surface of the catalyst and the polymer matrix, which is inflated with monomers throughout the polymerization [4].
Although catalysts for polymerizing ethylene are often heterogeneous, homogeneous catalysts are also used in some processes. Currently, four kinds of catalysts exist for polymerizing ethylene: Ziegler–Natta, Phillips, metallocenes, and late transition metal catalysts [2]. Only the first three types have commercial applications while the last one is yet to undergo examination and a developing phase. Ziegler–Natta catalysts vary greatly but are usually TiCl4 supported on MgCl2 and commonly employed for the polyethylene production in the industry [5].
In HDPE production practices, the ethylene index (EIX) is the critical controlling variable that indicates product characteristics. It is necessary to preserve the uniform features of HDPE throughout grade change practices to fulfill the different and strict requirements for HDPE products, including the EIX.
There are several approaches regarding polyethylene production estimation and correlation in the literature. Khare et al. [6] introduced steady-state and dynamic models for the HDPE slurry polymerization procedure for industrial applications. The authors utilized parallel reactors data to model the reactors in series order. They set the kinetic parameters by initial consideration of a single site catalyst followed by using the outcomes to optimize the kinetic parameters for multi-site-type catalyst. Neto et al. [7] designed a dynamic model for the LLDPE polymerization process on the basis of a multi-site catalyst that underwent development merely for a single slurry reactor.
Considering a myriad of studies for improving polyethylene productivity, modeling approaches stand superior to experiments due to several key factors, including time and running costs, test feasibility, and safety (without hazards) concerns [8,9,10,11,12]. As denoted in previous publications, modeling approaches comprise of regression models, mathematical models, and artificial intelligence models. Nevertheless, the accurate function of mathematical models is dubious, in particular, once these models deal with very uncertain polyethylene manufacture. Regression models, however, can still well predict the means of production systems per month. As it is necessary to accurately model the productive status of such systems, artificial neural network (ANN) models have been used to achieve this objective due to their ability to handle great uncertainty of such data. ANNs integrate industrial data for predicting the rate of production, but only a few investigations have been successful in utilizing the ANN models for such systems by integrating a diverse range of techniques such as the MLPNN (multi-layer perceptron), CFNN (cascaded forward), RBFNN (radial basis function), and GRNN (general regression) neural networks [13,14], among others.
MLP is a neural model with the highest prediction applicability, consisting of multiple layers, namely input, hidden layer(s), and output. Hidden layer(s) can also be set more precisely as having layers of nodes [15]. The MLPNN as the prime and plainest topology of the ANN, employed for updating the weight links through the learning step.
RBFNN is a common feedforward neural network, which has been demonstrated to be capable of global estimation without local minima problem. Besides, it possesses a plain construct and a rapid learning algorithm in comparison with other neural networks [16]. Despite the availability of multiple activation functions for radial basis neurons, the Gaussian function has received the uppermost popularity. Training is necessary before applying an RBF ANN, which is commonly achievable in two stages. Choosing centers among the data applied is in the first stage of the training. The second stage is to utilize the normal least squares for linear estimation of one weighting vector. The self-organized center election is the widespread learning approach employed for selecting RBF centers.
The input and output neurons begin the production of the CFNN topology. The output neurons are present in the neural network in advance; thus, novel neurons are provided for the network, as a result of which the network, in turn, attempts to increase the correlation level between the outputs and inputs by comparing the network residue with the novel measured error. This procedure goes ahead until reaching a smaller error value in the network, which explains the reasoning that it is labeled as a cascade [17]. CFNN generally comprises three major layers namely, input, hidden, and output layers. The variables in hidden layers are multiplied by the bias (1.0) and the weight (computed in the creation phase to decline the prediction error) followed by addition to the sum entering the neuron. The resultant value from this procedure will cross a transfer function to present the output value.
GRNN is a type of supervised network that works on the basis of the probabilistic model and is able to produce continuous outputs. It is a robust instrument for non-linear regression analysis based on the approximation of probability density functions using the Parzen window technique [18]. The GRNN architecture basically does not need an iterative process to simulate such results as back-propagation learning algorithms. GRNN is capable of estimating arbitrary functions among output and input datasets directly from training data.
The main uses of the RBF and GRNN topologies are with a rather small size of input data. Despite the common topology of neural networks, every neuron depends on the entire prior layer neurons in the CFNN. Additionally, the CFNN is able to carry on to a broad extent in case the input data possess a sizable memory capacity.
As revealed by a literature review, the ANN model has applications in predicting the performance of production rate. Nonetheless, only the MLP model was utilized to forecast the production rate [19]. The novelty of the present study is to utilize and compare the performance of several models including the GRNN and RBF models, which were not previously used for predicting the production rate. Accordingly, the current research mainly aims to introduce and assess a model for predicting the rate of polyethylene fabrication. The novelty of our model is characterized by its capability in predicting system productivity by taking the uncertainty issue into account.
2. Methods
2.1. HDPE Process
The HDPE plant comprises two procedures, through which the polymerization reaction is put into action. Figure 1 illustrates a representative gas phase polymerization procedure for producing HDPE introduced by this research. Every process involves two polymerization reactors. The polymerization reaction was very exothermal, with a heat reaction of around 1000 kcal/Kg of ethylene.
Figure 1.
A diagram of slurry polymerization for HDPE production [20].
There is, therefore, a need to provide suitable cooling systems that eliminate nearby 80% of heat in the polymerization process. Co-monomer (including 1-butene or higher alpha-olefin), ethylene, an activator, hydrogen, hexane, and a catalyst, as well as continually recycled original liquid, are supplied to reactors as reactants. Generally, the slurry phase occupies almost 90–95% of reactor volume. With building up of the reaction pressure, the polyethylene slurry is transferred to the next process apparatus and the reactor level is preserved within an allowable range. Separating the reaction slurry in the centrifugal separator yields cake, holding dilutants, after which dilutants are removed with hot nitrogen gas in a dryer. Thereafter, suitable additives are added depending on the final usages. After pelletizing in water, pellets are dried and placed in a homogenizer followed by cooling.
2.2. Artificial Neural Networks
The ANN is an AI (artificial intelligence) method, defined as the information processing model, which is exhilarated by the human nervous systems for information processing [21,22,23]. The ANN is capable of identifying patterns and learning from their interplays with the environment [24,25,26,27,28]. An ANN is constructed through three major fractions of the input, the output, and the hidden layer(s), all of which comprise parallel units named neurons [29]. The neurons are coupled with massive weight links, allowing the information to be transformed among the layers. The ANN model is essentially dependent upon two key steps for predicting the response of different systems: the training phase and the testing phase. In the training phase, the inputs are received by the neurons over their entering connections, after which these inputs are combined by a specific action with the output to discover the best the weight links values. Therefore, records of the association between output and input variables are taken to forecast the fresh data. In the testing phase, the system performance is tested using a portion of the input data and a comparison is made between the predicted data and the real data. The principal benefit of the ANN model is its ability in solving sophisticated problems that cannot be easily solved by traditional models; it is also capable of solving problems without an algorithmic solution or those with algorithmic solutions having sophisticated definitions.
The independent variables from outside resources in the input layer are processed by several mathematical operators and by sending values to the hidden layers. On the other hand, output neuron(s) are used to determine the dependent variables. The output value for all ANN structures can be defined as below:
(1) |
where this value is computed by the activation function of . Here, is the summation of the bias value, and is the weight of input. Figure 2 presents the schematic diagram of the proposed neural network for simulating the EIX.
Figure 2.
Schematic representation of the used model for neural networks [30].
2.3. Accuracy Assessment of AI Models
The current study has designed various AI-based approaches with diverse topologies to opt for the best model based on the accuracy of predictability. This can serve as a selection paradigm for network configuration to determine the number of hidden layers and neurons, the spread factor, and training algorithm. It is noted that the dependencies of the neurons number in the hidden layer on prediction intervals were also examined. The performance of ANN models using R2 (regression coefficient), MSE (mean square error), AARD% (average absolute relative deviation percent), and RMSE (root mean square error) are calculated, respectively, as the following:
(2) |
(3) |
(4) |
(5) |
where, is the actual and is predicted value. N and also denote the number of data points and the mean of actual values, respectively.
To evaluate mean squared error (MSE) for various parameters, the predicted data are also subject to statistical analysis. MSE measures the absolute deviation of the predicted and the actual values. Positive and negative values denote overestimation and underestimation of parameters, respectively. Based on the aforesaid description, the RMSE denotes that the model is efficient on the basis of the difference between the predicted and real data. Accordingly, a large positive RMSE indicates the presence of a high deviation between the predicted and real data and in a reverse order. The R2 index defines the proximity of the actual data points to the predicted values.
3. Results and Discussion
This section summarizes the actual databank gathered from the industrial polyethylene petrochemical company, by considering the significant independent variables and by utilizing the Pearson correlation matrix. Furthermore, this section deals with determining the best structures of different models and comparing the precisions of different models. The present section concludes by selecting the best model and analyzing the results.
3.1. Industrial Database
To investigate the EIX, eleven independent sets of input data namely, temperature, operating pressure, level, loop flow, ethylene flow, hydrogen flow, 1-butane flow, hydrogen concentration, 1-butane concentration, catalyst flow, and TEA (triethylaluminium) flow (i.e., inputs 1 to 11 denoted, respectively, by X1 to X11), and the EIX response (denoted by Y) are collected. Table 1 presents the information summary of the industrial data used in this work.
Table 1.
Ranges and of the Industrial inputs.
Independent Variables | Named | Minimum | Maximum |
---|---|---|---|
Temperature (°C) | Input 1 | 77 | 97.8 |
Pressure (bar) | Input 2 | 21.8 | 22.6 |
Level (%) | Input 3 | 68 | 77 |
Loop flow (Kg/h) | Input 4 | 646 | 724 |
Ethylene flow (Kg/h) | Input 5 | 1.8 | 15.6 |
Hydrogen flow (Kg/h) | Input 6 | 6.1 | 25 |
1-butane flow (Kg/h) | Input 7 | 17 | 970 |
Hydrogen concentration (mol%) | Input 8 | 16.02 | 19.03 |
1-butane concentration (mol%) | Input 9 | 0.49 | 1.38 |
Catalyst flow (Kg/h) | Input 10 | 2 | 3.7 |
TEA flow (Kg/h) | Input 11 | 1.8 | 3.7 |
According to the industrial databank, 93 data points were gathered at the steady-state conditions. The trained neural network needs validation for determining the precision of the introduced model. The network performance can be analyzed by cross-validation of an unidentified dataset. The network is validated through preserving a fraction of the dataset (e.g., 15%) for validation and the rest of the dataset is used for training. After the training phase, the data forecasted through the ANN topology and the measured data undergo a correlation analysis.
All ANN models were developed in MATLAB® with the Levenberg–Marquardt optimization algorithm. Besides, the choice of training algorithm and neuron transfer function has a major contribution to model precision. As researchers have shown, the Levenberg–Marquardt (LM) algorithm produces quicker responses for regression-type problems in overall facets of neural networks [31,32]. Most often, the LM training algorithm was reported to have the highest significant efficiency, fast convergence, and accuracy compared with other training algorithms.
3.2. Scaling the Data
To enhance the rate of convergence in the training step as well as to avoid parameter saturation in the intended ANNs, the entire actual data were subjected to mapping within the interval [0.01 0.99]. Data were normalized by Equation (6):
(6) |
where denotes an independent or dependent variable, represents the normal value, is the maximum, and is the minimum value of each variable.
3.3. Independent Variable Selection
Mathematical investigation for the dependency of two variables is possible through the correlation matrix study, the coefficients of which are usually measured from −1 to +1. This indicates that the two variables are correlated directly or indirectly given the signs of these coefficients, whereas the magnitude defines the robustness of their association. Our research surveyed a multivariate AI-based method with the Pearson correlation test for estimating the degree of relationship between each two variables [23]. Figure 3 displays the correlation coefficient values’ given probable pairs of variables.
Figure 3.
Pearson correlation coefficient values among independent and dependent variables.
The Pearson correlation coefficient as the variable ranking is described in choosing appropriate inputs for the neural network [33,34]. Values delivered by the Pearson method reveal the type and intensity of the association between every variable pair, with a value between −1 and +1 representing the uppermost converse relationship and the highest direct association, respectively. The coefficient takes a zero value in cases where the given variables do not have any association. Independent variables take non-zero correlation coefficients that verify their choices. The highest consideration is devoted to absolute average values because they have important strong associations. The values of Pearson correlation coefficient for each input are presented in Table 2.
Table 2.
The Pearson correlation coefficient for each input.
Independent Variables | Pearson Correlation Coefficient |
---|---|
Input1 | −0.0430 |
Input2 | 0.2452 |
Input3 | 0.0581 |
Input4 | 0.0252 |
Input5 | 0.1606 |
Input6 | 0.0956 |
Input7 | −0.1160 |
Input8 | 0.0052 |
Input9 | −0.0296 |
Input10 | 0.0935 |
Input11 | 0.1075 |
Accordingly, this examination confirmed that Input2, Input5, Input6, Input7, Input10, and Input11 had the uppermost direct dependency and that other inputs presented the lowermost indirect association. Hence, it is possible to model polyethylene as a function of pressure, ethylene flow, hydrogen flow, 1-butane, catalyst flow, and TEA flow. Therefore, we try to present a smart model to derive the following relation:
(7) |
Consequently, in the case of maximizing the AAPC (average of absolute Pearson’s coefficient) for a specific transformation on the dependent variable, it is inferred that it yields an association between the dependent and independent variables with the highest reliability. Every input variable of differing models was selected with the Pearson correlation coefficient. The inputs of every model are presented in Table 3.
Table 3.
Pearson coefficient values calculated for different transformations between dependent and independent variables.
Transformation | Pearson’s Coefficient | AAPC | |||||
---|---|---|---|---|---|---|---|
Input11 | Input10 | Input7 | Input6 | Input5 | Input2 | ||
Output15 | 0.09837 | 0.08478 | −0.12881 | 0.0877 | 0.17071 | 0.28188 | 0.14204 |
Output14 | 0.09943 | 0.08576 | −0.12845 | 0.08885 | 0.17077 | 0.27997 | 0.1422 |
Output13 | 0.10045 | 0.0867 | −0.12801 | 0.08992 | 0.17073 | 0.27795 | 0.14229 |
Output12 | 0.10143 | 0.0876 | −0.12750 | 0.09092 | 0.17058 | 0.27584 | 0.14231 |
Output11 | 0.10235 | 0.08845 | −0.12690 | 0.09184 | 0.17033 | 0.27361 | 0.14225 |
Output10 | 0.10321 | 0.08925 | −0.12623 | 0.09268 | 0.16996 | 0.27128 | 0.1421 |
Output2 | 0.1074 | 0.09331 | −0.11754 | 0.09572 | 0.16225 | 0.24851 | 0.13745 |
Output | 0.10751 | 0.09346 | −0.11603 | 0.09556 | 0.16063 | 0.24516 | 0.13639 |
Output0.75 | 0.10753 | 0.09349 | −0.11563 | 0.0955 | 0.1602 | 0.24431 | 0.13611 |
Output0.5 | 0.10753 | 0.0935 | −0.11523 | 0.09543 | 0.15976 | 0.24344 | 0.13582 |
Output0.25 | 0.10753 | 0.09351 | −0.11483 | 0.09535 | 0.15932 | 0.24258 | 0.13552 |
Output0.1 | 0.10753 | 0.09352 | −0.11458 | 0.0953 | 0.15904 | 0.24205 | 0.13534 |
Output−0.1 | −0.10752 | −0.09352 | 0.11425 | −0.09523 | −0.15867 | −0.24135 | 0.13509 |
Output−0.25 | −0.10751 | −0.09352 | 0.114 | −0.09517 | −0.15839 | −0.24082 | 0.1349 |
Output−0.5 | −0.10749 | −0.09351 | 0.11357 | −0.09507 | −0.15791 | −0.23994 | 0.13458 |
Output−0.75 | −0.10747 | −0.09350 | 0.11314 | −0.09496 | −0.15743 | −0.23904 | 0.13426 |
Output−1 | −0.10743 | −0.09348 | 0.11271 | −0.09484 | −0.15693 | −0.23814 | 0.13392 |
Output−2 | −0.10723 | −0.09335 | 0.11091 | −0.09427 | −0.15485 | −0.23449 | 0.13252 |
Output−10 | −0.10154 | −0.08867 | 0.09353 | −0.08469 | −0.13296 | −0.20259 | 0.11733 |
Output−11 | −0.10034 | −0.08765 | 0.09106 | −0.08289 | −0.12965 | −0.19840 | 0.115 |
Output−12 | −0.09903 | −0.08653 | 0.08855 | −0.08097 | −0.12624 | −0.19420 | 0.11259 |
Output−13 | −0.09763 | −0.08533 | 0.08599 | −0.07894 | −0.12273 | −0.19000 | 0.11011 |
Output−14 | −0.09614 | −0.08405 | 0.08341 | −0.07680 | −0.11914 | −0.18581 | 0.10756 |
Output−15 | −0.09456 | −0.08269 | 0.0808 | −0.07457 | −0.11548 | −0.18165 | 0.10496 |
exp(Output) | 0.08336 | 0.07114 | −0.12809 | 0.07017 | 0.16382 | 0.29757 | 0.13569 |
From this table, it can be concluded that it is better to output to the power of 12 instead of modeling the output itself. Although, at last, by inverse transformation, the dependent variable is calculated to compare with the actual values.
3.4. Configuration Selection for Different ANN Approaches
In this work, the EIX is predicted using a proper ANN model obtained from a logical procedure. As mentioned earlier, the number of hidden neurons has a major contribution to network performance. The majority of related investigations obtain the number of neurons through the trial and error approach. Training and generalization errors may highly happen when the numbers of hidden neurons are less than the optimum numbers. On the other hand, larger numbers of hidden neurons may result in over-fitting and considerable variations. Therefore, it is necessary to calculate the optimum number of hidden neurons for achieving the best performance of the network.
Subsequently, the ANN approaches were developed and then, for example, the MLP network was compared in terms of performance with CF, RBF, and GR neural networks. The numerical validation associates with the observed AARD%, R2, MSE, and RMSE between actual and estimated data. According to the literature, MLP network capability with one hidden layer was proven [35]. As such, an MLP network with only a single hidden layer is used for the analysis.
It is noted that the training data points should be at least twice the number of bias and weights. As a result, for the MLP with one dependent and six independent variables, the hidden neuron is computed as:
(8) |
Therefore, this number can change from 1 to 4 (the highest acceptable number) in this network, and is trained 50 times for each network. The best configuration of the hidden neurons in the MLP model is presented in Table 4.
Table 4.
The procedure for detecting the best configuration for the MLP model.
Hidden Neuron | Dataset | Statistical Index | |||
---|---|---|---|---|---|
AARD% | MSE | RMSE | R2 | ||
1 | Train | 0.6566 | 0.07274 | 0.2697 | 0.6043 |
Test | 0.7578 | 0.06665 | 0.2582 | 0.35034 | |
Total | 0.6715 | 0.07184 | 0.268 | 0.57942 | |
2 | Train | 0.568 | 0.06603 | 0.257 | 0.66405 |
Test | 0.7314 | 0.05583 | 0.2363 | 0.65288 | |
Total | 0.592 | 0.06454 | 0.254 | 0.644 | |
3 | Train | 0.3863 | 0.01982 | 0.1408 | 0.91723 |
Test | 0.6251 | 0.03581 | 0.1892 | 0.53962 | |
Total | 0.4213 | 0.02217 | 0.1489 | 0.89413 | |
4 | Train | 0.3397 | 0.02037 | 0.1427 | 0.90878 |
Test | 0.8084 | 0.07617 | 0.276 | 0.44131 | |
Total | 0.4085 | 0.02856 | 0.169 | 0.86377 |
The MLP network with three hidden neurons and the structure of 6-3-1 was determined as the most appropriate model. The MSE values of the MLP network with various numbers of hidden layer neurons are presented in Figure 4. The data reveal that the optimality of the three numbers of neurons owes to the uppermost value of R2 (0.89413) and the lowermost value of MSE (0.02217).
Figure 4.
The MSE for the multi-layer perceptron (MLP) network with 1 to 4 hidden neurons (50 networks per neuron) during the training stage.
According to Figure 4, MSE is lowest (0.07184) in the total MLP model with the presence of a single neuron in the hidden layer. The MLP model possesses the least MSE (0.02217) once six neurons exist in the hidden layer. Moreover, Figure 5 presents a comparison between the industrial datasets and the predicted values by the optimum MLP network. The fit performance was determined for every trained MLP with minimum MSE value on the basis of R2 values.
Figure 5.
The efficiency of the optimum MLP model for prediction of all the datasets.
3.5. Other Types of ANN
To find an appropriate model for evaluating the EIX, different topologies of artificial neural networks must be compare based on their performances. Therefore, the intended MLP approach developed with optimum configuration in terms of predictive accuracy was evaluated with other ANN models (GR, CF, and RBF). The sensitivity results for selecting the best number of hidden neurons are presented in Table 5, Table 6 and Table 7. It should be pointed out that determination of the number of hidden neurons in the other ANNs was the same as for the MLP model.
Table 5.
Determination of the best-hidden neurons for cascaded forward (CF).
Hidden Neuron | Dataset | Statistical Index | |||
---|---|---|---|---|---|
AARD% | MSE | RMSE | R2 | ||
1 | Train | 0.5604 | 0.07049 | 0.2655 | 0.62338 |
Test | 0.6377 | 0.05431 | 0.233 | 0.58962 | |
Total | 0.5717 | 0.06812 | 0.261 | 0.60867 | |
2 | Train | 0.4961 | 0.05227 | 0.2286 | 0.74815 |
Test | 0.6904 | 0.06022 | 0.2454 | 0.36408 | |
Total | 0.5246 | 0.05344 | 0.2312 | 0.71356 | |
3 | Train | 0.488 | 0.03638 | 0.1907 | 0.83777 |
Test | 0.5729 | 0.0552 | 0.235 | 0.48687 | |
Total | 0.5004 | 0.03914 | 0.1979 | 0.79892 |
Table 6.
Determination of the best hidden neurons for radial bias function (RBF).
Hidden Neuron | Spread | Dataset | Statistical Index | |||
---|---|---|---|---|---|---|
AARD% | MSE | RMSE | R2 | |||
1 | 0.41 | Train | 0.762 | 0.10257 | 0.3203 | 0.30383 |
Test | 0.8574 | 0.06878 | 0.2623 | 0.28658 | ||
Total | 0.776 | 0.09762 | 0.3124 | 0.312 | ||
2 | 0.81 | Train | 0.7563 | 0.09887 | 0.3144 | 0.42263 |
Test | 0.762 | 0.05584 | 0.2363 | −0.10195 | ||
Total | 0.7571 | 0.09255 | 0.3042 | 0.38919 | ||
3 | 1.01 | Train | 0.7311 | 0.09259 | 0.3043 | 0.45482 |
Test | 0.774 | 0.05724 | 0.2392 | 0.33916 | ||
Total | 0.7374 | 0.08741 | 0.2956 | 0.43965 | ||
4 | 0.21 | Train | 0.6708 | 0.08324 | 0.2885 | 0.51622 |
Test | 0.7579 | 0.08244 | 0.2871 | 0.4649 | ||
Total | 0.6836 | 0.08312 | 0.2883 | 0.48421 |
Table 7.
Sensitivity analyses on spread parameter for finding best hidden neurons for general regression (GR).
Spread | Dataset | Statistical Index | |||
---|---|---|---|---|---|
AARD% | MSE | RMSE | R2 | ||
4.81 | Train | 0.8361 | 0.10982 | 0.3314 | 0.32248 |
Test | 0.9242 | 0.098 | 0.313 | −0.20907 | |
Total | 0.849 | 0.10808 | 0.3288 | 0.22754 |
In GR, hidden neurons were not significant and the spread value needed to be set up. Subsequently, the spread value for the GR changes from 0.1 to 10 with 0.1 steps and 50 different GRs were considered, with statistical indices. The MSE is minimum (0.10808) in the GR model when the spread value was 4.81.
Considering the task, the best model contains the lowest value for MSE and AARD%. Table 8 clearly reveals that the MLP model can predict the EIX more accurately than other types of ANN models. Based on the statistical error values, the MSE for the MLP model (0.02217) is less than those for the MSE estimated with CF (0.03914), GR (0.10808), and RBF (0.09255) models, respectively. The above findings confirm that the MLP model is superior in the prediction of the EIX comparing to other ANN models.
Table 8.
A performance comparison of ANN models.
Model | Dataset | Statistical Index | |||
---|---|---|---|---|---|
AARD% | MSE | RMSE | R2 | ||
MLP | Train | 0.3863 | 0.01982 | 0.1408 | 0.91723 |
Test | 0.6251 | 0.03581 | 0.1892 | 0.53962 | |
Total | 0.4213 | 0.02217 | 0.1489 | 0.89413 | |
CF | Train | 0.488 | 0.03638 | 0.1907 | 0.83777 |
Test | 0.5729 | 0.0552 | 0.235 | 0.48687 | |
Total | 0.5004 | 0.03914 | 0.1979 | 0.79892 | |
GR | Train | 0.8361 | 0.10982 | 0.3314 | 0.32248 |
Test | 0.9242 | 0.098 | 0.313 | −0.20907 | |
Total | 0.849 | 0.10808 | 0.3288 | 0.22754 | |
RBF | Train | 0.7563 | 0.09887 | 0.3144 | 0.42263 |
Test | 0.762 | 0.05584 | 0.2363 | −0.10195 | |
Total | 0.7571 | 0.09255 | 0.3042 | 0.38919 |
The MLP model, trained by the Levenberg–Marquardt algorithm with 6-3-1 structure, has the logsig transfer function in the output and hidden layers. In fact, this model is chosen from 600 models (200 MLPNN models, 150 CFNN models, 50 GRNN models, and 200 RBFNN models). This selection is based on four statistical indices: AARD%, MSE, RMSE, and R2.
Table 9 summarizes the value of the weight and bias for the proposed MLP model. The MLP was trained using a training dataset by the adjustment of the biases and weights. The performance validity of the trained MLP was achieved according to the training and testing datasets (independent datasets). The optimal division ratio is 85:15 for segregating the data.
Table 9.
Weight and bias coefficients for the proposed MLP model.
Weights between Hidden Layer Neurons and Input Variables | Bias of Hidden Layer Neurons | Weight between the Hidden Layer and the Output Layer | Output Layer Bias | |||||
---|---|---|---|---|---|---|---|---|
Input2 | Input5 | Input6 | Input7 | Input10 | Input11 | |||
37.522 | 325.0769 | −1007.47 | −3.6761 | −330.513 | −168.349 | 176.6308 | −0.94888 | 0.69286 |
272.4066 | −70.6262 | −94.8238 | 4.1889 | −150.163 | 2.1714 | −16.8291 | 2.7494 | |
−1095.55 | 1458.018 | 70.1148 | 370.1613 | −539.429 | −12.1413 | 9.7747 | −2.276 |
4. Procedure for Simple Usage of the MLP Model
All independent variables normalize into an interval of [0.01 0.99] using Equation (9) and should be arranged as a 6×1 vector.
(9) |
-
2.
Multiply the first six columns of Table 9 by the variables achieved in step 1.
-
3.
The 7th column of Table 9 is added to the obtained values in step 2.
-
4.
Substitute all elements of step 3 in the following equation to calculate NOHL.
NOHL = 1/[1 + exp(−values obtained in step 3)] | (10) |
-
5.
Multiply the transposition of the 8th column of Table 9 by the obtained values in step 4.
-
6.
Add the value of the last column of Table 9, i.e., 0.69286, to the obtained values in step 5.
-
7.
Substitute the obtained value in step 6 in the following equation to calculate NOOL.
NOOL = 1/[1 + exp(−value obtained in step 6)] | (11) |
-
8.
Inverse transformation using NOOL ^ (1/12).
-
9.
Map the output values in the previous step into the actual range of dependent variables, i.e., [24.3 26.9], using the following equation.
Predicted variable = (Step9 − 0.01) × 2.6531 + 24.3 | (12) |
-
10.
The obtained value in step 9 shows the estimated value for the dependent variable by the proposed MLP approach.
Finally, the proposed MLP model was statistically checked for reliability, so the leverage approach was used. The following figure shows the outlier detection based on the MLP model using the Williams Plot method. It is clear from Figure 6 that only 4 data points (red dots) of the 75 data available are difficult to model. In fact, about 95% of the data is in the valid range (blue squares).
Figure 6.
Williams plot for the MLP model.
5. Conclusions
In HDPE procedures, the EIX is the critical controlling variable indicative of product quality. A large number of approaches are available for the estimation and correlation of EIX, since it is difficult to nonlinearly measure them. The current paper applied predicting methods to prediction schemes, including MLPNN (multi-layer perceptron), CFNN (cascaded forward), RBFNN (radial basis function), and GRNN (general regression) neural networks. Comparisons were made between the findings of different dynamic prediction schemes to assess the best performance. The superior performance of the present MLP model was demonstrated using the same case study dataset for predicting the EIX than other models. The results clearly suggest that three hidden neurons are the best number of neurons in the proposed MLP model. For this model, the MSE and R2 values of the total dataset are 0.02217 and 0.89413, respectively. The main advantages of using ANNs for the EIX are the ability to predict the production rate of the network quickly and to clarify the characteristics of high-density polyethylene with network inputs. Although these models have used complex computational algorithms, fast convergence along with accuracy is not always confirmed in some cases.
Nomenclature
b | Bias |
N | Number of actual data |
Xi | ith input variable |
Y | Response |
w | Weight |
Abbreviations
AARD% | Average absolute relative deviation percent |
AI | Artificial intelligence |
ANN | Artificial neural networks |
CFNN | Cascade feedforward neural networks |
CSTR | Continuously stirred tank reactors |
EIX | Ethylene index |
HDPE | High-density polyethylene |
MLP | Multi-layer perceptron |
MLPNN | Multi-layer perceptron neural networks |
MSE | Mean squared errors |
RBF | Radial basis neural networks |
RMSE | Root mean square errors |
R2 | Regression coefficient |
TEA | Triethylaluminium |
Subscripts/Superscripts
pre. | Predicted variable |
act. | Actual variable |
max | Maximum value |
min | Minimum value |
normal | Normalized values |
Author Contributions
Data curation, A.R.; Formal analysis, A.M.; Investigation, A.M.; Methodology, M.S.S.; Supervision, M.S.S.; Writing—original draft, A.R. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Nowlin T.E. Business and Technology of the Global Polyethylene Industry: An In-depth Look at the History, Technology, Catalysts, and Modern Commercial Manufacture of Polyethylene and Its Products. John Wiley & Sons; Hoboken, NJ, USA: 2014. [Google Scholar]
- 2.Malpass D.B. Handbook of Transition Metal Polymerization Catalysts. John Wiley & Sons; Hoboken, NJ, USA: 2018. Industrial Metal Alkyls and Their Use in Polyolefin Catalysts; pp. 1–30. [DOI] [Google Scholar]
- 3.Spalding M.A., Chatterjee A. Handbook of Industrial Polyethylene and Technology: Definitive Guide to Manufacturing, Properties, Processing, Applications and Markets Set. John Wiley & Sons; Hoboken, NJ, USA: 2017. [Google Scholar]
- 4.Patel R.M. Multilayer Flexible Packaging. Elsevier; Amsterdam, The Netherlands: 2016. Polyethylene; pp. 17–34. [Google Scholar]
- 5.Kissin Y.V., Nowlin T.E., Mink R.I. Handbook of Transition Metal Polymerization Catalysts. John Wiley & Sons; Hoboken, NJ, USA: 2018. Supported Titanium/Magnesium Ziegler Catalysts for the Production of Polyethylene; pp. 189–227. [DOI] [Google Scholar]
- 6.Khare N.P., Lucas B., Seavey K.C., Liu Y.A., Sirohi A., Ramanathan S., Lingard S., Song Y., Chen C.-C. Steady-State and Dynamic Modeling of Gas-Phase Polypropylene Processes Using Stirred-Bed Reactors. Ind. Eng. Chem. Res. 2004;43:884–900. doi: 10.1021/ie030714t. [DOI] [Google Scholar]
- 7.Neto A.G.M., Freitas M.F., Nele M., Pinto J.C. Modeling Ethylene/1-Butene Copolymerizations in Industrial Slurry Reactors. Ind. Eng. Chem. Res. 2005;44:2697–2715. doi: 10.1021/ie049588z. [DOI] [Google Scholar]
- 8.Shehzad N., Zeeshan A., Ellahi R., Vafai K. Convective heat transfer of nanofluid in a wavy channel: Buongiorno’s mathematical model. J. Mol. Liq. 2016;222:446–455. doi: 10.1016/j.molliq.2016.07.052. [DOI] [Google Scholar]
- 9.Ellahi R., Rahman S.U., Gulzar M.M., Nadeem S., Vafai K. A Mathematical Study of Non-Newtonian Micropolar Fluid in Arterial Blood Flow through Composite Stenosis. Appl. Math. Inf. Sci. 2014;8:1567–1573. doi: 10.12785/amis/080410. [DOI] [Google Scholar]
- 10.Nadeem S., Riaz A., Ellahi R., Akbar N.S. Mathematical model for the peristaltic flow of Jeffrey fluid with nanoparticles phenomenon through a rectangular duct. Appl. Nanosci. 2013;4:613–624. doi: 10.1007/s13204-013-0238-5. [DOI] [Google Scholar]
- 11.Hussain F., Ellahi R., Zeeshan A. Mathematical Models of Electro-Magnetohydrodynamic Multiphase Flows Synthesis with Nano-Sized Hafnium Particles. Appl. Sci. 2018;8:275. doi: 10.3390/app8020275. [DOI] [Google Scholar]
- 12.Bhatti M., Zeeshan A., Ellahi R., Shit G. Mathematical modeling of heat and mass transfer effects on MHD peristaltic propulsion of two-phase flow through a Darcy-Brinkman-Forchheimer porous medium. Adv. Powder Technol. 2018;29:1189–1197. doi: 10.1016/j.apt.2018.02.010. [DOI] [Google Scholar]
- 13.Liu W., Shadloo M.S., Tlili I., Maleki A., Bach Q.-V. The effect of alcohol–gasoline fuel blends on the engines’ performances and emissions. Fuel. 2020;276:117977. doi: 10.1016/j.fuel.2020.117977. [DOI] [Google Scholar]
- 14.Zheng Y., Shadloo M.S., Nasiri H., Maleki A., Karimipour A., Tlili I. Prediction of viscosity of biodiesel blends using various artificial model and comparison with empirical correlations. Renew. Energy. 2020;153:1296–1306. doi: 10.1016/j.renene.2020.02.087. [DOI] [Google Scholar]
- 15.Messikh N., Bousba S., Bougdah N. The use of a multilayer perceptron (MLP) for modelling the phenol removal by emulsion liquid membrane. J. Environ. Chem. Eng. 2017;5:3483–3489. doi: 10.1016/j.jece.2017.06.053. [DOI] [Google Scholar]
- 16.Zhao Z., Lou Y., Chen Y., Lin H., Li R., Yu G. Prediction of interfacial interactions related with membrane fouling in a membrane bioreactor based on radial basis function artificial neural network (ANN) Bioresour. Technol. 2019;282:262–268. doi: 10.1016/j.biortech.2019.03.044. [DOI] [PubMed] [Google Scholar]
- 17.Lashkarbolooki M., Vaferi B., Shariati A., Hezave A.Z. Investigating vapor–liquid equilibria of binary mixtures containing supercritical or near-critical carbon dioxide and a cyclic compound using cascade neural network. Fluid Phase Equilibria. 2013;343:24–29. doi: 10.1016/j.fluid.2013.01.012. [DOI] [Google Scholar]
- 18.Ghritlahre H.K., Prasad R.K. Exergetic performance prediction of solar air heater using MLP, GRNN and RBF models of artificial neural network technique. J. Environ. Manag. 2018;223:566–575. doi: 10.1016/j.jenvman.2018.06.033. [DOI] [PubMed] [Google Scholar]
- 19.Gonzaga J., Meleiro L.A.C., Kiang C., Filho R.M. ANN-based soft-sensor for real-time process monitoring and control of an industrial polymerization process. Comput. Chem. Eng. 2009;33:43–49. doi: 10.1016/j.compchemeng.2008.05.019. [DOI] [Google Scholar]
- 20.Allemeersch P. Polymerisation of Ethylene. Walter de Gruyter GmbH; Berlin, Germany: 2015. [Google Scholar]
- 21.Ghaffarian N., Eslamloueyan R., Vaferi B. Model identification for gas condensate reservoirs by using ANN method based on well test data. J. Pet. Sci. Eng. 2014;123:20–29. doi: 10.1016/j.petrol.2014.07.037. [DOI] [Google Scholar]
- 22.Aghel B., Rezaei A., Mohadesi M. Modeling and prediction of water quality parameters using a hybrid particle swarm optimization–neural fuzzy approach. Int. J. Environ. Sci. Technol. 2018;16:4823–4832. doi: 10.1007/s13762-018-1896-3. [DOI] [Google Scholar]
- 23.Maleki A., Elahi M., Assad M.E.H., Nazari M.A., Shadloo M.S., Nabipour N. Thermal conductivity modeling of nanofluids with ZnO particles by using approaches based on artificial neural network and MARS. J. Therm. Anal. Calorim. 2020:1–12. doi: 10.1007/s10973-020-09373-9. [DOI] [Google Scholar]
- 24.Amini Y., Gerdroodbary M.B., Pishvaie M.R., Moradi R., Monfared S.M. Optimal control of batch cooling crystallizers by using genetic algorithm. Case Stud. Therm. Eng. 2016;8:300–310. doi: 10.1016/j.csite.2016.09.001. [DOI] [Google Scholar]
- 25.Esfe M.H., Afrand M., Yan W.-M., Akbari M. Applicability of artificial neural network and nonlinear regression to predict thermal conductivity modeling of Al2O3–water nanofluids using experimental data. Int. Commun. Heat Mass Transf. 2015;66:246–249. doi: 10.1016/j.icheatmasstransfer.2015.06.002. [DOI] [Google Scholar]
- 26.Moayedi H., Aghel B., Vaferi B., Foong L.K., Bui D.T. The feasibility of Levenberg–Marquardt algorithm combined with imperialist competitive computational method predicting drag reduction in crude oil pipelines. J. Pet. Sci. Eng. 2020;185:106634. doi: 10.1016/j.petrol.2019.106634. [DOI] [Google Scholar]
- 27.Shadloo M.S., Rahmat A., Karimipour A., Wongwises S. Estimation of Pressure Drop of Two-Phase Flow in Horizontal Long Pipes Using Artificial Neural Networks. J. Energy Resour. Technol. 2020;142:1–21. doi: 10.1115/1.4047593. [DOI] [Google Scholar]
- 28.Komeilibirjandi A., Raffiee A.H., Maleki A., Nazari M.A., Shadloo M.S. Thermal conductivity prediction of nanofluids containing CuO nanoparticles by using correlation and artificial neural network. J. Therm. Anal. Calorim. 2019;139:2679–2689. doi: 10.1007/s10973-019-08838-w. [DOI] [Google Scholar]
- 29.Davoudi E., Vaferi B. Applying artificial neural networks for systematic estimation of degree of fouling in heat exchangers. Chem. Eng. Res. Des. 2018;130:138–153. doi: 10.1016/j.cherd.2017.12.017. [DOI] [Google Scholar]
- 30.Da Silva I.N., Spatti D.H., Flauzino R.A., Liboni L., Alves S.F.D.R. Artificial Neural Networks. Springer Science and Business Media; Berlin, Germany: 2016. Artificial Neural Network Architectures and Training Processes; pp. 21–28. [Google Scholar]
- 31.Kayri M. Predictive Abilities of Bayesian Regularization and Levenberg–Marquardt Algorithms in Artificial Neural Networks: A Comparative Empirical Study on Social Data. Math. Comput. Appl. 2016;21:20. doi: 10.3390/mca21020020. [DOI] [Google Scholar]
- 32.Zayani R., Bouallegue R., Roviras D. Levenberg-marquardt learning neural network for adaptive predistortion for time-varying HPA with memory in OFDM systems; Proceedings of the 16th European Signal Processing Conference; Lausanne, Switzerland. 25–29 August 2008; pp. 1–5. [Google Scholar]
- 33.Guyon I., Elisseeff A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003;3:1157–1182. [Google Scholar]
- 34.Wilby R., Abrahart R.J., Dawson C. Detection of conceptual model rainfall—Runoff processes inside an artificial neural network. Hydrol. Sci. J. 2003;48:163–181. doi: 10.1623/hysj.48.2.163.44699. [DOI] [Google Scholar]
- 35.Kamadinata J.O., Tan L.K., Suwa T. Sky image-based solar irradiance prediction methodologies using artificial neural networks. Renew. Energy. 2019;134:837–845. doi: 10.1016/j.renene.2018.11.056. [DOI] [Google Scholar]