Skip to main content
Sage Choice logoLink to Sage Choice
. 2021 Nov 9;24(2):536–551. doi: 10.1177/14680874211055546

Support vector machine based emissions modeling using particle swarm optimization for homogeneous charge compression ignition engine

David Gordon 1,, Armin Norouzi 1, Gero Blomeyer 1,2, Julian Bedei 2, Masoud Aliramezani 1, Jakob Andert 2, Charles R Koch 1
PMCID: PMC9902987  PMID: 36776419

Abstract

The internal combustion engine faces increasing societal and governmental pressure to improve both efficiency and engine out emissions. Currently, research has moved from traditional combustion methods to new highly efficient combustion strategies such as Homogeneous Charge Compression Ignition (HCCI). However, predicting the exact value of engine out emissions using conventional physics-based or data-driven models is still a challenge for engine researchers due to the complexity the of combustion and emission formation. Research has focused on using Artificial Neural Networks (ANN) for this problem but ANN’s require large training datasets for acceptable accuracy. This work addresses this problem by presenting the development of a simple model for predicting the steady-state emissions of a single cylinder HCCI engine which is created using an metaheuristic optimization based Support Vector Machine (SVM). The selection of input variables to the SVM model is explored using five different feature sets, considering up to seven engine inputs. The best results are achieved with a model combining linear and squared inputs as well as cross correlations and their squares totaling 26 features. In this case the model fit represented by R2 values were between 0.72 and 0.95. The best model fits were achieved for CO and CO2, while HC and NOx models have reduced model performance. Linear and non-linear SVM models were then compared to an ANN model. This comparison showed that SVM based models were more robust to changes in feature selection and better able to avoid local minimums compared to the ANN models leading to a more consistent model prediction when limited training data is available. The proposed machine learning based HCCI emission models and the feature selection approach provide insight into optimizing the model accuracy while minimizing the computational costs.

Keywords: Emission predictive modeling, homogeneous charge compression ignition, support vector machine, machine learning, particle swarm optimization, artificial neural network

Introduction

Improving internal combustion engine efficiency and reducing their emissions has the potential to improve air quality in urban centers and reduce greenhouse gas emissions. This has led to governments around the world to introducing ever more stringent environmental legislation leading automobile manufactures to turn to new combustion methods in an attempt to meet these targets. Homogeneous Charge Compression Ignition (HCCI) is a low temperature internal combustion engine mode that has the potential to significantly reduce engine-out emissions and fuel usage.1,2 HCCI is characterized by compression induced autoignition of a lean homogeneous air-fuel mixture.3 Reduced wall heat losses due to the reduced combustion temperature and short combustion duration provide HCCI with improved fuel efficiency benefits compared to conventional combustion modes.47

HCCI has shown promising engine out emissions reductions, however, the lack of a direct timing control is a major control and modeling challenge.811 Furthermore, increased Hydrocarbon (HC) and Carbon Monoxide (CO) emissions have also been observed.12 The main combustion mechanism for HCCI is compression induced autoignition of a premixed charge, leading to a high dependency on the in-cylinder gas mixture properties. To meet current and the upcoming emission regulations a deep understanding of HCCI engine emission formation is essential. To capture the behavior of HCCI combustion, various simulation models including stochastic, multi-zone and physical models have been developed to predict the gas exchange and combustion processes.1316 These models are beneficial as they provide accurate results over a wide operating range while requiring minimal validation data especially for engine performance parameters.17 However, predicting the exact value of engine out emissions using conventional physics-based models is still a challenge for engine researchers due to the complexity of combustion and emission formation modeling.18,19 Furthermore, detailed physical models are typically too computationally intensive for use in real-time engine applications and are often linearized around a specific operating point for implementation in processor based engine controllers.20,21 This has led researchers to consider machine learning (ML) based methods which help to provide an accurate model while minimizing the computational requirements.

ML techniques have been widely used for addressing engine performance, emission modeling and control.2224 To this end, different ML methods have been tested and used for HCCI performance, combustion phasing, and emission modeling using an Artificial Neural Network (ANN),2529 Extreme Learning Machine (ELM),3033 Bayesian Neural Network (BNN),34 Deep Neural Networks (DNN),35 and Least Squared Support Vector Machine (LS-SVM).3638 Among these methods, most researchers have focused on the prediction of engine performance, consisting of Indicated Mean Effective Pressure (IMEP) and CA50 (crank angle where 50% of heat energy has been released)27,3033,3638 while a limited number of researchers have studied emission prediction.25,26,28,29 ANN has been the ML method of choice and has been widely used for emission and performance prediction for Spark Ignition (SI) and Compression Ignition (CI) engines.28,3941 This has led researchers to consider ANN the baseline ML method for engine modeling and control implementation. However, to create an accurate model ANN requires a large data set which requires significant engine testing time and results in high testing costs.

One of the most powerful machine learning methods that has shown remarkable accuracy in the prediction of Internal Combustion Engine (ICE) emissions and performance is Support Vector Machine (SVM).22,23,39,42,43 SVM is a machine learning approach which has been used for both classification and regression problems.44,45 By providing the SVM with a set of input and output pairs, it approximates a hyperplane to retrieve a pattern that exists between given inputs and the corresponding outputs. For HCCI, SVM has been used to predict combustion phasing, misfire, and high pressure rise rates.46 For example, it has accurately predicted CA50 with an error of 1.9% for transient load changes,46 and cyclic combustion variability.47 Transit Linear Parameter Varying (LPV) based models were developed to predict CA50 and IMEP.3638 The accurate prediction capabilities and low computational requirements of SVM has proven it is a powerful technique for predicting the complex and highly nonlinear phenomena of other systems and this study looks to apply this strategy to emission formation in HCCI engines and compare the results to ANN. SVM has been used to predict the performance and emissions of SI48,49 and diesel22,23,42,43 engines but to date has not been comprehensively investigated for HCCI emissions prediction.

Hyperparameter tuning, is typically the most tricky part of every machine learning approach. In ANN, a grid search for the number of neurons and number of hidden layers is usually used to find optimal hyperparameters. Depending on the depth of the network, a random search could be added during optimization.26 Metaheuristic approaches were also used to tune ANN hyperparameters such as Particle Swarm Optimization (PSO)50,51 and Genetic Algorithm (GA).52 Compared with GA, PSO is a relatively new heuristic search method based on collaborative behavior and swarming in biological populations. Both GA and PSO are population-based search approaches that depend on information sharing among their population members. Although PSO and GA have a similar performance in terms of the accuracy of the solution, it has been proven that PSO is computationally more efficient, and requires fewer parameters that need to be defined for optimization.53,54 In SVM, there are three main parameters to tune which are tolerated error, kernel function parameters, and regularization coefficient. In this study, the PSO algorithm is used to tune these hyperparameters.

In this paper, a SVM technique is used to find correlations between key manipulated variables of an HCCI engine and the engine out emissions. First the linear SVM and nonlinear SVM will be compared to an ANN model for four engine out emissions. Second, a detailed investigation into the feature selection will be performed to identify which engine inputs should be used for SVM design. Then finally, the chosen model will be tested for its prediction capabilities. Knowledge about the correlation between process inputs and emissions can be then used in future control applications for HCCI model based emissions control strategies.

With the overall goal of creating a control oriented emissions prediction model for HCCI combustion. The main contributions of this work can be summarized as follows:

  • Developing a novel homogeneous charge compression ignition emission model using metaheuristic optimization based SVM,

  • Implementing particle swarm optimization method for optimizing SVM hyperparameters,

  • Analyzing feature sets based on physical understanding of the HCCI combustion process for each engine out emission component (CO, CO2, HC, and NOx),

  • Evaluating the linear and nonlinear kernels for SVM and providing a detailed comparison to an artificial neural network,

  • Proposing an accurate steady state simple emissions model design for future control applications.

Experimental setup

A Single Cylinder Research Engine (SCRE) outfitted with a fully variable Electro-Magnetic Valve Train (EMVT) is used to collect the experimental data. The flexibility of the valve timing allows for engine operation with various valve strategies including symmetric Negative Valve Overlap (NVO), which is used in this paper. Symmetric NVO is chosen where Exhaust Valve Closing (EVC) and Intake Valve Opening (IVO) are varied evenly around gas exchange Top Dead Center (TDC). This ensures that no intake or exhaust re-breathing takes place. The NVO duration can be changed every cycle if desired using this valve train.

Fuel is directly injected into the SCRE through a piezoelectric outward-opening hollow cone injector. Conventional European Research Octane Number (RON) 96 gasoline containing 10% ethanol is used and the fuel pressure is maintained at 100 bar. Cylinder pressure is measured by a Kistler 6041 piezoelectric pressure transducer which is used to calculate the Indicated Mean Effective Pressure (IMEP) as in Heywood et al.56 The air-fuel equivalence ratio λ is measured by a production Bosch wide-band oxygen sensor.

The exhaust gas measurement is done using two measurement devices. The first is an Eco Physics CLD700REht for Nitrogen Oxide (NO) and Oxides of Nitrogen (NOx) measurement and the second is a Rosemount NGA 2000 which provides measurements of unburnt Hydrocarbons (HC), Carbon Monoxide (CO), Carbon Dioxide (CO2), and Oxygen (O2) concentration. The emission analysis equipment provides an averaged emission reading due to the transport delay and mixing during transport. Therefore, the emissions values presented are average emissions over a 30 s measurement for a steady state operating point. The specifications of the emission measurement system are provided in Table 1 with full details provided in Gordon et al.10

Table 1.

Accuracy of emissions measurement system.55

Gas Maximum Detection level Resolution Accuracy
NOx 10,000 ppm 0.1 ppm 0.1 ppm 1% of reading
uHC 5% 0.04 ppm 0.1 ppm 1% of reading
CO (low) 2500 ppm 0.1ppm 0.1 ppm 1% of reading
CO (high) 10% 0.1% 0.1% 1% of reading
CO2 18% 0.1% 0.1% 1% of reading
O2 25% 0.1% 0.1% 1% of reading

It is well understood that HCCI has a narrow operating range and performs best within a specific operating conditions.5759 At first this appears as a disadvantage, however, with the transition to hybrid and electric range extender applications a few efficient load and speed operating points are acceptable as the electric systems are used to handle transient loads. To simulate a steady state operating point the engine is operated in a conditioned environment that keeps rotational speed, load, intake pressure and temperature, oil and coolant temperature, and exhaust pressure constant to minimize the effect of these confounding variables. As only one load and speed is selected this helps to reduce the experimental space in order to show the effectiveness of the proposed SVM based model. The engine geometry and chosen operating condition are listed in Table 2.

Table 2.

Single cylinder research engine parameters.10

Parameter Value
Displacement volume 0.499 l
Stroke 90 mm
Bore 84 mm
Compression ratio 12:1
Exhaust valve opening 160° aTDC
Intake valve closing 545° aTDC
Intake/exhaust pressure 1014 ± 4 mbar
Oil and coolant temperature 100 ± 1°C
Engine speed 1500 ± 5 rpm
Fuel rail pressure 100 ± 2 bar
Intake temperature 52 ± 1°C

Active input factors to the HCCI combustion process include injected fuel mass, injection timing, and valve timings. These variables were chosen to be varied as they significantly affect the combustion process and the resulting engine out emissions at a given load and speed operating point. They also significantly affect the combustion stability of the HCCI process which has a significant effect on the engine out emissions.10,60 As the HCCI process is extremely sensitive to operating conditions a relatively small change in input parameters results in a significant change in engine output parameters. These engine input parameters have also been explored in previous works regarding HCCI emissions modeling.40 The variation in engine inputs can be seen in Table 3.

Table 3.

Variation in HCCI engine input parameters.

Engine Input Min Max Mean
SOI (°aTDC) 455.9 495 473.4
mf (mg) 2.7 2.96 2.86
NVO (°) 173 201.6 187.7

Methodology

In this section, the main methodology of this study will be discussed. First, the required data for emission modeling of the HCCI engine was collected as discussed in Section 2. Figure 1 schematically shows the data collection and emission modeling. During data collection, fuel amount ( mf ), negative valve overlap duration ( NVO ), and start of fuel injection ( SOI ) are system’s main inputs while intake pressure ( Pin ) and intake temperature ( Tin ) are conditioned to keep them constant during data collection. The power output of the engine is represented by IMEP and the operating points are chosen to keep the output load approximately constant for all tests. There is a slight variation in IMEP over the operating range of 3.5 ± 0.2 bar. However, as IMEP is held relatively constant over the sweep of mf , SOI, and NVO inputs this results in varying combustion efficiency and a trade-off between different emissions.

Figure 1.

Figure 1.

Schematic of data collection and proposed emission modeling using PSO-based LSVM and NLSVM.

Then, λ , CO, CO2, NOx, and HC emission were collected. All manipulated and conditioned input and engine output variables in the emission modeling section are given to a data-driven system as inputs. Then different feature sets by interpolation of these inputs are created, and these features are the main inputs of the PSO-based SVM method. In this study, both Nonlinear SVM (NLSVM) and Linear SVM (LSVM) are considered for emission modeling using different feature sets, and PSO is used to optimize the hyperparameters of both the LSVM and NSVM. This method has been compared with the benchmark ML emission modeling, Artificial Neural Network (ANN) as proposed in literature.26

Support vector machine

The main idea of the regression form of SVM, also called Support Vector Regression (SVR) is to find an optimal hyperplane, y(ui) , such that y(ui) is as flat as possible and it has the maximum deviation of ϵ for all training data.45 In other words, the optimization problem is to find the flattest function with the maximum error tolerance ϵ . Therefore, the optimal hyperplane which describes the training data, { ui,zi }, can be defined as:

y(ui)=wTui+b (1)

where ui and zi are input and target of the training data and w and b are found by solving the SVM algorithm for regression problems. The optimization problem to find the optimum hyperplane y(ui) is defined as:

Minimize:12||w||22Subjectto:{ziwTuibϵwTui+bziϵi=1,...,n (2)

where the flatest function is achieved by minimizing 12||w||22 and the tolerance is achieved by solving for the defined constraints. A schematic of SVM regression is shown in Figure 2 where the main objective of SVM is shown as the orange line estimating a proper function by the maximum deviation of ϵ .

Figure 2.

Figure 2.

SVM regression and support vectors (based on Norouzi et al.22).

For those data points within a defined tolerance ( ϵ ), y(ui) has been found such that it predicts all pairs of learning data within a defined error. If all the data points lay within the defined tolerance, the optimization problem is feasible. However, occasionally the algorithm cannot converge within the defined constraints and the current optimization problem becomes infeasible. To overcome the infeasibility of equation (2), a penalty variable ( ζi ) or so called slack variable has been added to the original optimization problem as:

ϵζiziyiϵ+ζi+ (3)

To consider these penalty variables, the Soft Margin Loss Function (SMLF) has been added to optimization problem which is defined as61:

Minimize:12||w||22+Ci=1n(ζi++ζi)Subjectto:{ziwTuibϵ+ζi+wTui+bziϵ+ζii=1,...,nζi,ζi+0 (4)

where C is a regulatory parameter to set the trade off between tolerated error and the smoothness of the model.

To consider constraints in the optimization problem the Lagrangian function is calculated as

L=12||w||22+Ci=1N(ζi+ζi+)i=1Nαi+(zi+yi+ϵ+ζi+)i=1Nμi+ζi+i=1Nαi(ziyi+ϵ+ζi)i=1Nμiζi (5)

where αi+ , αi , μi+ , and μi are the non-negative Lagrangian Multipliers. The Lagrangian is solved by calculating the partial differential with respect to the optimization variables as

Lw=0w=i=1N(αi+αi)ui (6a)
Lb=0i=1N(αi+αi)=0 (6b)
Lζi+=0αi++μi+=C (6c)
Lζi=0αi+μi=C (6d)

where equations (6a)–(6c) are SVM expansion, bias constraints, and the box constraint, respectively.22 By substituting equations (6a)–(6d) into equation (5) the Quadratic Programming (QP) problem can be defined by

Minimize:L=12i=1Nj=1N(αi+αi)(αj+αj)uiTuji=1N(αi+αi)zi+ϵi=1N(αi++αi)Subjectto:{i=1N(αi+αi)=00αi+C0αiC (7)

which can be used in a compact version following62:

Minimize:12αTHα+fTαSubjectto:Aeqα=Beq (8)

where

α=[α+α],H=[HHHH],f=[zi+ϵzi+ϵ],H=[uiTuj],Aeq=[1...11...1],Beq=[0] (9)

In order to calculate b , the Karush–Kuhn–Tucker (KKT) conditions are used where63,64:

αi+(zi+yi+ϵ+ζi+)=0 (10a)
αi(ziyi+ϵ+ζi)=0 (10b)
μi+ζi+=(Cαi+)ζi+ (10c)
μiζi=(Cαi)ζi (10d)

must be fulfilled at the optimum point. Based on these equations, only five following cases are possible as

αi+=αi=0 (11a)
0<αi+<C,αi=0 (11b)
0<αi<C,αi+=0 (11c)
αi+=C,αi=0 (11d)
αi=C,αi+=0 (11e)

To find the support vector, where |ziyi| is exactly equal to ϵ , only 0<αi+<C,αi=0 and 0<αi<C,αi+=0 must be fulfilled. Therefore, b can be calculated as

b=1|S|iSS(ziwTuisign(αi+αi)ϵ) (12)

where S represents the support vector set based on equations (18) and (19) as:

S={i|0<αi+αi+<C} (13)

Therefore, by solving equations (12) and (8) and substituting w and b into equation (1), y(ui) is obtained as22:

y(u)=i=1N(αi+αi)uiTu+1|S|iSS(ziwTuisign(αi+αi)ϵ) (14)

Kernel tricks

Although the structure of the dot product in equation (14), is a simple linear kernel, however, it fails to capture any nonlinear behavior of the process. Therefore, by replacing the linear kernel with a nonlinear kernel, using so called kernel tricks, brings nonlinear pattern recognition at a reasonable computational cost.65 Thus, the inner product of equation (14), uiTuj , is replaced by nonlinear kernel as K(ui,uj) . In this study the RBF (Radial basis function) kernel function is used as

K(ui,uj)=exp(||uiuj||222σ2) (15)

where σ is the Gaussian variance and ||.||2 is the two norm. Therefore, the prediction function, y is calculated as65:

y=i=1n(αi+αi)K(ui,u)+b (16)

This study examines different interpolations of different features that also play the precise role of the polynomial kernel. Therefore, three main kernel types, including linear, RBF, and polynomial are considered in this study.

Hyperparameters optimization: Particle swarm optimization (PSO)

To calculate the hyperparameters for both the LSVM and the NLSVM, Particle Swarm Optimization (PSO) has been used. The LSVM and NLSVM hyperparameters are (CLSVM,ϵLSVM) and (CNLSVM,ϵNLSVM,σ) , respectively. PSO is an optimization method that optimizes a candidate solution iteratively with regard to the given cost or merit function.66,67 To train the SVM models, a total of 70 engine operating points were available. Then 80% of the data was used for training, 10% for cross-validation, and 10% as test data. Cross-validation data is used to tune the hyperparameters of the optimization methods. The cost function to find the LSVM and NLSVM is defined based on the Mean Square Error (MSE) of training and cross validation datasets. Hence, the hyperparameter calculation is defined as the following optimization problem:

[CLSVM,ϵLSVM]=argmin(1ntri=1ntr(ztr,iytr,i)2+1ncvi=1ncv(zcv,iycv,i)2)[CLSVM,ϵLSVM,σ]=argmin(1ntri=1ntr(ztr,iytr,i)2+1ncvi=1ncv(zcv,iycv,i)2) (17)

Where CLSVM and CNLSVM are the regulatory parameters for linear SVM and nonlinear kernel SVM, respectively. The index tr and cv represent training and cross-validation data set and n denotes number of data points in the data-set (i.e. ntr is number of training data points). The tolerated error for linear SVM is ϵLSVM and for nonlinear kernel SVM is ϵNLSVM . The target data and prediction data are illustrated by z and y , respectively and σ is the Gaussian variance of RBF kernel. The PSO algorithm was used to solve for the hyperparameters. The PSO-based SVM algorithm is shown in Algorithm 1 and Algorithm 2 for linear and RBF kernel of SVM, respectively. The number of particles in the swarm set for both the LSVM and NLSVM model is set to 200 while the maximum iteration number is limited to 400 and 600 for LSVM and NLSVM, respectively.

Algorithm 1: PSO based linear kernel SVM algorithm
Result: HCCI emission model: y(u)
training data set: {u,z};
splitting data set: training {utr,ztr} , cross-validation
{ucv,zcv} , and test {uts,zts};
Random hyperparameters: CLSVM,ϵLSVM;
Run Quadratic Programming of equation (11) to calculate α  and α+;
Calculate support vector sets based on equation (13)
Calculate model based on random hyperparameters using
Equation (14)
Set PSO options: MaxIterations1 ,
MaxStallIterations2 , FunctionTolerance3 ,
 and SwarmSize4 ;
whilei MaxIterations or d5 over
MaxStallIterations FunctionTolerance
do
  Calculate cost function:
   J(CLSVM,ϵLSVM)=(1ntri=1ntr(ztr,iytr,i)2+1ncvi=1ncv(zcv,iycv,i)2)
  Run PSO algorithm to minimize J(CLSVM,ϵLSVM) and
   find hyperparameters
  Update hyperparameters: CLSVM,ϵLSVM
  Run Quadratic Programming of equation (11) to calculate α
   and α+
  Calculate support vector sets based on equation (13)
  Calculate model based on Updated hyperparameters
   using equation (14)
   i=i+1
end
1. Maximum number of iterations for optimization ( =400 ), 2. Positive integer ( =20 ), 3. Non-negative scalar: Iterations end when the relative change in cost function value over the last MaxStallIterations iterations is less than FunctionTolerance =1e6 , 4. Number of particles in the swarm =200 , 5. Relative change
Algorithm 2: PSO based RBF kernel SVM algorithm
Result: HCCI emission model: y(u)
training data set: {u,z}
splitting data set: training {utr,ztr} , cross-validation
{ucv,zcv} , and test {uts,zts}
Random hyperparameters: CLSVM,ϵLSVM,σ
Run Quadratic Programming of equation (11) to calculate α  and α+
Calculate support vector sets based on equation (13)
Calculate model based on random hyperparameters using
 Equations (15) and (16)
Set PSO options: MaxIterations1 ,
MaxStallIterations2 , FunctionTolerance3 ,
 and SwarmSize4 ;
whilei MaxIterations or d5 over
MaxStallIterations FunctionTolerance
do
  Calculate cost function:
   J(CLSVM,ϵLSVM,σ)=(1ntri=1ntr(ztr,iytr,i)2+1ncvi=1ncv(zcv,iycv,i)2)
  Run PSO algorithm to minimize J(CLSVM,ϵLSVM,σ)
   and find hyperparameters
  Update hyperparameters: CLSVM,ϵLSVM,σ
  Run Quadratic Programming of equation (11) to calculate α
   and α+
  Calculate support vector sets based on equation (23)
  Calculate model based on Updated hyperparameters
   using equations (25) and (26)
   i=i+1
 end
1. Maximum number of iterations for optimization ( =600 ), 2. Positive integer ( =20 ), 3. Non- negative scalar: Iterations end when the relative change in cost function value over the last MaxStallIterations iterations is less than FunctionTolerance =1e6 , 4. Number of particles in the swarm =200 , 5. Relative change

Artificial neural network (ANN)

In this study, the proposed methods will be compared to the conventional ANN methods presented in literature. A feed-forward backpropagation network with single hidden layer and 15 neurons in each hidden layer using Levenberg–Marquardt backpropagation training method has been used in this study. This model with the same structure and number of neurons was previously developed for a single cylinder HCCI Ricardo engine.26 This is a relatively shallow network which was chosen as there is a limited amount of data available. The model training has been completed using the same parameters as used in Rezaei et al.26

Feature selection: Physical insights

A steady state emissions model is developed to predict the steady-state HCCI engine emissions values of carbon dioxide (CO2), carbon monoxide (CO), unburnt hydrocarbons (HC), and nitrogen oxides (NOx). The structure of the model is defined by equation (1) where w and b are obtained by solving the SVM algorithm for a given training data set, { u¯,z }. Here, u¯ is the normalized Feature Set (FS). In total five different FS are tested. The training target set, z , is defined based on measured steady-state CO2, CO, HC, and NOx values. To develop the model, 70 experimental data points are available where 56 points (80%) are used to train the model and 14 (20%) points to test the model.

Due to the lack of direct ignition control in HCCI, unlike conventional spark ignition in gasoline engines, the start of combustion depends on the in-cylinder conditions including pressure, temperature and gas mixture. However, these factors can only be influenced indirectly. The inputs used in this publication to set cylinder conditions and therefore affect the combustion are: Negative Valve Overlap ( NVO ); Injected Fuel Mass per cycle ( mf ); and Start of Injection ( SOI ).

Symmetric NVO is used to change the percentage of fresh air and exhaust gas within the cylinder, called Exhaust Gas Recirculation (EGR). This changes both the amount of oxygen in the cylinder as well the temperature of the cylinder charge. Generally, a lean mixture is desired for reduced NOx emissions, however, at very lean mixtures the fuel flammability limit of the fuel may be exceeded leading to combustion instability which results in high cyclic variability and increased HC emissions. NVO also impacts the cylinder temperature after compression, where a higher cylinder temperature results in an earlier auto-ignition process. The injected fuel mass directly changes the amount of fuel that is added to the cylinder. However, it is important to note that some unburnt fuel is transferred between cycles due to the trapped EGR within the cylinder. As shown in17 the amount of transferred fuel changes depending on the combustion efficiency of the last cycle. The start of injection impacts the mixture homogeneity and can lead to stratified mixtures. This has an impact on the start of combustion as well as the emissions levels. These three parameters were chosen to be varied as they provide a wide range of cylinder conditions before combustion.

Additional factors that are held as constant as possible are: intake temperature ( Tin ); intake pressure ( Pin ); and indicated mean effective pressure ( IMEP ) which is representative of applied engine load. IMEP , Tin , and Pin are active input factors to the HCCI process but they were controlled for this measurement set to reduce the number of input variables. These three factors are included to account for unwanted fluctuations, and to provide a meaningful comparison between different operating conditions. For modeling in this work the measured lambda value is used, however, it can also be accurately estimated using measured intake air flow and injected fuel demand or calculated using an online gas exchange model making it a causal variable which is useful for future control applications.17

The first sets use seven inputs to create a linear model, L7. A first extension of the FS is considering cross correlations between the variables ( mf×NVO,mf×SOI,NVO×SOI,λ×NVO,λ×SOI,λ×mf ) resulting in the L13 FS. The cross correlations with Tin , Pin , and IMEP are not taken into account as to not over interpret the effect of possible fluctuations. Then higher order correlations are also considered by adding the squares of the input variables, FS S14. Additionally, two more FS are added (S20 and S26) that consider the square of the cross correlations. Details of the five FS’s can be found in Table 4. From a machine learning point of view, these FS’s plays the exact role of a polynomial feature set. The only difference is that the redundant higher dimensional feature has been removed based on physical insight expertise.

Table 4.

Features u1u26 for the five different feature sets L7–S26. u1u7 are linear features, u8u14 are squared features, u15u20 are cross correlations, and u21u26 are the squared cross correlations.

name →
feature ↓
L7 L13 S14 S20 S26
u1=mf x x x x x
u2=NVO x x x x x
u3=SOI x x x x x
u4=Tin x x x x x
u5=Pin x x x x x
u6=IMEP x x x x x
u7=λ x x x x x
u8=mf2 x x x
u9=NVO2 x x x
u10=SOI2 x x x
u11=Tin2 x x x
u12=Pin2 x x x
u13=IMEP2 x x x
u14=λ2 x x x
u15=mf×NVO x x x
u16=mf×SOI x x x
u17=NVO×SOI x x x
u18=λ×NVO x x x
u19=λ×SOI x x x
u20=λ×mf x x x
u21=(mf×NVO)2 x
u22=(mf×SOI)2 x
u23=(NVO×SOI)2 x
u24=(λ×NVO)2 x
u25=(λ×SOI)2 x
u26=(λ×mf)2 x

L stands for linear and S stands for squared.

As the dimensions and the range of the features are quite different, all of the features should be normalized to improve the training performance.68 Here the min-max normalization method is used to normalize the features

u¯=umin(u)max(u)min(u) (18)

All of features from Table 4 are normalized for ANN and SVM methods to eliminate relative orders of magnitude difference between the features. By solving the SVM algorithm for the training data set, { u¯,z }, the approximate function, yss is obtained to predict the steady-state values of CO2, CO, HC, and NOx.

Results and discussion

To illustrate the method, the model for the CO emissions will be discussed in detail with the other emissions being similar. The recorded data points are randomly split into three sections where 80% of the collected data is used as training data to develop the models. Then 10% of the data is used for model cross-validation. Training and cross-validation data sets are used to train the model and calculate hyperparamters. In LSVM and NLVM, as discussed in 3.4, the PSO algorithm is employed to calculate the hyperparameters by solving the optimization problem of equation (17). The same training and cross-validation data are used to train an ANN model using Levenberg–Marquardt algorithm. The remaining 10% of the data is allocated for assessment of the models where the same data is used for assessing all models including LSVM, NLSVM, and ANN. To do this, the randomly chosen data points for each of the three data sets is then kept constant between all models and feature sets to allow for a fair comparison.

To rate the model quality the coefficient of determination ( R2 ) is used. It is defined by

R2=1(ziyi)2(ziz¯)2 (19)

with zi being a measured value in the data set, yi being the models response to the accompanying zi and z¯ being the mean of the measured data. The closer the R2 value is to 1 the better the model fits the data. The R2 estimate of the relationship between the dependent variables based on an independent variable may fail to tell the goodness of fit. Therefore, the Normalized Root Mean Square Error (NRMSE) is used to capture the error between the model and actual values. The Normalized version of RMSE is used to remove the dependency of RMSE to scale output and generalize the model easily. NRMSE is defined by RMSEσ where σ is the standard deviation and RMSE is defined as RMSE=Σi=1n(yiy^)n where yi is experimental value and y^ is predicted value. This criteria provides a good representation of how far the model prediction is away from the real data. Therefore, the lower the NRMSE the closer the model is to the real value. Both of these methods help to quantify the model fit.

Figure 3 shows both the R2 and NRMSE values for the training and test data for the CO model. As expected the R2 and NRMSE values are the best for the training data as the models were trained on this data set. As the model has never been trained on the testing data this reduced prediction accuracy is expected and provides the best representation of the model fit.

Figure 3.

Figure 3.

Comparison between R2 and Normalized RMSE values for CO for NLSVM and LSVM with benchmark ANN method designed based on Rezaei et al.26 in dependence of the different feature sets.

Model comparison

When comparing the coefficients of determination ( R2 ) of the LSVM, NLSVM, and ANN models in Figure 3 a few key differences can be seen. First, when only considering the training data the NLSVM and ANN models result in an improved R2 value over the simplified LSVM model. Although this does not result in a significantly improved model prediction performance when given the test data. Actually, the LSVM outperforms the ANN model in most feature sets. Showing that the ANN model can suffer from over fitting which is not seen with the simple LSVM model when presented with unknown training data. This problem with a small network, such as conventional ANN, can be reduced when using large datasets; however, when limited data is available the traditional machine learning algorithms such as SVM show a better prediction capability.22

When comparing the different feature sets, all the three models result in fairly consistent prediction accuracy even as the number of features is increased. This is especially true for both linear and non-linear SVM models which only vary by 12.0% and 14.1% R2 as the number of features is increased from 7 to 26. This is likely the result of the SVM algorithm always converging to the global minimum while the ANN model can converge to a local minimum as seen by the decrease in ANN model performance going from L7 to L13. The convergence of the ANN model is highly dependent of the initial choice of weights and bias values. This guarantee of global convergence is one the major advantages of the SVM method.7,22,42,69 The main reason for global optimization is that SVM uses Quadratic programming, which includes optimizing a function according to linear constraints. As ANN uses Gradient descent, it makes ANN sensitive to randomization of weights parameters. This means that if initial weights put cost function close to a local minimum, the accuracy of the model will never increase past a certain threshold.39 To avoid this, each ANN model is trained in a loop with multiple randomization values, where the randomization is reset until it reaches acceptable accuracy.

The findings from CO emissions can then be extended to HC, NOx, and CO2 as shown in Figures 4 to 6. The trends seen between modeling methods vary slightly between specific emissions as expected due to the physical differences in their production mechanism. To do this, a Criterion for Methods Selection (CMS), JCMS(R2) , is defined as

Figure 4.

Figure 4.

Comparison between R2 and Normalized RMSE values for HC for NLSVM and LSVM with benchmark ANN method designed based on Rezaei et al.26 in dependence of the different feature sets.

Figure 6.

Figure 6.

Comparison between R2 and Normalized RMSE values for CO2 for NLSVM and LSVM with benchmark ANN method designed based on Rezaei et al.26 in dependence of the different feature sets.

JCMS(R2)=R¯(FS)2σ(R(FS)2) (20)

Figure 5.

Figure 5.

Comparison between R2 and Normalized RMSE values for NOx for NLSVM and LSVM with benchmark ANN method designed based on Rezaei et al.26 in dependence of the different feature sets.

where σ(R(FS)2) is standard deviation of R2 and R¯(FS)2 is average value of R2 for selected feature set, L7, L13, S14, S20, and S26. Table 5 shows criterion for method selection, JCMS(R2) . This represents the lower bound of one standard deviation of uncertainty of the model fit. This helps to select a model with the best fit while ensuring the robustness of the model to changing feature sets. The goal is to have the value closest to 1. Here the best model fit score is highlighted in green and the worst is shown in red.

Table 5.

Criteria for method selection.

CMS(R2) LSVM NLSVM ANN
CO 0.809 0.818 0.641
HC 0.612 0.864 0.838
NOx 0.877 0.710 0.799
CO2 0.884 0.944 0.725

Here three of the four emissions are best represented using the NLSVM model and the other is best fit using LSVM. This shows that the SVM based models provide a stable prediction over the range of feature sets considered. A detailed analysis of the feature set will be performed next.

Feature selection

One important aspect to training the ML methods is the proper feature selection. It is important to include any features that have a correlation to the outputs of interest. However, the addition of extra features increase the model complexity and training time which is undesirable for real-time model implementation. Figure 3 shows the effect of feature selection on the model performance for CO emissions. Each feature set increases in the number of features from left to right.

The best Rtraining2 value in all cases occurs for FS L7 ( RANN2=0.999 ), while Rtest2 is maximized at S20 ( RANN2=0.959 ) for the ANN model. The R2 values are all very close for the training data at approximately R20.98 , however, a significant difference can be seen between the R2 values of the test dataset. Generally as more features are added model performance improves as seen in Figure 3 in the test data for the ANN model. As the feature set is increased from L7 to S26 a continued increase can be seen, with the exception of L13 which has a decreased model performance with the training data using the ANN model. Improved model performance does not necessarily result from increased features.

For CO emissions the best model performance on the test data occurs when using the ANN model with S20 feature set. However, for simplified control purposes the L7 feature set using the NLSVM model provides good a prediction capability with a 15.6% reduction in model fit, Rtest2 . As the main goal is to provide a real-time model for control applications this simplified and robust NLSVM prediction model is the desired choice for CO emissions prediction.

This feature analysis can then be extended extended to HC, NOx, and CO2 as seen in Figures 4 to 6. To compare the increased feature sets to the base feature set ( FS=L7 ) a percent accuracy increase in R2 value is defined, as Criterion for Feature Selection (CFS), JCFS(R2) , as:

JCFS(R2)=RFS2RL72RL72×100% (21)

This provides the relative increase in performance compared to the simplest model with lambda for the model type selected previously. Table 6 shows the improvement based on different feature sets. Here the simplest model is chosen that provides a significant increase in prediction performance ( ΔR2>2% ).

Table 6.

Criteria for feature selection.

Feature set NLSVM- CO (%) NLSVM- HC (%) LSVM- NOx (%) NLSVM- CO2 (%)
JCFS(R2) FS = L13 4.53 0.41 7.56 −1.75
FS = S14 5.12 2.60 −2.57 −1.38
FS = S20 4.78 0.59 5.09 0.34
FS = S26 4.66 0.28 4.02 −0.17
Selected FS L13 L7 L13 L7

Overall, proper feature selection is required to gain the maximum model performance. This does not mean including any and all features but rather a proper feature exploration and selection is required. In this study, the emission model for control purposes for CO, HC, NOx, and CO2 are NLSVM-L13, NLSVM-L7, LSVM-L13, and NLSVM-L7, respectively. This shows that the inclusion of more features does not necessarily result in better model prediction performance. Additionally, this shows that based on the data collected there is not a single modeling method that should be used for all emissions.

Optimization and model training time

As the propose of the proposed emissions models is hardware implementation, it is necessary to evaluate their time requirements. To evaluate this possible problem the time it takes for optimization of the hyperparameters and evaluation of the model based on optimized hyperparameters are evaluated for the CO model as shown in Figure 7(a) and (b). As shown in Figure 7(a), PSO-based NSVM requires more optimization time than LSVM. Part of this increase is because more optimization variables need to be determined using PSO compared with LSVM. As shown, ANN has a optimization time that is between NSVM and LSVM. For the ANN model, the optimization time includes multiple ANN training runs to reduce the effect of the randomized starting weights as described in the “Model Comparison” section. However, in addition to this optimization time the ANN model also requires a grid search between the number of neurons and the hidden layer size that can add up to a significant computation time. However, as in this study, the structure of the ANN is chosen based on a benchmark model for comparison purposes based on Rezaei et al.26 we did not require the grid search. The optimization part of modeling, even for the ANN grid search, does not affect the real-time implementation for two main reasons: (1) in real-time, only the already trained model is evaluated and (2) even with online learning, that is, updating model in real-time, the model will be updated based on optimized hyperparameters.

Figure 7.

Figure 7.

Optimization and evaluation time comparison between LSVM, NSVM, and ANN: (a) CO – optimization time and (b) CO – evaluation time.

The model evaluation time is based on already optimized hyperparameters and this evaluation time plays a crucial role in real-time implementation. As shown in Figure 7(b), LSVM needs 67% and 32% lower evaluation time compared to ANN and NLSVM, respectively. NLSVM also takes 52% lower computation time than ANN. These results can be extended to the other NOx, CO2, and HC models which result in an average reduction in evaluation time for the LSVM model of 64% and 28% compared to the ANN and NLSVM models, respectively. On average for the four emissions the NLSVM requires 45% lower evaluation time than the ANN model.

Chosen model performance

The model type and feature set selected in the previous sections for each of the four emissions are evaluated compared against the experimental data. Figure 8(a) to (d) show the prediction performance of the selected models along with a ±5% band shown in red.

Figure 8.

Figure 8.

Actual versus experimental- HCCI emission model. Dashed red line represent ± 5% of experimental data value: (a) CO – NLSVM – L13 – actual versus experimental, (b) HC – NLSVM – L7 – actual versus experimental, (c) NOx – LSVM – L13 – actual versus experimental, and (d) CO2 – NLSVM – L7 – actual versus experimental.

The CO2 model has all predicted values within the ±5% error bands. For the CO, HC and NOx models there is 56%, 97%, and 56% of the data points within the error bands, respectively. For the CO model there is a relatively large spread in the cross-validation and test data. However, as there is a large spread in the CO levels over the testing data points the model is still able to provide the modeling trends.

The NOx model has a larger spread in all of the data points. This could be a result of the low level of NOx emissions from 35 to 70 ppm and the stochastic variation in the HCCI combustion that is not captured in the steady state modeling. A single or only a few cycles within a measurement can greatly increase the average emissions levels.

Conclusions

This paper shows the effect of different machine learning approaches and feature sets on the model quality for HCCI emissions prediction. The goal of this work was to select an accurate model while also selecting the simplest model that still has an acceptable prediction capability for future realtime control implementation. First, linear and non-linear SVM models were compared to a traditional ANN model. This comparison showed for a small data set that SVM based models were more robust to changes in feature selection and better able to avoid local minimums compared to ANN leading to a more consistent model prediction. For each of the four emissions examined the best model type was determined by taking the highest average R2 value less the variance in R2 over the various feature sets. This led to the NLSVM being selected for three of the emissions and LSVM for NOx prediction.

Then the individual feature sets were examined. The base feature sets were extended by multiplying individual features together to explore in-feature interactions. By comparing the individual features with the base feature set (L7) the feature set with an improved accuracy that is acceptable given the increase in model complexity was chosen. In this study, the emission models chosen for control purposes for CO, HC, NOx, and CO2 are NLSVM-L13, NLSVM-L7, LSVM-L13, and NLSVM-L7, respectively. The NOx and CO models have the largest prediction error while the HC and CO2 models are quite accurate. The NOx model produced the least accurate results however it was still able to capture the trends in NOx production.

The presented SVM approach allows for emissions predictions that could be used as the basis for future real-time control applications. The inclusion of offline and online trained SVM models in engine controllers allows for real-time adaption to system aging and changes in operating conditions. Using the modeling methods identified in this work additional operating points can be tested and modeled. Additionally, the presented SVM model could be enhanced with the addition of a transient emissions model to better calculate engine out emissions during rapid load and speed changes. Implementing hybrid emission modeling by combining data-driven models with a physical-based model that provides more features from the physics of system through a chemical kinetics mechanism will be next step of this study to improve the emission model further.

Appendix

Notation

Pin Intake Pressure

Tin Intake Temperature

Λ Air-fuel Equivalence Ratio

ANN Artificial Neural Network

BNN Bayesian Neural Network

CA50 Crank angle where fifty percent of heat energy has been released

CFS Criterion for Feature Selection

CI Compression Ignition

CMS Criterion for Methods Selection

CO Carbon Monoxide

CO2 Carbon Dioxide

DNN Deep Neural Networks

EGR Exhaust Gas Recirculation

ELM Extreme Learning Machine

EMVT Fully Variable Electro-magnetic Valve Train

EVC Exhaust Valve Closing

FS Feature Set

GA Genetic Algorithm

HC HydrocarbonHCCI Homogeneous Charge Compression Ignition

ICE Internal Combustion Engine

IMEP Indicated Mean Effective Pressure

IVO Intake Valve Opening

LSVM Linear Support Vector Machine

ML Machine Learning

MSE Mean Square Error

NLSVM Nonlinear Support Vector Machine

NOx Nitrogen Oxide

NRMSE Normalized Root Means Square Error

NVO Symmetric Negative Valve Overlap

PSO Particle Swarm Optimization

RBF Radial Basis Function

RON Research Octane Number

SCRE Single Cylinder Research Engine

SOI Start of Fuel Injection

SVM Support Vector Machine

SVR Support Vector Regression

TDC Top Dead Center

Footnotes

The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding: The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: The research was performed as part of the Research Group (Forschungsgruppe) FOR 2401 “Optimization based Multiscale Control for Low Temperature Combustion Engines,” which is funded by the German Research Association (Deutsche Forschungsgemeinschaft, DFG) and with Natural Sciences Research Council of Canada Grant 2016-04646. Partial funding from Future Energy Systems at the University of Alberta is also gratefully acknowledged.

References

  • 1. Pachiannan T, Zhong W, Rajkumar S, He Z, Leng X, Wang Q. A literature review of fuel effects on performance and emission characteristics of low-temperature combustion strategies. Appl Energy 2019; 251: 113380. [Google Scholar]
  • 2. Hellstrom E, Larimore J, Jade S, Stefanopoulou AG. Reducing cyclic variability while regulating combustion phasing in a four-cylinder HCCI engine. IEEE Trans Control Syst Technol 2014; 22(3): 1190–1197. [Google Scholar]
  • 3. Yao M, Zheng Z, Liu H. Progress and recent trends in homogeneous charge compression ignition (HCCI) engines. Prog Energy Combust Sci 2009; 35(5): 398–437. [Google Scholar]
  • 4. Lehrheuer B, Pischinger S, Wick M, et al. A study on in-cycle combustion control for gasoline controlled autoignition. SAE technical paper 2016-01-0754, 2016. [Google Scholar]
  • 5. Tasoujian S, Ebrahimi B, Grigoriadis K, Franchek M. Parameter-varying loop-shaping for delayed air-fuel ratio control in lean-burn SI engines. In: Dynamic systems and control conference, Minneapolis, Minnesota, USA, 12–14 October 2016, vol. 50695, American Society of Mechanical Engineers, p.V001T01A009. [Google Scholar]
  • 6. Norouzi A, Ebrahimi K, Koch CR. Integral discrete-time sliding mode control of homogeneous charge compression ignition (HCCI) engine load and combustion timing. IFAC-PapersOnLine 2019; 52(5): 153–158. [Google Scholar]
  • 7. Shahpouri S, Norouzi A, Hayduk C, Rezaei R, Shahbakhti M, Koch CR. Soot emission modeling of a compression ignition engine using machine learning. In: IFAC-PapersOnLineModeling, estimation and control conference (MECC 2021), Austin, Texas, USA, 24-27 October 2021. [Google Scholar]
  • 8. Shahbakhti M, Koch CR. Characterizing the cyclic variability of ignition timing in a homogeneous charge compression ignition engine fuelled with n-heptane/iso-octane blend fuels. Int J Engine Res 2008; 9(5): 361–397. [Google Scholar]
  • 9. Fathi M, Jahanian O, Shahbakhti M. Modeling and controller design architecture for cycle-by-cycle combustion control of homogeneous charge compression ignition (HCCI) engines – a comprehensive review. Energy Convers Manag 2017; 139: 1–19. [Google Scholar]
  • 10. Gordon D, Wouters C, Wick M, et al. Development and experimental validation of a field programmable gate array–based in-cycle direct water injection control strategy for homogeneous charge compression ignition combustion stability. Int J Engine Res 2019; 20: 1101–1113. [Google Scholar]
  • 11. Norouzi A, Heidarifar H, Shahbakhti M, Koch CR, Borhan H. Model predictive control of internal combustion engines: a review and future directions. Energies 2021; 14: 6251. [Google Scholar]
  • 12. Gordon D, Wouters C, Kinoshita S, et al. Homogeneous charge compression ignition combustion stability improvement using a rapid ignition system. Int J Engine Res 2020; 21(10): 1846–1856. [Google Scholar]
  • 13. Ritter D, Andert J, Abel D, Albin T. Model-based control of gasoline-controlled auto-ignition. Int J Engine Res 2018; 19(2): 189–201. [Google Scholar]
  • 14. Andert J, Wick M, Lehrheuer B, Sohn C, Albin T, Pischinger S. Autoregressive modeling of cycle-to-cycle correlations in homogeneous charge compression ignition combustion. Int J Engine Res 2018; 19(7): 790–802. [Google Scholar]
  • 15. Nuss E, Ritter D, Wick M, Andert J, Abel D, Albin T. Reduced order modeling for multi-scale control of low temperature combustion engines. In: Rudibert K.(ed.) Active flow and combustion control 2018. Berlin, Germany: Springer, 2019, pp.167–181. [Google Scholar]
  • 16. Morcinkowski B. Simulative analyse von zyklischen schwankungen der kontrollierten ottomotorischen Selbstzündung, Dissertation, RWTH Aachen University, Aachen, 2015. [Google Scholar]
  • 17. Gordon D, Wouters C, Wick M, et al. Development and experimental validation of a real-time capable field programmable gate array–based gas exchange model for negative valve overlap. Int J Engine Res 2020; 21: 421–436. [Google Scholar]
  • 18. Bidarvatan M, Thakkar V, Shahbakhti M, Bahri B, Abdul Aziz A. Grey-box modeling of HCCI engines. Appl Therm Eng 2014; 70(1): 397–409. [Google Scholar]
  • 19. Hasan MM, Rahman MM. Homogeneous charge compression ignition combustion: advantages over compression ignition combustion, challenges and solutions. Renew Sustain Energ Rev 2016; 57: 282–291. [Google Scholar]
  • 20. Ebrahimi K, Koch C. Model predictive control for combustion timing and load control in HCCI engines. SAE technical paper 2015-01-0822, 2015. [Google Scholar]
  • 21. Choi S, Ki M, Min K. Development of an on-line model to predict the in-cylinder residual gas fraction by using the measured intake/exhaust and cylinder pressures. Int J Autom Technol 2010; 11(6): 773–781. [Google Scholar]
  • 22. Norouzi A, Aliramezani M, Koch CR. A correlation-based model order reduction approach for a diesel engine NOx and brake mean effective pressure dynamic model using machine learning. Int J Engine Res 2021; 22(8): 2654–2672. [Google Scholar]
  • 23. Norouzi A, Gordon D, Aliramezani M, Koch CR. Machine learning-based diesel engine-out NOx reduction using a plug-in PD-type Iterative learning control. In: Proceedings of the 4th IEEE conference on control technology and applications (CCTA 2020), Montreal, Canada, 2020. [Google Scholar]
  • 24. Yu M, Tang X, Lin Y, Wang X. Diesel engine modeling based on recurrent neural networks for a hardware-in-the-loop simulation system of diesel generator sets. Neurocomputing 2018; 283: 9–19. [Google Scholar]
  • 25. Javed S, Satyanarayana Murthy YV, Baig RU, Prasada Rao D. Development of ANN model for prediction of performance and emission characteristics of hydrogen dual fueled diesel engine with jatropha methyl ester biodiesel blends. J Nat Gas Sci Eng 2015; 26: 549–557. [Google Scholar]
  • 26. Rezaei J, Shahbakhti M, Bahri B, Aziz AA. Performance prediction of HCCI engines with oxygenated fuels using artificial neural networks. Appl Energy 2015; 138: 460–473. [Google Scholar]
  • 27. Basina LNA, Irdmousa BK, Velni JM, Borhan H, Naber JD, Shahbakhti M. An online transfer learning approach for identification and predictive control design with application to RCCI engines. In: Proceedings of the ASME 2020 dynamic systems and control conference, Virtual, Online, 5–7 October 2020. ASME. [Google Scholar]
  • 28. Bendu H, Deepak BB, Murugan S. Application of GRNN for the prediction of performance and exhaust emissions in HCCI engine using ethanol. Energy Convers Manag 2016; 122: 165–173. [Google Scholar]
  • 29. Pan W, Korkmaz M, Beeckmann J, Pitsch H. Nonlinear identification modeling for PCCI engine emissions prediction using unsupervised learning and neural networks. SAE technical paper 2020-01-0558, 2020. [Google Scholar]
  • 30. Janakiraman VM, Nguyen X, Assanis D. Stochastic gradient based extreme learning machines for stable online learning of advanced combustion engines. Neurocomputing 2016; 177: 304–316. [Google Scholar]
  • 31. Janakiraman VM, Nguyen X, Assanis D. An ELM based predictive control method for HCCI engines. Eng Appl Artif Intell 2016; 48: 106–118. [Google Scholar]
  • 32. Vaughan A, Bohac SV. Real-time, adaptive machine learning for non-stationary, near chaotic gasoline engine combustion time series. Neural Netw 2015; 70: 18–26. [DOI] [PubMed] [Google Scholar]
  • 33. Janakiraman VM, Nguyen X, Assanis D. Nonlinear model predictive control of a gasoline HCCI engine using extreme learning machines. arXiv preprint arXiv 2015; 1501.03969. [Google Scholar]
  • 34. Bao Y, Velni JM, Shahbakhti M. Epistemic uncertainty quantification in state-space LPV model identification using Bayesian neural networks. IEEE Control Systems Letters 2021; 5(2): 719–724. [Google Scholar]
  • 35. Yaşar H, Çağıl G, Torkul O, Şişci M. Cylinder pressure prediction of an HCCI engine using deep learning. Chin J Mech Eng 2021; 34(1): 1–8. [Google Scholar]
  • 36. Irdmousa BK, Rizvi SZ, Veini JM, Nabert JD, Shahbakhti M. Data-driven modeling and predictive control of combustion phasing for RCCI engines. In: 2019 American Control Conference (ACC), Philadelphia, PA, USA, 10–12 July 2019; 1617–1622. [Google Scholar]
  • 37. Basina LA, Irdmousa BK, Velni JM, Borhan H, Naber JD, Shahbakhti M. Data-driven modeling and predictive control of maximum pressure rise rate in RCCI engines. In: Proceedings of the 2020 IEEE conference on control technology and applications (CCTA), Montreal, QC, Canada, 24–26 August 2020, pp.94–99. New York: IEEE, 2020. [Google Scholar]
  • 38. Raut A, Irdmousa BK, Shahbakhti M. Dynamic modeling and model predictive control of an RCCI engine. Control Eng Pract 2018; 81: 129–144. [Google Scholar]
  • 39. Niu X, Yang C, Wang H, Wang Y. Investigation of ANN and SVM based on limited samples for performance and emissions prediction of a CRDI-assisted marine diesel engine. Appl Therm Eng 2017; 111: 1353–1364. [Google Scholar]
  • 40. Ahmed E, Usman M, Anwar S, Ahmad HM, Nasir MW, Malik MAI. Application of ANN to predict performance and emissions of SI engine using gasoline-methanol blends. Sci Prog 2021; 104(1): 00368504211002345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Bhatt AN, Shrivastava N. Application of artificial neural network for internal combustion engines: a state of the art review. Arch Comput Methods Eng 2021; 1–23. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Mohammad A, Rezaei R, Hayduk C, Delebinski TO, Shahpouri S, Shahbakhti M. Hybrid physical and machine learning-oriented modeling approach to predict emissions in a diesel compression ignition engine. SAE technical paper 2021-01-0496, 2021. [Google Scholar]
  • 43. Aliramezani M, Norouzi A, Koch CR. Support vector machine for a diesel engine performance and NOx emission control-oriented model. IFAC-PapersOnLine 2020; 53(2): 13976–13981. [Google Scholar]
  • 44. Vapnik V, Lerner A. Generalized portrait method for pattern recognition. Autom Remote Control 1963; 24(6): 774–780. [Google Scholar]
  • 45. Cortes C, Vapnik V. Support-vector networks. Mach Learn 1995; 20(3): 273–297. [Google Scholar]
  • 46. Janakiraman VM, Nguyen X, Sterniak J, Assanis D. A system identification framework for modeling complex combustion dynamics using support vector machines. In: 9th International Conference, ICINCO 2012, Rome, Italy, 28–31 July 2012, pp.297–313. Springer. [Google Scholar]
  • 47. Janakiraman VM, Nguyen X, Sterniak J, Assanis D. Identification of the dynamic operating envelope of HCCI engines using class imbalance learning. IEEE Trans Neural Netw Learn Syst 2015; 26(1): 98–112. [DOI] [PubMed] [Google Scholar]
  • 48. Gani E, Manzie C. Indicated torque reconstruction from instantaneous engine speed in a six-cylinder SI engine using support vector machines. SAE technical paper 2005-01-0030, 2005. [Google Scholar]
  • 49. Najafi G, Ghobadian B, Moosavian A, et al. SVM and ANFIS for prediction of performance and exhaust emissions of a SI engine with gasoline–ethanol blended fuels. Appl Therm Eng 2016; 95: 186–203. [Google Scholar]
  • 50. Bendu H, Deepak BB, Murugan S. Multi-objective optimization of ethanol fuelled HCCI engine performance using hybrid GRNN–pso. Appl Energy 2017; 187: 601–611. [Google Scholar]
  • 51. Li X, Wu S, Li X, Yuan H, Zhao D. Particle swarm optimization-support vector machine model for machinery fault diagnoses in high-voltage circuit breakers. Chin J Mech Eng 2020; 33(1): 1–10. [Google Scholar]
  • 52. Taghavi M, Gharehghani A, Nejad FB, Mirsalim M. Developing a model to predict the start of combustion in HCCI engine using ANN-GA approach. Energy Convers Manag 2019; 195: 57–69. [Google Scholar]
  • 53. Hassan R, Cohanim B, DeWeck O, Venter G. A comparison of particle swarm optimization and the genetic algorithm. In: Proceedings of the 46th AIAA/ASME/ASCE/AHS/ASC structures, structural dynamics and materials conference, Austin, TX, 18–21 April 2005, p.1897. [Google Scholar]
  • 54. Kouziokas GN. SVM kernel based on particle swarm optimized vector and bayesian optimized SVM in atmospheric particulate matter forecasting. Appl Soft Comput 2020; 93: 106410. [Google Scholar]
  • 55. Gordon D. Modeling and Control Strategies Utilizing Water Injection. M.Sc. Thesis, University of Alberta, 2018. [Google Scholar]
  • 56. Heywood JB. Internal combustion engine fundamentals, vol. 930. New York: Mcgraw-hill, 1988. [Google Scholar]
  • 57. Lee K, Cho S, Kim N, Min K. A study on combustion control and operating range expansion of gasoline HCCI. Energy 2015; 91: 1038–1048. [Google Scholar]
  • 58. Stanglmaier RH, Roberts CE. Homogeneous charge compression ignition (HCCI): benefits, compromises, and future engine applications. SAE Trans 1999; 108: 2138–2145. [Google Scholar]
  • 59. Duan X, Lai M-C, Jansons M, Guo G, Liu J. A review of controlling strategies of the ignition timing and combustion phase in homogeneous charge compression ignition (HCCI) engine. Fuel 2021; 285: 119142. [Google Scholar]
  • 60. Wouters C, Ottenwälder T, Lehrheuer B, et al. Evaluation of the potential of direct water injection in HCCI combustion. SAE technical paper 2019-01-2165, 20192019. [Google Scholar]
  • 61. Smola AJ, Schölkopf B. A tutorial on support vector regression. Stat Comput 2004; 14(3): 199–222. [Google Scholar]
  • 62. Bellman R. The theory of dynamic programming. Bull Am Math Soc 1954; 60(6): 503–516. [Google Scholar]
  • 63. Karush W. Minima of functions of several variables with inequalities as side constraints. MSc Dissertation. University of Chicago, Chicago, IL, 1939. [Google Scholar]
  • 64. Kuhn HW, Tucker AW. Nonlinear programming. In: Proceedings of the second Berkeley symposium on mathematical statistics and probability (Neyman J, ed.), University of California Press, Berkeley, CA, 1951, pp.481–492. [Google Scholar]
  • 65. Aliramezani M, Norouzi A, Koch CR. A grey-box machine learning based model of an electrochemical gas sensor. Sens Actuators B Chem 2020; 321: 128414. [Google Scholar]
  • 66. Norouzi A, Masoumi M, Barari A, Farrokhpour Sani S. Lateral control of an autonomous vehicle using integrated backstepping and sliding mode controller. Proc IMechE, Part K: J Multi-body Dynamics 2019; 233(1): 141–151. [Google Scholar]
  • 67. Norouzi A, Adibi-Asl H, Kazemi R, Hafshejani PF. Adaptive sliding mode control of a four-wheel-steering autonomous vehicle with uncertainty using parallel orientation and position control. Int J Heavy Veh Syst 2020; 27(4): 499–518. [Google Scholar]
  • 68. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv 2015; 1502.03167. [Google Scholar]
  • 69. Aliramezani M, Norouzi A, Koch CR. Support vector machine for a diesel engine performance and NOx emission control-oriented model. In: Proceedings of the 21st IFAC world congress, IFACPapersOnLine, Berlin, Germany, 11–17 July, 2020. Elsevier hosted at the ScienceDirect web service. [Google Scholar]

Articles from International Journal of Engine Research are provided here courtesy of SAGE Publications

RESOURCES