Abstract
This paper aims to further increase the reliability of optimal results by setting the simulation conditions to be as close as possible to the real or actual operation to create a Cyber-Physical System (CPS) view for the installation of the Fractional-Order PID (FOPID) controller. For this purpose, we consider two different sources of variability in such a CPS control model. The first source refers to the changeability of a target of the control model (multiple setpoints) because of environmental noise factors and the second source refers to an anomaly in sensors that is raised in a feedback loop. We develop a new approach to optimize two objective functions under uncertainty including signal energy control and response error control while obtaining the robustness among the source of variability with the lowest computational cost. A new hybrid surrogate-metaheuristic approach is developed using Particle Swarm Optimization (PSO) to update the Gaussian Process (GP) surrogate for a sequential improvement of the robust optimal result. The application of efficient global optimization is extended to estimate surrogate prediction error with less computational cost using a jackknife leave-one-out estimator. This paper examines the challenges of such a robust multi-objective optimization for FOPID control of a five-bar linkage robot manipulator. The results show the applicability and effectiveness of our proposed method in obtaining robustness and reliability in a CPS control system by tackling required computational efforts.
1. Introduction
Nowadays, developing processes in the engineering world is strongly associated with computer simulations. These computer codes can collect appropriate information about the characteristics of engineering problems before actually running the process. Computer simulations can provide a rapid investigation of various alternative designs to decrease the required time to improve the system. In addition, most numerical analyses for engineering problems make a well-suited use of mathematical programming. The main goals of simulation include what-if study of a model or sensitivity analysis and optimization and validation of the model [1]. The essential benefit of simulation is its ability to cover complex processes, either deterministic or random while eliminating mathematical sophistication [2]. Clearly, because of the complexity of mathematical formulation analyzing in many real-world optimization problems, simulation-optimization methods become necessary to find more interest and popularity than other optimization methods [3–5].
Cyber-Physical System (CPS) combines physical objects or systems with integrated computational facilities and data storage [6]. CPS is a key enabling technology in systems intelligence. In CPS embedded computers and networks, the physical processes are controlled usually with feedback loops where physical processes affect computations and vice versa [7–10]. CPS is a multidimensional and complex system that integrates the cyber world with the dynamic physical world. Integrating physical processes with computer systems is the main challenge presented in CPS as the computational cyber part continuously senses the state of the physical system and applies decisions and actions for its control [11]. The integration and collaboration of three terms including computing, communication, and control are known as “3C” [12, 13], CPS provides sensing, real-time optimization, information feedback, dynamic control, and other services, see Fig 1. In recent years, the application of CPS has been widely considered in different fields such as aerospace [14–16], defense [17, 18], energy systems [19, 20], healthcare [21–24], vehicle [25–27], and others [28–30].
Fig 1. Overall representation of Cyber-Physical System (CPS).
In industrial practice, many CPS systems have been designed by decoupling the control system design. In this way, CPS and real-time interaction are achieved in order to monitor and control physical entities in a reliable, safe, collaborative, robust, and efficient way [12, 13]. Using precise calculations to control a seemingly unpredictable physical environment is a great challenge [31]. After the CPS control system is designed and modeled by extensive simulation, tuning methods need to be expanded to address uncertainty and random disturbances in the system. In addition, ignoring the impact of uncertainty on the optimization model, the obtained optimal results may be far from the true optimum settings [32]. One of the main features in a reliable CPS design is the stability feature (robustness), which means no matter how the environment generates noise and uncertain factors, the control system should always reach a stable decision result eventually [33]. Robustness in the CPS control system seeks to achieve a certain level of performance with possible modeling errors in the forms of parametric or nonparametric uncertainties [34]. However, considering uncertainty and random disturbances, while keeping the function and operation of the system, has been computationally time-consuming and costly.
Because of uncertainty, more complexity in the real-time control implementation of CPS is unavoidable. So, looking for less expensive computational methods of optimization considering uncertainties has become interesting among most engineering applications [35]. To overcome such computational difficulties, researchers have applied surrogate-based learning methods (e.g. polynomial regression, GP, and radial basis function) [36–39]. Surrogate-based methods can ‘learn’ the problem behaviors and approximate the function value. These approximation models can accelerate the function evaluation as well as the estimation of the function value with acceptable accuracy. Also, they can improve the optimization performance and provide a better final solution. Various types of real-world engineering optimization problems have been developed by applying surrogate-based methods. These optimization problems include dynamic and stochastic control system design, sub-communities in machine learning problems, discrete event systems (e.g. queues, operations, and networks), manufacturing, medicine and biology, engineering, computer science, electronics, transportation, and logistics, see [3, 5, 38, 40–42]. However, several studies have systematically illustrated the applications of surrogate-based optimization algorithms [38, 39, 43–45].
In this paper, a new outline of robust real-time optimization in the CPS control model under the effect of environmental factors (also known as noise factors or uncertainty, see [4, 5]), and variability in feedback loop due to sensor’s anomaly is studied. The main contributions of this study are as follows:
In this paper, we propose a new CPS framework of the control system for a five-bar linkage robot manipulator by considering the effect of uncertainty (sources of variability) in the stochastic model. The first source is related to a real-time setpoint that is predicted by learning from collected data (e.g. surrogate) over CPS environmental factors and the second variability is found in output’s feedback due to anomaly in sensors. Besides, energy consumption and response error are optimized as a robust multi-objective optimization model by Pareto frontier estimation in the real-time computational part of the CPS model.
A new hybrid surrogate/metaheuristic algorithm for robust tuning of FOPID controller in the stochastic control system is proposed. The proposed hybrid GP/PSO algorithm has the advantages of both GP surrogates in learning the behavior of the model in an efficient global optimization with PSO metaheuristic in convergence searching for optimum results. We apply the straightforward jackknife leave-one-out technique to estimate surrogate prediction error applied in efficient global optimization.
The proposed algorithm can analyze the sensitivity of the obtained optimal results in such stochastic environments using the same collected data obtained among optimization procedure and simulation doesn’t need to be run anymore for computing the confidence intervals of robust optimal results (i.e. this algorithm does not to increase the number of function evaluations for sensitivity analysis).
The rest of this paper is organized as follows. Section 2 provides more details about real-time FOPID control when two types of uncertainties (noises) including environmental factors and sensor anomaly are considered in a CPS framework. Materials and methods of the proposed algorithm to handle robust multi-objective optimization of a CPS control system are elaborated in Section 3. In Section 4, the applicability and effectiveness of the proposed approach are examined to provide robustness and reliability in the robust optimal design of the FOPID controller in the CPS framework of a five-bar linkage robot manipulator. Finally, this paper is concluded in Section 6.
2. Point of view
The existing uncertainties and anomalies in the cyber environment have resulted in emerging concerns about the traditional control system [34]. In real-time control of CPS, physical process variables are monitored and processed by intelligent controllers for keeping the values of safety parameters between the given thresholds. Environmental conditions can affect system dynamics and also the controller function [9]. The precision of computing must interface with the uncertainty and the noise in the physical environment [46]. The physical world, however, is not entirely predictable. Normally, the CPS does not operate in a controlled environment. So, it must be robust to uncertainty (unexpected conditions) and adaptable to subsystem failures [8].
2.1 Nomenclature
The main parameters and symbols used in the proposed algorithm are revealed in Table 1.
Table 1. The table of nomenclature.
| Notation | Description |
|---|---|
| Ki, Kp, Kd, λ, μ | FOPID gain parameters (decision variables in an optimization model in this study). |
| e(t) | The error of the control system in the time moment t (distance between the output of the system with the desired set point in time moment t). |
| Desired setpoint in the control system (first uncertain variable in this study). | |
| Percent of variability in the feedback loop of the control system (second uncertain variable in this study). | |
| Ls, Us | The lower and upper limit for the uncertain variable . |
| Lα, Uα | The lower and upper limit for the uncertain variable . |
| y(t) | The output of the plant in a control system in the time moment t. |
| SEC | Signal Energy Control |
| REC | Response Error Control |
| F1, F2, OF | First, second, and overall objective functions in the optimization model respectively. |
| θ | A user-defined weighting factor used in overall function formulation and shows the tendency of the model toward F1 or F2 functions. |
| Means, Stds | Mean and standard deviation of sth input combination. Regarding the crossed array design, these statistical parameters are computed through repetitions of sth input combination over different uncertainty scenarios. |
| SNRs | Signal to noise ratio of sth input combination and computed by . |
| ω | A weighting parameter that is introduced to allow for individual emphasis on the minimization of variations in SNRs formulation. |
| l × m | The number of simulation experiments regarding the structure of crossed array design with l input combinations and m uncertainty scenarios. |
| EI(c) | The expected improvement that can be considered for the candidate point c from the best point so far. |
| γ | Type I error and shows the probability of becoming infeasible from estimated confidence intervals. |
| Rp,s | Performance measure criteria |
| CIs | Confidence intervals for optimal result using the augmented bootstrapping technique (employ the same set of data used for optimization procedure). |
| GP surrogate prediction error used in expected improvement EI(c) formulation. In this paper, the surrogate prediction error computed by the Jackknife leave-one-out approach. |
2.2 FOPID controller
In this paper, for better control, a fractional-order PIλDμ controller is used. Currently, fractional-order controllers are being extensively used by many scientists to achieve the most robust performance of the systems [47]. The main reason for choosing FOPID controllers is their additional degrees of freedom that result in a better control performance [48, 49]. A generalized FOPID controller was first introduced by [50] which proposed PIλDμ controller involving a λ order integer and a μ order differentiator. The differential equation of a fractional-order PIλDμ controller is defined by:
| (1) |
where e(t) calculates an error value as the difference between a desired setpoint and a measured process output in the time moment t. The controller attempts to minimize the error over time by adjustment of a control variable u(t). The reliability of the FOPID controller depends on the optimal design of three gain parameters (Ki, Kp, Kd) and two order parameters (λ, μ). However, we try to further increase the reliability of the tuning result by setting the traits of the simulation model to be as close as possible to the practical condition to make a CPS outline for the FOPID controller. The FOPID control system with a single setpoint does not express the aspects of the behavior that are essential to the system in the context of CPS. Moreover, we challenge the robust control to achieve CPS stability when the uncertainty in environmental conditions is the source of variability of the setpoints in the control system. In addition, an uncertain anomaly in sensors causes the noise (variability) in the control feedback loop. Moreover, we aim to tune the FOPID controller robustly in such a CPS control system with real-time setpoints and noise in the model’s feedback. Fig 2 shows the control outline of CPS with real-time setpoints and noise in the model’s feedback. The application of the integer-order and fractional-order of the PID controller in CPSs has been studied in [49, 51–53].
Fig 2. The control framework of CPS with real-time setpoints and noise in model’s feedback.
The environmental factors would be predicted and applied as a real-time setpoint and anomaly in sensor is estimated in feedback loop. Gain parameters and order parameters in FOPID controller are tuned to be robust against source of variability.
2.3 Uncertainty in the CPS control model
Assume are the environmental (uncertain) factors in such a CPS control outline. It should be noted that a real-time setpoint of the control system at time moment t is affected by variability on the environmental (uncertain) factors. Furthermore, the decision policy needs to be able to predict real-time setpoints regarding the data collected from the uncertain environmental factors so far. Here, we use supervised learning of data collected so far from the environment (e.g. polynomial regression function, and predict the real-time setpoint in the control system. In addition, an anomaly in the sensor to convey response feedback is assumed as uncertainty that causes the variability in the tuning of the FOPID controller. Assume that the true response of model y(t) is varied by where is uncertain variable , thus is a true response that is transmitted to the controller at time-step t. Fig 3 shows a block diagram representation of the CPS control system by considering both types of uncertainty including environmental factors, and sensor anomaly.
Fig 3. The block diagram of robust FOPID control in CPS framework with real-time setpoints and noise in model’s feedback.
Real-time setpoint is estimated by approximation function of environmental factors (). Anomaly in sensor’s feedback is function of uncertain variable . FOPID gain parameters and order parameters are tuned robustly in such a way to make CPS insensitive against sources of variability in system.
2.4 Objective functions
This study aims at optimizing a robust multi-objective model of the FOPID tuning in the CPS framework by considering two different objective functions (e.g. performance criteria). The authors in [54] considered the amount of energy consumed to control the plant. They applied this measure to compare the optimal results obtained by different methods. However, they don’t use the energy consumption factor for the tuning procedure. Here, we define the Signal Energy Control (SEC) as the first objective function to optimize the energy that is consumed in the time domain 0 ≤ t ≤ T:
| (2) |
and
| (3) |
where M1 is a big value that is defined by user and is used to normalize the first objective function in [0,1], so that M1 > log(max0≤t≤T SEC).
We define the second objective function, namely Response Error Control (REC) inspired by common integral absolute error criteria [47, 48] as below:
| (4) |
and
| (5) |
where M2 shows a big value that is defined by user and is used to normalize the second objective function in [0,1], so that M2 > log(max0≤t≤T REC). Notably, we use a logarithmic scale for both objective functions to smooth the large differences between the values (i.e. cases in which one or a few points are much larger than the bulk of the data). As mentioned earlier, the real-time setpoint in Eq (5) can be predicted on-time by easy-to-apply supervised learning like polynomial regression as a function of environmental uncertain factor(s).
2.5 Overall objective function
To combine both objective functions including the signal energy control (see Eq (2)) with response error control (see Eq (4)) to be used as a single objective model, we apply Lp-mertic approach by p = 2 (i.e. for more information about Lp-mertic approach in multi-objective optimization, refer to [55]). Assume s = (1,2, …, l) is the vector of input combination, then we define the Overall Function (OF) as below:
| (6) |
where θ is a user-defined weighting factor (0 ≤ θ ≤ 1) that indicates the tendency of the model toward optimization based on each objective function F1 and F2, see Eqs (2) and (4). Fluctuating this magnitude (θ) provides the capture of Pareto frontier (also called Pareto optimal efficiency) to make a trade-off between each objective function. This approach is a classical method to solve optimization problems when the model is faced with multiple criteria [56]. In fact, the set of optimal solutions obtained from fluctuating θ in [0,1] provides an estimate of the Pareto frontier.
3. Proposed algorithm
In this section, we propose a promising technique for optimization under uncertainty using augmented efficient global optimization using the jackknife leave-one-out technique to estimate GP prediction error hybrid GP/PSO method. For this purpose, we first explain the main materials and methods used in the proposed algorithm briefly and then sketch the algorithmic steps in the proposed approach.
3.1 Materials and methods
3.1.1 Gaussian Process (GP) surrogate
GP which is also known as kriging is a non-parametric Bayesian approach to supervised learning [57]. GP is an interpolation method that can cover deterministic data and is highly flexible due to its ability to employ various ranges of correlation functions [58]. In a GP model, a combination of a polynomial model and the realization of a stationary point are assumed by the form of:
| (7) |
| (8) |
where the polynomial terms of fp(X) are typically the first or the second-order response surface approach and coefficients are regression parameters (p = 0,1, …, k). This type of GP approximation is called the universal GP, while in the ordinary GP, instead of f(X), the constant mean μ = E(y(x)) is used. The term ε describes the approximation error and the term Z(X) represents the realization of a stochastic process which in general is a normally distributed Gaussian random process with zero mean, variance σ2, and non-zero covariance. The correlation function of Z(X) is defined by:
| (9) |
where σ2 is the process variance and R(xk, xj) is the correlation function that can be chosen from different correlation functions which were proposed in the literature (e.g. exponential, Gaussian, linear, spherical, cubic, and spline), see [59, 60]. Today, GP surrogate has been used as a widespread global approximation technique that is applied widely in control engineering design problems [40, 61].
3.1.2 Particle Swarm Optimizer (PSO)
The canonical PSO algorithm was proposed by [62] and is inspired by the social behavior of swarms such as bird flocking or fish schooling. The parameters of PSO consist of the number of particles, position of agent in the solution space, velocity, and neighborhood of particles (communication of topology). The PSO algorithm begins with initializing the population. The second step is to calculate the fitness values of each particle, followed by updating individual and global bests as the third step. Then, velocity and the position of the particles become updated (step four). The second to fourth steps are repeated until the termination condition is satisfied [63, 64]. The PSO algorithm is formulated as follows [62–64]:
| (10) |
where w is the inertia weight factor, and are particle velocity and particle position respectively. d is the dimension in the search space, i is the particle index, and t is the iteration number. Expressions c1 and c2 represent the speeds of regulating the length when flying towards the most optimal particles of the whole swarm and the most optimal individual particle. The term pi is the best position achieved by particle i so far and pg is the best position found by the neighbors of particle i. The expression rand(0,1) shows the random values between 0 and 1. The exploration happens if either or both of the differences between the particle’s best and previous particle’s position , and between the population’s all-time best and previous particle’s position are large. In addition, exploitation occurs when these two values are both small. PSO has attracted wide attention in control engineering design problems due to its algorithmic simplicity and powerful search performance [54, 65]. However, PSO algorithm that requires a large number of fitness evaluations before locating the global optimum is often prevented from being applied to computationally expensive real-world problems [66]. Therefore, surrogate-assisted PSO metaheuristic optimization algorithms have been focused in the literature, see [66–68].
3.1.3 Uncertainty management
Here, we follow [39, 69, 70] and inspire Taguchi’s overview of robust design [71] for dealing with uncertainty as a source of variability in the model. However, we expand Taguchi’s robust design terminology and apply its definition for environmental noise factors in such a CPS control system. But in this study, we replace the statistical approach of Taguchi viewpoint with augmented efficient global optimization using jackknife leave-one-out technique and hybrid GP/POS approach. Furthermore, we first intersect two sampling design sets. One sampling design is for decision variables (inner array) and another is for uncertain variables (outer array). Given that s = (1, 2, …, l) is the vector of sample points over decision variables, and r = (1, 2, …, m) is the vector of uncertainty scenarios, so l × m input combinations are designed, and the real model (or true simulation model) are evaluated l × m times to collect relevant simulation outputs, see Fig 4. Assume Y is the l × m matrix of simulation outputs (i.e. in this study the simulation outputs include OF values that can be obtained regarding Eq (6)), thus mean and standard deviation (Std) for each arrow in Y can be computed by the following equations:
| (11) |
| (12) |
Fig 4. Crossing two sets of DOE dealing with uncertainty in a model, one DOE (l samples) over decision variables of the model and second DOE (m samples) over uncertain variables in the model.
Signal-to-Noise Ratio (SNR) as introduced by Taguchi [71, 72] is a robustness criterion based on the mean and the Std of a system response Y. Given that, Y is the smaller the better type, Taguchi assumed zero as the minimal possible response value. Accordingly, he formulated the following SNR as the robustness criterion:
| (13) |
Since we performed a minimization of the model’s output (here is the overall function, see Eq (6)) to find the optimal parameters of the FOPID controller, the formulation of the SNR in Eq (13) has the opposite sign by Taguchi formulation. Additionally, a weighting parameter ω is introduced to allow for individual emphasis on the minimization of variations. The smallest value of SNR in Eq (13) depicts the better point with smaller relevant simulation output and higher insensitivity to the source of variability (robustness).
3.1.4 Efficient global optimization using a jackknife leave-one-out strategy
A common formulation of efficient global optimization has been developed in the outline of the expected improvement criterion, see [73, 74]. The expected improvement (EI) method has been developed in engineering design problems to adaptively improve the local and global search of optimal points (i.e. control a trade-off between exploration and exploitation properties) [75]. This method has been combined with two main parts. The first statistical part consists of the design of experiments and surrogate techniques and the second part involves evolutionary algorithms. If SNRc is considered for the arbitrary point c, an improvement function over the best point that is so far computed with SNRb is defined as max{0, (SNRb − SNRc)}. A common formulation of efficient global optimization in term of expected improvement creation is constructed as below:
| (14) |
where indicates the estimation of GP perdition’s error on candidate point c. The expression SNRb shows the value of the best signal to noise ratio that is obtained from true data of the original simulation model, and is GP surrogate prediction on candidate point c. The terms Φ and ∅ depict the cumulative distribution function (CDF) and probability density function (PDF) of a standard normal distribution respectively. The first phrase (Φ) in Eq (14) is related to the local search and the second phrase (∅) is related to a global search. In the search for the next best point among all the candidate points, the point with maximum EI term in Eq (14) is selected and replaced with the best point so far obtained. This procedure is continued until Max REI − 0 ≤ ε, where ε is a user-defined threshold, or an allocated computational cost (e.g. fixed number of repetitions) is reached. However, to find the neighbor points (candidate points) around the current best point, different strategies of sampling design methods can be used such as factorial designs [76] and space-filling design [77]. Here, we apply PSO global optimizer to investigate the maximum EI among the whole design space.
In order to estimate the surrogate prediction error for cth candidate point in Eq (14), simulation experiments can be resampled [73, 74]. The authors in [77] have used the bootstrapping technique to obtain perdition error for a GP surrogate using resampling to refit surrogate and obtain prediction error. However, resampling imposes extra computational cost due to the additional number of required simulation experiments (function evaluations). Here, to estimate the surrogate prediction error for each candidate point, we apply the jackknife leave-one-out approach. This approach applied an available set of I/O data and doesn’t need resampling and extra simulation experiments.
3.1.5 Jackknife leave-one-out approach
Jackknife was first introduced by Quenouille (1949) [78] and named by Tukey (1958) [79]. The application of the jackknife method involves a leave-one-out strategy for the estimation of a parameter (e.g. the variance) in a dataset [80]. In this study, we are motivated to use the jackknife leave-one-out approach to estimate surrogate prediction error required in Eq (14) formulation, because this method uses an existing data and does not require to re-run the expensive simulation model. Here, this method is used to predict GP prediction error while it can be used for other surrogates as well. Let denotes the prediction of GP surrogate that fitted over all l samples (input combinations), therefore the GP perdition error in cth candidate point can be estimated through the jackknife leave-one-out approach as the steps in Algorithm 1.
3.2 Algorithmic framework
In this study, we develop a new hybrid surrogate/metaheuristic method applied in robust efficient global optimization and optimization under uncertainty. We apply a PSO metaheuristic to update a GP surrogate for sequential investigation of a robust optimal point. The proposed algorithm can handle robust efficient global optimization by the exhaustive search method that can be applied in real operation of CPS control frameworks. The algorithmic representation of the proposed approach is presented in Fig 5. The main steps involved in the proposed algorithm are presented in Algorithm 2. Note that, we assume the approximation function fitted over environmental factors can be used to estimate upper (Us) and lower bound (Ls) for by varying in their relevant ranges (upper and lower bounds of each relevant environmental factor) in time-step t. Here, these bounds are predefined and existed as inputs of the program.
Fig 5. Algorithmic representation of proposed approach for hybrid GP-PSO based robust simulation-optimization under uncertainty.
Algorithm 1. Jackknife leave-one-out approach.
Input: Set of input combinations and relevant output (SNR).
Output: Estimation of surrogate prediction error for cth candidate point.
begin
Step 1. Select lc samples from the complete set of l combinations (s = 1, 2, …,l) when ic = l − k and k is a set of samples located in vertices (i.e. we aim to avoid extrapolating of GP surrogate).
Step2: Drop uth samples (simulation experiment) and relevant SNR output when (u = 1, 2, …, ic).
Step 3: Fit a new GP surrogate over (lc − 1 + k) remaining samples.
Step 4: Predict output for cth candidate point ( using the GP surrogate constructed from the previous step.
Step 5: Implement three previous steps for all lc samples computing lc relevant predictions.
Step 6: Apply the jackknife estimator to obtain the estimation of surrogate prediction error for cth candidate point as below:
End
Algorithm 2. Proposed robust simulation-optimization approach.
Input: Estimated upper (Us) and lower bound (Ls) for system’s setpoint and upper (Uα) and lower bound (Lα) for due to anomaly in sensor feedback.
Output: The estimation of the Pareto frontier by a set of robust optimal points found by the proposed approach.
begin
Step 1. Design crossed array (using the space-filling design) by crossing two sets of experiments with dimensions l × m as below:
An inner array matrix with dimension (l × nx) where l is the number of sample points for decision variables and nx is number of decision variables (e.g. in FOPID tuning nx = 5 decision variables including three gain Ki, Kp, Kd and two order λ, μ parameters).
An outer array matrix with dimension (m × nz) where m is the number of sample points (uncertainty scenarios) for nz uncertain variables (e.g. here in represented CPS control system nz = 2 including and ).
Step 2. Run the CPS model (i.e. here we use simulation model) for each crossed (l × m) combination and obtain the relevant output regarding each objective function, when s = (1, 2, …, l) and r = (1, 2, …, m).
Step 3. Compute overall function (OF) values for all l × m input combinations using Eq (6).
Step 4. Compute Means and Stds of overall function using Eqs (11) and (12) for each s = (1, 2, …, l) sample point in inner array and compute relevant SNRs using Eq (13).
Step 5. Fit a GP surrogate over sets of I/O data (with l input combinations and relevant SNRs values).
Step 6. Define an initial best point among the set of I/O data obtained from Step 4 (the point with the smallest SNR regarding Eq (13)).
Step 7. Set expected improvement criterion (see Eq (14)) as an objective function in PSO optimizer algorithm (i.e. with minimizing of −EI(c)) and obtaining a winner point.
Step 8. Run the real CPS model (e.g. original simulation model) in the winner point for m combinations of uncertainty (scenarios) designed in Step 1 and obtain relevant outputs for each objective function.
Step 9. Obtain OF values for the winner point regarding m uncertainty scenarios.
Step 10. Compute mean and Std of the winner point using Eqs (11) and (12).
Step 11. Update the set of I/O data s = (1, 2, …, l + i), when i is the number of the sequential runs.
Step 12. Fit a new GP surrogate over an updated set of I/O data (with l + i training points and SNR as outputs).
Step 13. Update (if needed) the best point obtained so far to a point with smallest SNR ratio among all the sample points (including initial training points and points which are added so far for updating of surrogate, see Step 11) and repeat Step 7 till Step 12 until stopping rules are satisfied (e.g. stop sequential updating if Max EI − 0 ≤ ε, or i ≤ k, where ε and k are user-defined thresholds).
Step 14. If stopping rule(s) is satisfied, then set the best point obtained so far as a robust optimal point of the model. The best point so far has the smallest SNR value among all sample points including initial samples and updating sample points.
Step 15. Obtain estimation of Pareto frontier by varying the weight scale θ in [0,1] (see Eq (6)) and repeating Step 1 to Step 14.
end
In this study, we use a common space-filling design method named Latin hypercube sampling (LHS) with the desired correlational function to design simulation experiments. LHS was first introduced by McKay and colleagues [81]. It is a strategy to generate random sample points while guaranteeing that all the portions of the design space are depicted. LHS has been commonly defined for designing computer experiments based on the space-filling concept. In general, for n input variables, m sample points are produced randomly in m intervals or scenarios (with equal probability). Inspired by [82] in the case of non-independent multivariate input variables, the desired correlation matrix can be used to produce distribution-free sample points in LHS. For more information, refer to [39, 83].
In the represented CPS control system in this study, outputs for two separate F1 and F1 objective functions need to be obtained regarding response error control and signal energy control, see Eqs (2) and (4). Notably, in Step 4, each s = (1,2, …, l) sample point is repeated m times through m different combinations (scenarios) of uncertain variables, see the framework of uncertainty management in Section 3.1.3. In Step 7, the fitted GP surrogate over SNR constructed in Step 5 is used to approximate the relevant SNR of each search point produced by PSO.
3.3 Augmented bootstrapping approach (sensitivity analysis)
In this study, the main idea behind the proposed algorithm is to perform sensitivity analysis to expand the information obtained from robust efficient global optimization. Estimating a single optimal point using a particular response may be inaccurate because of variability in the surrogate. Thus, we derive a series of possible responses that take into account a degree of uncertainty by providing confidence regions or prediction intervals. The author in [84] has mentioned two alternative strategies for bootstrapped resampling as follows:
In each set of bootstrapping, both sets of input (design) combination (X) and noise (uncertain) combination (Z) are resampled randomly.
The resampling is adapted to noise or uncertain component (Z) only while keeping the deterministic input combination (X) fixed.
Here, to find the bootstrapped set of data, a model is resampled B times (b = 1,2, …, B) (sampling with replacement), while B is the number of resampled or bootstrapped sample size. Moreover, B separate surrogates are fitted on B different sets of sample points with the same size (n design points). It is assumed that d+ is a robust optimal solution which is obtained from the original (non-bootstrapped) surrogate. All output values in point d+ are estimated using all the B bootstrapped surrogates. The distribution-free bootstrapped Confidence Intervals (CIs) can be computed as below, [59, 85]:
| (15) |
The superscript ‘*’ is a common symbol for bootstrapped values [59]. The expression γ gives two-sided CIs. Bonferroni’s inequality suggests that Type I error rate for each interval per output is divided by the number of outputs (here is SNR). If the values of bootstrap estimate SNR(d+)* are sorted from low to high, then ⎿.⏌ and ⎾.⏋ respectively denote floor and ceiling function to achieve the integer part and round upwards.
Here, inspired by [70, 86], the particular augmented bootstrapping approach is used for costly simulation running. In such a case, assume the set of sample points is fixed and only old data to fit surrogate with enough replication is available and new simulation replicating is very expensive. This augmented bootstrapping approach does not imply extra computational cost because of resampling and required simulation running to find a bootstrapped set of data. xs (s = 1,2, …, l) denotes the set of sample points and each xs is repeated m times (r = 1,2, …, m). We assume that the original set of data obtained from the original simulation model is available (size l × m) when m is the number of scenarios for uncertainty and l is the number of input combinations. Moreover, the augmented bootstrapping procedure is sketched in Algorithm 3.
Algorithm 3. The augmented bootstrapping procedure.
Input: Set of I/O data, and robust optimal point.
Output: Estimation of CIs.
begin
Step 1. Set s = 1 and r = 1.
Step 2. Choose (with replacement) one random number from the collection of {r* = 1, 2, …, m}.
Step 3. Replace the rth original output ys,r (selected from the old data) with the bootstrap output.
Step 4. Set r = r + 1 and continue Step 3 and Step 4 till r = m.
Step 5. Set s = s + 1 and continue Step 3, Step 4 and Step 5 till s = l.
Step 6: Compute , and SNR* using Eqs (11), (12) and (13) respectively for (s = 1, 2, …l) and fit a GP surrogate over new set of I/O data.
Step 7: Continue resampling B times (b = 1, 2, …, B) where B is the number of resampling or bootstrap sample size and compute .
Step 8: Compute bootstrapped CIs using Eq (15) for the robust optimal point obtained by the proposed algorithm as elucidated in Section 3.2).
end
Note that, regarding the Step 1 till Step 5, it can be seen that a random number with replacement in [1, m] is selected, and regarding the selected number, we choose the relevant response in an outer array (see the structure of the crossed array design explained in Section 3.1.3) that was previously collected from the original simulation model and has the same column number. For the same input combination, we repeat this procedure m times and collect m different responses or may have the same responses (i.e. because the random selection is done with replacement). This procedure is also repeated for other input combinations. Therefore, the data matrix with l row and m column is constructed.
4. Numerical example
Here, the proposed algorithm is specified for the robust optimal design of FOPID controller in CPS control of five-bar linkage robot manipulators. In the continue, we first explain the dynamics of the five-bar linkage robot manipulators. Next, the robust optimal design of FOPID controller in the CPS framework of a five-bar linkage robot manipulator is obtained using the proposed algorithm in this paper.
4.1 Dynamics of a five-bar linkage robot manipulator
Robotic manipulators, classic examples of nonlinear systems, are extensively used in the industry to automate various aspects of the production process of goods, thereby improving the quality of human life [87]. With the changing dynamics of these manipulators and their increasing complexity arising from their greater use, there has been considerable interest in their control technique fields. Robotic manipulators are Multi-Input Multi-Output (MIMO) systems with highly coupled nonlinear dynamics, posing a challenge to the development of their control scheme [88]. A five-bar linkage manipulator is a special class of parallel manipulators where a minimum of two kinematic chains control the motion of end-effectors [89]. The mechanism of a five-bar linkage is shown in Fig 6 [90].
Fig 6. Five bar linkage robot manipulator.
Even though there are four links being moved, there are in fact only two degrees-of-freedom that are defined as q1 and q2. qi and τi are the joint variable and torque of the ith motor respectively. Likewise, Ii, li, lci, and mi are the inertia matrix, length, distance to the center of gravity, and mass of the ith link respectively. In addition, if m3l2lc3 = m4l1lc4, then the inertia matrix is diagonal and constant. As a consequence, the dynamic model of the manipulator is derived by the following equations [90]:
| (16) |
where g is gravitational constant and d11 and d22 are as follows:
| (17) |
It should be noted that τ1 depends only on q1 but not on q2. Similarly, τ2 depends only on q2 but not on q1. This discussion helps to explain the popularity of the parallelogram configuration in industrial robots. If m3l2lc3 = m4l1lc4, then two angles q1 and q2 can be adjusted independently without worrying about interactions between the two angles.
4.2 Simulation and algorithm setup
Here, the main goal is to obtain a robust optimal design of FOPID controller in such a CPS control model as elucidated in Section 2. We simulate the five-bar linkage robot manipulator using Eq (16) in Matlab®/Simulink environment. Simulink does not have a library for the FOPID. Therefore, the controller from the library of FOMCON: a Matlab® toolbox for fractional-order system identification and control [91], which allows for the computation of the fractional-order derivative and integration is used. Numeric values of the parameters of the five-bar manipulator dynamics are taken from [54, 92] as shown in Table 2. From the data that are driven in Table 2, it is revealed that m3l2lc3 = m4l1lc4, thus we can perform robust optimal design of controller for one motor, where the same results are also valid for the second motor. In FOPID tuning, five decision variables including Ki, Kp, Kd, λ, and μ are considered as decision variables. The search procedure for the robust optimal result is done in the ranges as [54]:
Table 2. Numeric values of the parameters of the five-bar manipulator dynamics.
| Link | Mass (Kg) | Length (m) | C of g (m) | Inertia (Kgm2) |
|---|---|---|---|---|
| 1 | 0.2880 | 0.33 | 0.166 | 1 |
| 2 | 0.0324 | 0.12 | 0.060 | 2 |
| 3 | 0.3702 | 0.33 | 0.166 | 1 |
| 4 | 0.2981 | 0.45 | 0.075 | 2 |
Two performance criteria are considered as outputs of the model including Eqs (2) and (4) in time-step t (here the size of time-step is fixed at 0.01) and time domain (simulation time) T = 20. In addition, for uncertain variables, we assume that varies in [0.5,2.5] and varies in [−0.05,0.05]. However, we implement the proposed algorithm in CPS control framework of a five-bar linkage robot manipulator.
The following process is done to determine the robust optimal values of the FOPID parameters (Ki, Kp, Kd, λ, and μ) using the proposed algorithm. First, we design a set of experiments with the size of l = 15 samples using LHS. Another sampling design is constructed for uncertain factors and (here we choose m = 9 samples as the size of uncertainty scenarios). Two Matlab® functions “lhsdesign” and “gridsamp” are used to design training sample points with minimum correlation and to design uncertainty scenarios (different combinations of uncertain factors) respectively. We cross both sets of experiments to follow the crossed array design framework as elaborated in Section 3.1.3. Each input combination in the inner array s = (1,2, …, l = 15) including designed values of Ki, Kp, Kd, λ, and μ are sent to Simulink block for m times regarding each uncertainty scenarios r = (1, 2, …, m = 9) and the values of SEC and REC in time domain are collected. So, 15 × 9 simulation outputs are collected according to 135 simulation runs (function evaluations). We, use the collected data to obtain F1, F2, and, OF regarding Eqs (2), (4) and (6) respectively. We set both M1 and M2 parameters equal to 10 used in Eqs (2) and (4). Regarding m uncertainty scenarios, we repeat each input combination m times, and compute the relevant mean and Std of OF using Eqs (11) and (12). Then, we calculate SNRs for each input combination s = (1,2, …, l = 15) using Eq (13) while assume ω = 3. Afterwards, we fit a GP surrogate over set of input combinations and set of SNRs outputs. The DACE [93], Matlab® toolbox has been employed to construct GP surrogate. In the current study, first-order polynomial regression and Gaussian correlation functions are adjusted to fit GP surrogate. The correlation parameter is fixed on 0.1 (i.e. in the DACE toolbox, the correlation parameter is forced to vary in the range between 0.01 to 20).
Next, we perform the procedure of sequential expected improvement to estimate the robust optimal point after n sequential EI iterations. Among all the SNR values in the set of SNRs, a sample with the smallest SNR value and its relevant input combination including the relevant vector of [Ki, Kp, Kd, λ, μ] is considered as an initial best point. Regarding our proposed algorithm, we apply the PSO optimizer to search for a winner point in each sequential EI iteration. For setting the PSO parameters, the maximum iteration number is fixed 200 and the swarm is initialized with 30 particles. Notably, as we use GP surrogate instead of the true (original) simulation model as an objective function in PSO, thus we don’t worry about the computational cost due to running of a true simulation model. At the end of any sequential EI iteration, the program checks the stopping rule(s). Here, we stop the EI procedure when the EI criterion becomes smaller than 0.01, or the number of sequential runs reaches 15 iterations. Also, at the end of any sequential EI iteration, two terms of the program are updated, i) the set of training sample points by adding a winner point and relevant SNR output that is computed accordingly from the original simulation model, ii) the best sample point obtained so far with the smallest SNR among all the training points and updating points. Moreover, according to an updated set of training samples, a new GP surrogate is constructed after each sequential EI iteration. It is important to note that we avoid extrapolation of GP surrogate in each sequential iteration by setting two different rules, i) we consider a death penalty for any point that is investigated by PSO and is located out of bounds of training points, ii) to estimate GP prediction error using jackknife leave-one-out approach, we only remove input combination’s rows that don’t locate on the margin of design space (see Section 3.1.5). The obtained results from the proposed algorithm and relevant sensitivity analysis are discussed in the following sections.
4.3 Robust optimal results
We perform the proposed algorithm for three different values of θ = 0.25, θ = 0.5, and θ = 0.75 in computing OF (see Eq (6)). To evaluate the effect of randomness in sampling design methods, each optimization set was repeated 10 times. Tables 3–5 show the obtained results using the proposed algorithm for estimating robust FOPID optimal design over 10 different repetitions for θ = 0.25, θ = 0.5, and θ = 0.75 respectively. As mentioned in Section 4.2, the obtained SNR values are computed by repeating each set of FOPID gain parameters over 9 different uncertainty scenarios designed by the grid sampling method. In those tables, two expressions “In.sa” and “Up.sa” indicate the initial sampling design and updating samples that are added to the training set through the procedure of sequential improvement respectively.
Table 3. Robust FOPID optimal results using proposed algorithm for 10 repetitions for θ = 0.25 (the results obtained over 9 different uncertainty scenarios).
| Repeat No | FOPID Optimal Parameters | Simulation Experiments | Optimum SNR value | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Kp | Ki | Kd | μ | λ | Total | In.sa | Up.sa | SNR (ω = 3) | Ave | Std | |
| 1 | 4.133 | 1.411 | 1.839 | 0.967 | 0.967 | 171 | 135 | 36 | -20.564 | -20.263 | 0.219 |
| 2 | 1.000 | 2.638 | 2.925 | 0.033 | 0.790 | 180 | 135 | 45 | -20.209 | ||
| 3 | 3.321 | 1.911 | 3.875 | 0.033 | 0.804 | 198 | 135 | 63 | -19.911 | ||
| 4 | 1.000 | 4.494 | 2.854 | 0.875 | 0.831 | 171 | 135 | 36 | -20.143 | ||
| 5 | 1.000 | 2.250 | 2.559 | 0.387 | 0.852 | 207 | 135 | 72 | -20.621 | ||
| 6 | 3.311 | 2.387 | 3.206 | 0.305 | 0.896 | 207 | 135 | 72 | -20.119 | ||
| 7 | 1.509 | 0.167 | 2.922 | 0.458 | 0.615 | 216 | 135 | 81 | -20.208 | ||
| 8 | 1.000 | 2.182 | 3.430 | 0.575 | 0.726 | 198 | 135 | 63 | -20.527 | ||
| 9 | 3.311 | 2.387 | 3.206 | 0.305 | 0.896 | 207 | 135 | 72 | -20.119 | ||
| 10 | 1.509 | 0.167 | 2.922 | 0.458 | 0.615 | 216 | 135 | 81 | -20.208 | ||
Table 5. Robust FOPID optimal results using proposed algorithm for 10 repetitions for θ = 0.75 (the results obtained over 9 different uncertainty scenarios).
| Repeat No | FOPID Optimal Parameters | Simulation Experiments | Optimum SNR value | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Kp | Ki | Kd | μ | λ | Total | In.sa | Up.sa | SNR (ω = 3) | Ave | Std | |
| 1 | 8.238 | 2.300 | 3.748 | 0.426 | 0.919 | 261 | 135 | 126 | -23.868 | -24.041 | 0.140 |
| 2 | 1.267 | 3.101 | 3.332 | 0.525 | 0.835 | 270 | 135 | 135 | -24.238 | ||
| 3 | 4.285 | 1.998 | 3.510 | 0.612 | 0.938 | 189 | 135 | 54 | -24.032 | ||
| 4 | 7.803 | 3.239 | 3.257 | 0.944 | 0.918 | 216 | 135 | 81 | -24.191 | ||
| 5 | 1.025 | 2.144 | 3.072 | 0.567 | 0.860 | 270 | 135 | 135 | -24.151 | ||
| 6 | 6.113 | 1.314 | 3.039 | 0.709 | 0.967 | 243 | 135 | 108 | -24.022 | ||
| 7 | 5.775 | 2.308 | 3.689 | 0.348 | 0.913 | 234 | 135 | 99 | -23.879 | ||
| 8 | 5.201 | 2.485 | 4.124 | 0.823 | 0.794 | 261 | 135 | 126 | -24.213 | ||
| 9 | 5.775 | 2.308 | 3.689 | 0.348 | 0.913 | 234 | 135 | 99 | -23.879 | ||
| 10 | 11.00 | 1.833 | 2.833 | 0.900 | 0.967 | 270 | 135 | 135 | -23.939 | ||
Table 4. Robust FOPID optimal results using proposed algorithm for 10 repetitions for θ = 0.5 (the results obtained over 9 different uncertainty scenarios).
| Repeat No | FOPID Optimal Parameters | Simulation Experiments | Optimum SNR value | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Kp | Ki | Kd | μ | λ | Total | In.sa | Up.sa | SNR (ω = 3) | Ave | Std | |
| 1 | 3.321 | 1.549 | 2.634 | 0.967 | 0.967 | 225 | 135 | 90 | -21.838 | -21.840 | 0.245 |
| 2 | 1.000 | 0.713 | 2.017 | 0.802 | 0.726 | 261 | 135 | 126 | -22.249 | ||
| 3 | 1.000 | 2.668 | 3.604 | 0.575 | 0.832 | 225 | 135 | 90 | -21.879 | ||
| 4 | 1.000 | 1.228 | 2.436 | 0.657 | 0.844 | 189 | 135 | 54 | -22.120 | ||
| 5 | 1.000 | 2.667 | 3.947 | 0.696 | 0.757 | 225 | 135 | 90 | -21.949 | ||
| 6 | 3.429 | 1.782 | 3.410 | 0.298 | 0.861 | 243 | 135 | 108 | -21.562 | ||
| 7 | 7.000 | 1.833 | 2.833 | 0.433 | 0.967 | 198 | 135 | 63 | -21.448 | ||
| 8 | 1.000 | 1.964 | 3.909 | 0.775 | 0.705 | 180 | 135 | 45 | -22.003 | ||
| 9 | 3.740 | 1.702 | 2.151 | 0.033 | 0.915 | 189 | 135 | 54 | -21.551 | ||
| 10 | 2.233 | 1.624 | 2.000 | 0.318 | 0.967 | 180 | 135 | 45 | -21.803 | ||
As can be seen, for θ = 0.25, the best SNR (−20.621) is obtained through the fifth repetition with a total of 207 function evaluations (15 × 9 runs regarding initial crossed sampling design and 8 × 9 simulation runs regarding sequential updating of the training sample set). For θ = 0.5, the best SNR (-22.249) is obtained from the second repetition and with a total of 261 simulation experiments (135 initial samples plus 126 updating samples). The best SNR value (-24.238) for θ = 0.75 is obtained through the second repetition by a total of 270 function evaluations (15 initial input combinations and 15 update combinations that are crossed by 9 uncertainty scenarios). We consider all the three best points over all 10 repetitions as robust optimal points for the FOPID controller using the proposed algorithm for each θ = 0.25, θ = 0.5, and θ = 0.75 separately. Fig 7 shows the magnitudes of the EI criterion and the best SNR obtained by sequential expected improvement over 10 different repetitions of the proposed algorithm for θ = 0.25, θ = 0.5, and θ = 0.75. Also, the mean and Std of OF related to the best point so far (smaller SNR) in each sequential EI iteration are shown in Fig 8. It should be noted that two stopping rules are adjusted, the EI value becomes smaller than 0.01, or the sequential procedure reaches 15 sequential iterations. Fig 9 shows the step responses of the robot manipulator with 9 different uncertainty scenarios ( and ) for θ = 0.25, θ = 0.5, and θ = 0.75.
Fig 7. EI criterion magnitudes and best SNR obtained by sequential expected improvement over 10 different repetition of proposed algorithm for θ = 0.25, θ = 0.5 and θ = 0.75.
Two stopping rules are adjusted, EI value becomes smaller than 0.01 or reach 15 sequential iterations.
Fig 8. Mean and Std of overall function (OF) related to best point so far (smaller SNR) obtained by sequential expected improvement over 10 different repetition of proposed algorithm for θ = 0.25, θ = 0.5 and θ = 0.75.
Two stopping rules are adjusted, EI value becomes smaller than 0.01 or reach 15 sequential iterations.
Fig 9. The step responses of the robot manipulator with 9 different uncertainty scenarios ( and ) for θ = 0.25, θ = 0.5 and θ = 0.75.
4.4 Sensitivity analysis
To analyze the sensitivity of robust optimal results obtained by the proposed algorithm and estimate the variability which occurred due to randomness in designing sample points, we used the augmented bootstrapping method explained in Section 3.3. Here, based on the obtained results from the original GP surrogate, the FOPID parameters in robust optimum point (d+) for θ = 0.25, θ = 0.5, and θ = 0.75 are defined as below:
- For, θ = 0.25
- For θ = 0.5
- For θ = 0.75
With B predicted values from B bootstrapped GP surrogates, we can quantify the CIs (bootstrapped confidence intervals) for d+. For the current case, we selected the bootstrapped size B = 50. We predicted SNR by each B = 50 bootstrapped surrogates in the robust optimal point which are obtained by original GP surrogates. From these 50 values for SNR, we estimated CIs for SNR by applying Eq (15). We quantified these confidence regions for γ = 0.05 (i.e. γ denotes type I error and shows the probability of becoming infeasible from estimated confidence regions). As we estimated the robustness as a consequence of the uncertainty in the model, it becomes important to implement further analyses of the statistical variation. The CIs shows that the original estimated SNR may still give variety regarding its threshold due to the variability of surrogates’ predictions. However, 95% two-sided approximations of CIs obtained by bootstrapped GP surrogates for SNR values in robust optimal points of FOPID controller regarding θ = 0.25, θ = 0.5, and θ = 0.75 are as follows:
-
For θ = 0.25
Lower bound:
Upper bound:
-
For θ = 0.5
Lower bound:
Upper bound:
-
For θ = 0.75
Lower bound:
Upper bound:
Figs 10–12 show the sensitivity analysis obtained by bootstrapping for θ = 0.25, θ = 0.5, and θ = 0.75 respectively regarding FOPID gain parameters (Kp, Ki, Kd) and order parameters (μ, λ).
Fig 10. Sensitivity analysis via 50 bootstrapped GP surrogate and 95% Confidence Intervals (CIs) over robust optimal point obtained by original GP surrogate for θ = 0.25.
Augmented parametric bootstrapping is performed using on hand set of input/output data provided among original optimization program.
Fig 12. Sensitivity analysis via 50 bootstrapped GP surrogate and 95% Confidence Intervals (CIs) over robust optimal point obtained by original GP surrogate for θ = 0.75.
Augmented parametric bootstrapping is performed using on hand set of input/output data provided among original optimization program.
Fig 11. Sensitivity analysis via 50 bootstrapped GP surrogate and 95% Confidence Intervals (CIs) over robust optimal point obtained by original GP surrogate for θ = 0.5.
Augmented parametric bootstrapping is performed using on hand set of input/output data provided among original optimization program.
4.5 Proposed algorithm versus three common optimizers in FOPID tuning
In this section, we compare the results obtained by a proposed algorithm with three common FOPID optimization methods including the PSO metaheuristic [62] (i.e. when directly used in optimization procedure), Grey Wolf Optimizer (GWO) [94], and Ant Lion Optimizer (ALO) [95]. These methods have been widely used in the literature for optimal control systems [49, 54, 96, 97]. We compare both levels of accuracy (lower objective function) and the robustness of each method in the tuning of stochastic controllers. Here, we assume that the model is limited to only 270 simulation experiments (function evaluations) to obtain a robust optimal design of stochastic FOPID controller for θ = 0.25, θ = 0.5, and θ = 0.75. So, we let each optimizer employ a maximum of 270 simulation experiments. It should be noted that we also allowed our proposed algorithm to use maximum 270 simulation experiments to search for the optimal point, see Section 4.2. The parameters settings for all the three optimizers (PSO, GWO, and ALO) are as follows:
The number of iterations is considered to be 30 and initial papulation is adjusted to 9. The other parameters for PSO [62, 63] are also selected as follows:
Min of inertia weight equals to 0.4; max of inertia weight equals to 0.9; all the three factors of velocity clamping factor, cognitive constant, and social constant are set to 2. To run each of the three mentioned optimizers in each relevant iteration by optimizer, we randomly (with replacement) produce a scenario of uncertainty and compute output of the original simulation including SEC and REC and compute OF as an objective (fitness) function of optimizer. To make a fair comparison between the proposed algorithm and three optimizers (PSO, GWO, and ALO) in the stochastic FOPID control system, we repeat each of the three optimizers 10 times (as mentioned in Section 4.2 and 4.3, we repeated the proposed algorithm 10 times). To compare the obtained results using proposed algorithm and the three global optimizers (PSO, GWO, and ALO), we produce 100 different combinations (scenarios) of two uncertain factors including and using grid sampling design approach. Afterwards, for each set of optimal FOPID parameters according to the obtained results by proposed algorithm and three global optimizers, we run true simulation model regarding each uncertainty scenario (total 100 simulation runs for each set of FOPID optimal point).
4.5.1 Comparison results
Tables 6–8 provide the statistical comparison results between the proposed algorithm and three common optimizers in 10 separate repetitions for θ = 0.25, θ = 0.5, and θ = 0.75 respectively. In these tables, the level of accuracy (lower objective function) and robustness for the stochastic FOPID tuning are compared. The FOPID tuning results using different methods are obtained over 100 different uncertainty scenarios. Note that the expression “SE” in the tables indicates the total number of simulation experiments (function evaluations) employed for the optimization procedure. It should be also noted that we allow each optimizer to use a maximum of 270 function evaluations. For the proposed algorithm, we consider two stopping rules for the relevant procedure of sequential improvement. The EI criterion becomes smaller than 0.01 or the algorithm reaches 15 sequential iterations (See Section 4.2).
Table 6. Comparison results for FOPID tuning using different methods over 100 different uncertainty scenarios for θ = 0.25.
| Repeat No | Proposed algorithm | PSO | GWO | ALO | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave / Std | SE | |
| 1 | -20.806 | -20.499 / 0.238 | 1971 | -19.325 | -20.404 / 0.630 | 2700 | -21.351 | -20.944 / 0.545 | 2700 | -20.666 | -20.205 / 0.642 | 2700 |
| 2 | -20.450 | -20.620 | -21.338 | -18.721 | ||||||||
| 3 | -20.112 | -20.337 | -21.076 | -20.131 | ||||||||
| 4 | -20.349 | -19.647 | -20.877 | -20.066 | ||||||||
| 5 | -20.894 | -20.352 | -19.477 | -20.742 | ||||||||
| 6 | -20.329 | -21.162 | -21.484 | -20.736 | ||||||||
| 7 | -20.470 | -20.841 | -20.695 | -20.217 | ||||||||
| 8 | -20.782 | -19.806 | -21.036 | -20.460 | ||||||||
| 9 | -20.329 | -21.426 | -20.841 | -20.875 | ||||||||
| 10 | -20.470 | -20.524 | -21.260 | -19.434 | ||||||||
Table 8. Comparison results for FOPID tuning using different methods over 100 different uncertainty scenarios for θ = 0.75.
| Repeat No | Proposed algorithm | PSO | GWO | ALO | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave / Std | SE | |
| 1 | -24.094 | -24.292 / 0.150 | 2448 | -23.774 | -23.198 / 1.675 | 2700 | -24.012 | -24.028 / 0424 | 2700 | -24.004 | -24.032 / 0.424 | 2700 |
| 2 | -24.523 | -19.045 | -24.262 | -23.300 | ||||||||
| 3 | -24.295 | -24.050 | -24.607 | -24.253 | ||||||||
| 4 | -24.395 | -24.077 | -23.667 | -24.608 | ||||||||
| 5 | -24.471 | -21.273 | -23.075 | -24.596 | ||||||||
| 6 | -24.276 | -24.416 | -24.504 | -23.998 | ||||||||
| 7 | -24.136 | -24.385 | -23.845 | -24.064 | ||||||||
| 8 | -24.444 | -22.591 | -23.855 | -24.191 | ||||||||
| 9 | -24.136 | -24.384 | -24.258 | -24.004 | ||||||||
| 10 | -24.153 | -23.986 | -24.198 | -23.300 | ||||||||
As can be seen from Table 6, the lower SNR is obtained by GWO with -20.944, and it is followed by our proposed algorithm with -20.499. The PSO and ALO optimizers also provided competitive results. Regarding the effect of randomness on 10 repetitions (stochastic optimization), the proposed algorithm shows the most robustness behavior (Std = 0.238) in comparison to the other three optimizers. The GWO, PSO, and ALO optimizers are the second, third, and fourth robust methods respectively.
From the results in Table 7, it is clear that the proposed algorithm provides competitive result in obtaining the lower SNR value with -22.112 compared to the lowest SNR value (-22.165) by GWO. The ALO and PSO are ranked third and fourth in obtaining lower objective function (SNR). In Table 7, it is readily apparent that the most robust method against randomness in the model is the proposed algorithm with Std = 0.280 compared to the other three optimizers. The next robust methods are GWO, ALO, and PSO optimizers respectively.
Table 7. Comparison results for FOPID tuning using different methods over 100 different uncertainty scenarios for θ = 0.5.
| Repeat No | Proposed algorithm | PSO | GWO | ALO | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave / Std | SE | SNR | Ave | SE | |
| 1 | -22.076 | -22.112 / 0.280 | 2115 | -20.915 | -21.101 / 1.066 | 2700 | -22.376 | -22.165 / 0.377 | 2700 | -21.801 | -21.824 / 0.562 | 2700 |
| 2 | -22.616 | -21.053 | -22.470 | -21.903 | ||||||||
| 3 | -22.136 | -19.600 | -21.797 | -21.780 | ||||||||
| 4 | -22.448 | -19.120 | -22.417 | -22.017 | ||||||||
| 5 | -22.201 | -20.193 | -21.358 | -21.806 | ||||||||
| 6 | -21.810 | -22.173 | -22.372 | -21.824 | ||||||||
| 7 | -21.653 | -22.005 | -22.550 | -22.540 | ||||||||
| 8 | -22.262 | -21.965 | -22.257 | -20.281 | ||||||||
| 9 | -21.821 | -22.057 | -22.338 | -22.216 | ||||||||
| 10 | -22.096 | -21.929 | -21.714 | -22.077 | ||||||||
It is apparent from Table 8 that the lower SNR value (objective function) with -24.292 is obtained by the proposed algorithm compared to the other three optimizers. The ALO, GWO, and PSO optimizers are ranked second, third, and fourth respectively in obtaining lower SNR value. Regarding robustness against randomness in the stochastic optimization model, the proposed algorithm also shows the most robust behavior compared to the other three methods. The PSO optimizer shows a weak performance to obtain robustness (see Std statistics) in the current case study.
To perform further comparison of the proposed algorithm with three different optimizers including PSO, GWO, and ALO, we apply two common statistical tests including the t-test and the Wilcoxon signed ranks test [98]. Inside the field of inferential statistics, hypothesis testing [99] can be employed to draw inferences about one or more populations from given samples (results). In order to do that, two hypotheses, the null hypothesis H0 and the alternative hypothesis H1, are defined. The null hypothesis is a statement of no effect or no difference, whereas the alternative hypothesis represents the presence of an effect or a difference (in our case, significant differences between algorithms). Table 9 provides the p-values using the t-test and the Wilcoxon signed ranks test. These statistical test results computed for all the pairwise comparisons concerning the proposed algorithm compared with three optimizers of PSO, GWO, and ALO. In general, the Wilcoxon signed ranks test is safer than the t-test because it does not assume normal distributions. Also, the outliers (exceptionally good/bad performances of an algorithm in a few repetitions) have less effect on the Wilcoxon test than on the t-test [100]. As Table 9 states, the proposed algorithm shows a significant improvement over GWO for θ = 0.25 with a level of significance α = 0.05 regarding the t-test and with α = 0.01 regarding the Wilcoxon test. For θ = 0.5, the results for both types of test indicate an improvement of proposed algorithm over PSO with a level of significance α = 0.01 and an improvement over ALO with α = 0.1. The statistical comparison results for θ = 0.75 using t-test shows an improvement of proposed algorithm over PSO with a level of significance α = 0.05 and over GWO and ALO with α = 0.1. The Wilcoxon test also for θ = 0.75 indicates a significant improvement of proposed algorithm over PSO with a level of significance α = 0.01, an improvement over GWO with α = 0.1, and an improvement over ALO with α = 0.05.
Table 9. p-values of the t-test and the Wilcoxon signed rank test for pairwise comparison of proposed algorithm with three common stochastic optimizers over 10 repetitions.
| Optimizer | t-test | Wilcoxon signed rank test | ||||
|---|---|---|---|---|---|---|
| θ = 0.25 | θ = 0.5 | θ = 0.75 | θ = 0.25 | θ = 0.5 | θ = 0.75 | |
| PSO | 0.339826 | 0.00999 | 0.04112 | 0.5 | 0.00778 | 0.00955 |
| GWO | 0.02200 | 0.36948 | 0.05283 | 0.00408 | 0.24815 | 0.07546 |
| ALO | 0.11113 | 0.09626 | 0.05497 | 0.16288 | 0.08681 | 0.04815 |
4.5.2 Performance measure
In many studies on optimization, the strength of an optimization technique is measured by comparing the final solution achieved by different algorithms [101, 102]. This approach only provides information about the quality of the results and neglects the speed of convergence which is a very important measure for expensive optimization problems. Comparing the convergence curves (number of function evaluations) is also one of the common benchmarking approaches [103]. A convergence curve provides good information about the final quality of the optimization results in terms of computational cost, even though it can be used to compare the performance of several algorithms only for one problem. Moré and Wild [104] have suggested performance measure for any pair (p, s) of problem p and solver s to analyze the performance of any optimization algorithm as below:
| (18) |
where P is a set of problems, S is a set of solvers, and tp,s is the required number of the function evaluations for solver s ∈ S to solves a particular problem p ∈ P. In Eq (18), larger values of tp,s indicate a worse performance. The convention rp,s = ∞ is used when solver s fails to satisfy the convergence test on problem p. However, Eq (18) considers the required budget to solve the expensive optimization problem. In this study, inspired by [104], a new performance measure is used to consider two aspects of the algorithm’s performance including the level of accuracy and computational cost as below:
| (19) |
where lp,s indicates the level of accuracy (i.e. lower objective function) for solver s in an expensive problem p and β(0 ≤ β ≤ 1) is the weight scale. Note that the best solver for a particular problem p obtains the lower bound Rp,s = 1. In β = 1, Eq (19) makes the same measure with suggested rp,s based on [104] in Eq (18). For β = 0, only the accuracy of solver in obtaining lower objective function is considered as a performance measure for comparing all the optimization methods. However, the computational cost (i.e. the number of the function evaluation) is not attended. Here, we measure the performance of the proposed algorithm and other solvers for both average and standard deviation of the optimal results obtained for 10 repetitions regarding the results presented in Tables 6–8. Moreover, we apply the Eq (19) when β varies in [0,1] to obtain performance measures. The results are shown in Fig 13. This figure reports an appropriate performance (Rp,s) of the proposed algorithm for the stochastic FOPID tuning or robot manipulator compared to the other solvers when the budget of optimization for an expensive control system is limited to a small number of function evaluations. As mentioned before, in the current case we assume a maximum of 270 function evaluations.
Fig 13. The performance comparison of proposed algorithm with other three solvers in the literature for tuning of stochastic FOPID controller.
The performance criterion Rp,s measured based on two terms, accuracy of solution (lower objective function) and number of function evaluations (computational cost), see Eq (19).
5. Discussion
This paper followed two main purposes i) sketching a new framework of a stochastic control system for robot FOPID control under uncertainty in a CPS framework ii) developing a new optimization algorithm associated with such stochastic control systems. Regarding the second purpose, there are some rationales including:
Most existing methods have been developed to apply in the deterministic control systems rather than the stochastic or random models [47, 54, 96]. Moreover, we firstly developed a new straightforward algorithm that can handle uncertainty for any stochastic behavior of control systems [35]. Notably, here we show the applicability of the proposed method for robust FOPID tuning of a robot manipulator in a CPS framework.
Most practical control systems in CPS are complex in terms of dynamic mathematical sophistication or time-consuming simulation experiments [36, 47]. Therefore, we aimed to propose a new less-expensive method for complex black-box simulation models when a limited (small) number of input-output data needs be applied in the control system (i.e. the model is limited to a few numbers of simulation experiments or function evaluations).
Besides optimal design (lower objective function) and robustness against sources of variability (uncertainty) with a small number of simulation experiments, we are also interested to perform the sensitivity analysis (bootstrapping) for the obtained results in the stochastic (random) control system. The proposed algorithm in this study can compute the two-sided confidence intervals for the obtained optimal results using the same set of data produced among optimization procedure and it doesn’t need extra simulation experiments (function evaluations) for sensitivity analysis.
As elucidated in the No Free Lunch (NFL) theorem by [105], a particular optimization method may show very promising results on a set of problems, but the same algorithm may show poor performance on a different set of problems [94]. This is also another motivation for conduction this study.
The proposed hybrid algorithm has been compared with the three most well-known optimizers (PSO, GWO, and ALO, see [49, 54, 96, 97] that are commonly used in control tunings. The comparison results have been provided in Section 4.5. The performance of the proposed algorithm can be evaluated by three factors as below:
Level of accuracy (lower objective function): The proposed algorithm provides competitive results to obtain a lower objective function with three aforementioned optimizers.
Robustness: The proposed algorithm shows a higher robustness behavior against randomness in the stochastic model of the control system.
The number of simulation experiments (computational cost) that are required for optimization procedure.
The proposed hybrid algorithm has multi-disciplinary applications. In other words, the proposed algorithm can be applied in a variety of engineering design problems under the effect of uncertainty with expensive black-box simulation models. In this paper, we show the application of the proposed algorithm in the stochastic robot control system (robust tuning of FOPID controller for five-bar linkage robot manipulator).
The main limitations of the current study are as below:
The proposed algorithm employed a Gaussian process (Kriging) surrogate to train the control system using space-filling sampling strategies. Therefore, the approximate errors could not be ignored when solving simulation-based optimization problems particularly with complex function and nonlinear structure. It is well known that GP is an ideal choice for smooth models. If the functions are non-smooth or noisy, it is likely that GP surrogate degrades rapidly and overfits due to its interpolating behavior. A challenge for optimization under restricted budgets will be to find the right degree of approximation (smoothing factor) from a limited number of samples [102].
In this study, the proposed algorithm was evaluated in a stochastic control system with five design variables and two uncertain variables. The proposed algorithm should be evaluated more in other practical stochastic problems with a higher dimension and degree of uncertainty.
Here, the two main weight-scale parameters including θ in Eq (6) and ω in Eq (13) mostly influenced the performance of the proposed algorithm in obtaining the robust optimal solution. The range of parameter θ was between zero and one (0 ≤ θ ≤ 1) and any positive value could be assigned to the parameter ω. In this study, we considered three different values for θ including θ = 0.25, θ = 0.5, and θ = 0.75 and performed the optimization procedure separately. While the parameter ω was assumed to be a fix value ω = 3 for all the values of θ. This paper was limited to evaluate the effect of parameter θ only in obtaining the robust optimal solutions by the algorithm. However, more analysis is required to study the effect of both parameters of θ and ω at the same time and in the same platform of problem.
6. Conclusion
In this paper, a new CPS framework of fractional-order PID controller is developed by considering uncertainty in the control system. To optimize such a stochastic control system, a new hybrid surrogate/metaheuristic-based robust simulation-optimization algorithm is proposed that possesses the advantages of both GP surrogate in learning the behavior of the model for efficient global optimization and PSO metaheuristic in convergence searching of optimum results. We smooth the application of PSO using GP surrogate instead of the original simulation model to diminish computational cost due to a large number of fitness evaluations required for the global optimizer when used individually. Also, this simple modified algorithm is developed in such a way to handle computational complexity to obtain optimal and robust FOPID design in the CPS control system. In such a system, we also consider the conflict between multiple objective functions and uncertainty in the model. The proposed algorithm can analyze the sensitivity of the computed robust optimal results (i.e. it obtains two-sided confidence intervals). Here, we apply this approach to the robust optimal control design of a five-bar linkage robot manipulator to depict the applicability and effectiveness of the proposed algorithm. Comparative simulation results reveal that the proposed hybrid GP/PSO-based robust efficient global optimization algorithm can effectively and robustly tune the parameters of the FOPID controllers. From an application point of view, the introduced technique is simple and fast and has a suitable control over the error and energy of a system and it can be easily implemented in real-world applications of CPS control systems. Future research may address the issues regarding the limitations of the current study mentioned in Section 5.
Supporting information
(ZIP)
Data Availability
The supplementary materials including all Matlab® codes and functions and Excel data set and analysis required to reproducing and replicating of our study’s findings have been shared publicly in the Zenodo repository via DOI and URL as below: DOI: 10.5281/zenodo.4266126 URL: https://doi.org/10.5281/zenodo.4266126.
Funding Statement
This research project is supported by Second Century Fund (C2F), Chulalongkorn University, Bangkok. The authors acknowledge the support by Second Century Fund (C2F), Chulalongkorn University, Bangkok for funding and postdoctoral fellowship to Amir Parnianifard (first and corresponding author). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.van Beers W. C. M. and Kleijnen J. P. C., Kriging for interpolation in random simulation, Journal of the Operational Research Society, 54, (3), pp. 255–262, 2003. [Google Scholar]
- 2.Figueira G. and Almada-Lobo B., Hybrid simulation optimization methods a taxonomy and discussion, Simulation Modelling Practice and Theory, 46, pp. 118–134, 2014. [Google Scholar]
- 3.Amaran S., Sahinidis N. V., Sharda B., and Bury S. J., Simulation optimization: a review of algorithms and applications, Annals of Operations Research, 240, (1), pp. 351–380, 2016. [Google Scholar]
- 4.Parnianifard A., Azfanizam A. S., Ariffin M. K. A., and Ismail M. I. S., An overview on robust design hybrid metamodeling: Advanced methodology in process optimization under uncertainty, International Journal of Industrial Engineering Computations, 9, (1), pp. 1–32, 2018. [Google Scholar]
- 5.Parnianifard A., Azfanizam A., Ariffin M., Ismail M., and Ebrahim N., Recent developments in metamodel based robust black-box simulation optimization: An overview, Decision Science Letters, 8, (1), pp. 17–44, 2019. [Google Scholar]
- 6.Skowroński R., The open blockchain-aided multi-agent symbiotic cyber–physical systems, Future Generation Computer Systems, 94, pp. 430–443, 2019. [Google Scholar]
- 7.Lee E. A., The past, present and future of cyber-physical systems: A focus on models, Sensors (Switzerland), 15, (3), pp. 4837–4869, 2015. 10.3390/s150304837 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.E. A. Lee, Cyber physical systems: Design challenges, in Proceedings—11th IEEE Symposium on Object/Component/Service-Oriented Real-Time Distributed Computing, ISORC 2008, 2008, pp. 363–369.
- 9.E. a Lee, Computing Foundations and Practice for Cyber- Physical Systems: A Preliminary Report, Electrical Engineering, (UCB/EECS-2007-72), pp. 1–27, 2007.
- 10.M. Zamani, Control of cyber-physical systems using incremental properties of physical systems, 2012.
- 11.Koulamas C. and Kalogeras A., Cyber-physical systems and digital twins in the industrial internet of things, Computer, 51, (11), pp. 95–98, 2018. [Google Scholar]
- 12. Tao F., Qi Q., Wang L., and Nee A. Y. C., Digital Twins and Cyber–Physical Systems toward Smart Manufacturing and Industry 4.0: Correlation and Comparison, Engineering, 5, (4), pp. 653–661, 2019. [Google Scholar]
- 13.L. Hu, N. Xie, Z. Kuang, and K. Zhao, Review of cyber-physical system architecture, Proceedings—2012 15th IEEE International Symposium on Object/Component/Service-Oriented Real-Time Distributed Computing Workshops, ISORCW 2012, pp. 25–30, 2012.
- 14.Sampigethaya K. and Poovendran R., Aviation cyber–physical systems: Foundations for future aircraft and air transport, Proceedings of the IEEE, 101, (8), pp. 1834–1855, 2013. [Google Scholar]
- 15.I. S. Sacala, M. A. Moisescu, and D. Repta, Towards the development of the future internet based enterprise in the context of cyber-physical systems, in 2013 19th International Conference on Control Systems and Computer Science, 2013, pp. 405–412.
- 16.K. Sampigethaya and R. Poovendran, Cyber-physical integration in future aviation information systems, in 2012 IEEE/AIAA 31st Digital Avionics Systems Conference (DASC), 2012, pp. 7C2-1.
- 17.Banerjee A., Venkatasubramanian K. K., Mukherjee T., and Gupta S. K. S., Ensuring safety, security, and sustainability of mission-critical cyber–physical systems, Proceedings of the IEEE, 100, (1), pp. 283–299, 2011. [Google Scholar]
- 18.C. W. Axelrod, Managing the risks of cyber-physical systems, in 2013 IEEE Long Island Systems, Applications and Technology Conference (LISAT), 2013, pp. 1–6.
- 19.M. J. Stanovich et al., Development of a smart-grid cyber-physical systems testbed, in 2013 IEEE PES Innovative Smart Grid Technologies Conference (ISGT), 2013, pp. 1–6.
- 20.J. Taneja, R. Katz, and D. Culler, Defining cps challenges in a sustainable electricity grid, in 2012 IEEE/ACM Third International Conference on Cyber-Physical Systems, 2012, pp. 119–128.
- 21.M. Ghorbani and P. Bogdan, A cyber-physical system approach to artificial pancreas design, 2013 International Conference on Hardware/Software Codesign and System Synthesis, CODES+ISSS 2013, pp. 1–10, 2013.
- 22.H. Wang, X. Deng, and F. Tian, WiP abstract: A human-centered cyber-physical systematic approach for post-stroke monitoring, in 2012 IEEE/ACM Third International Conference on Cyber-Physical Systems, 2012, p. 209.
- 23.A. Banerjee and S. K. S. Gupta, Spatio-temporal hybrid automata for safe cyber-physical systems: A medical case study, in 2013 ACM/IEEE International Conference on Cyber-Physical Systems (ICCPS), 2013, pp. 71–80.
- 24.C. Sankavaram, A. Kodali, and K. Pattipati, An integrated health management process for automotive cyber-physical systems, in 2013 International Conference on Computing, Networking and Communications (ICNC), 2013, pp. 82–86.
- 25.Y. P. Fallah and R. Sengupta, A cyber-physical systems approach to the design of vehicle safety networks, in 2012 32nd International Conference on Distributed Computing Systems Workshops, 2012, pp. 324–329.
- 26.Li X. et al. , A holistic approach to service delivery in driver-in-the-loop vehicular CPS, IEEE Journal on Selected Areas in Communications, 31, (9), pp. 513–522, 2013. [Google Scholar]
- 27.M. Lukasiewycz et al., Cyber-physical systems design for electric vehicles, in 2012 15th Euromicro Conference on Digital System Design, 2012, pp. 477–484.
- 28.Schirner G., Erdogmus D., Chowdhury K., and Padir T., The future of human-in-the-loop cyber-physical systems, Computer, 46, (1), pp. 36–45, 2013. [Google Scholar]
- 29.M. Franke, C. Seidl, and T. Schlegel, A seamless integration, semantic middleware for cyber-physical systems, in 2013 10th IEEE International Conference on Networking, Sensing and Control (ICNSC), 2013, pp. 627–632.
- 30.S. El-Tawab and S. Olariu, Communication protocols in FRIEND: A cyber-physical system for traffic Flow Related Information Aggregation and Dissemination, in 2013 IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops), 2013, pp. 447–452.
- 31.A. Aminifar, P. Eles, Z. Peng, and A. Cervin, Control-quality driven design of cyber-physical systems with robustness guarantees, in 2013 Design, Automation & Test in Europe Conference & Exhibition (DATE), 2013, pp. 1093–1098.
- 32.Feng Z., Wang J., Ma Y., and Tu Y., Robust parameter design based on Gaussian process with model uncertainty, International Journal of Production Research, 0, (0), pp. 1–17, 2020. [Google Scholar]
- 33.Hu F. et al. , Robust Cyber-Physical Systems: Concept, models, and implementation, Future Generation Computer Systems, 56, pp. 449–475, 2016. [Google Scholar]
- 34.Q. Zhu, C. Rieger, and T. Başar, A hierarchical security architecture for cyber-physical systems, Proceedings—ISRCS 2011: 4th International Symposium on Resilient Control Systems, pp. 15–20, 2011.
- 35.M. J. Blondin, J. S. Sáez, and P. M. Pardalos, Control Engineering from Classical to Intelligent Control Theory—An Overview, in Computational Intelligence and Optimization Methods for Control Engineering, Springer, 2019, pp. 1–30.
- 36.Blondin M. J. and Pardalos P. M., Computational Intelligence and Optimization Methods for Control Engineering, 150, (September). 2019. [Google Scholar]
- 37.B. Ali Asghar, Computational Intelligence and Its Applications in Uncertainty-Based Design Optimization, in Bridge Optimization-Inspection and Condition Monitoring, IntechOpen, 2019.
- 38.Wang G. and Shan S., Review of Metamodeling Techniques in Support of Engineering Design Optimization, Journal of Mechanical Design, 129, (4), pp. 370–380, 2007. [Google Scholar]
- 39.Parnianifard A., Azfanizam A. S., Ariffin M. K. A., and Ismail M. I. S., Comparative study of metamodeling and sampling design for expensive and semi-expensive simulation models under uncertainty, SIMULATION, 96, (1), pp. 89–110, May 2019. [Google Scholar]
- 40.Parnianifard A. and Azfanizam A., Metamodel‐based robust simulation‐optimization assisted optimal design of multiloop integer and fractional‐order PID controller, International Journal of Numerical Modelling: Electronic Networks, Devices and Fields, 33, (1), p. e2679, 2020. [Google Scholar]
- 41.Parnianifard A., Azfanizam A. S., Ariffin M. K. A., and Ismail M. I. S., Kriging-Assisted Robust Black-Box Simulation Optimization in Direct Speed Control of DC Motor Under Uncertainty, IEEE Transactions on Magnetics, 54, (7), pp. 1–10, 2018. [Google Scholar]
- 42.Simpson T. W., Poplinski J. D., Koch P. N., and Allen J. K., Metamodels for Computer-based Engineering Design: Survey and recommendations, Engineering With Computers, 17, (2), pp. 129–150, 2001. [Google Scholar]
- 43.Li Y. F., Ng S. H., Xie M., and Goh T. N., A systematic comparison of metamodeling techniques for simulation optimization in Decision Support Systems, Applied Soft Computing, 10, (4), pp. 1257–1273, 2010. [Google Scholar]
- 44.Jin R., Du X., and Chen W., The use of metamodeling techniques for optimization under uncertainty, 25, (2). 2003. [Google Scholar]
- 45.Yondo R., Andrés E., and Valero E., A review on design of experiments and surrogate models in aircraft real-time and many-query aerodynamic analyses, Progress in Aerospace Sciences, 96, pp. 23–61, 2018. [Google Scholar]
- 46.R. Rajkumar, I. Lee, L. Sha, and J. Stankovic, Cyber-physical systems: The next computing revolution, Proceedings—Design Automation Conference, pp. 731–736, 2010.
- 47.Shah P. and Agashe S., Review of fractional PID controller, Mechatronics, 38, (January 2020), pp. 29–41, 2016. [Google Scholar]
- 48.Ranganayakulu R., Uday Bhaskar Babu G., Seshagiri Rao A., and Patle D. S., A comparative study of fractional order PIλ/PIλDµ tuning rules for stable first order plus time delay processes, Resource-Efficient Technologies, 2, pp. S136–S152, 2016. [Google Scholar]
- 49.Tepljakov A., Alagoz B. B., Yeroglu C., Gonzalez E., HosseinNia S. H., and Petlenkov E., FOPID Controllers and Their Industrial Applications: A Survey of Recent Results 1, IFAC-PapersOnLine, 51, (4), pp. 25–30, 2018. [Google Scholar]
- 50.Podlubny I., Fractional-order systems and fractional-order controllers, Institute of Experimental Physics, Slovak Academy of Sciences, Kosice, 12, (3), pp. 1–18, 1994. [Google Scholar]
- 51.M. A. Clark and K. S. Rattan, Piecewise affine hybrid automata representation of a multistage fuzzy PID controller, AAAI Spring Symposium—Technical Report, SS-14-02, pp. 104–109, 2014.
- 52.W. W. Shein, Y. Tan, and A. O. Lim, PID controller for temperature control with multiple actuators in cyber-physical home system, in 2012 15th International Conference on Network-Based Information Systems, 2012, pp. 423–428.
- 53.Wang W., Di Maio F., and Zio E., Hybrid fuzzy-PID control of a nuclear Cyber-Physical System working under varying environmental conditions, Nuclear Engineering and Design, 331, (December 2017), pp. 54–67, 2018. [Google Scholar]
- 54.Aghababa M. P., Optimal design of fractional-order PID controller for five bar linkage robot using a new particle swarm optimization algorithm, Soft Computing, 20, (10), pp. 4055–4067, 2016. [Google Scholar]
- 55.K. Miettinen, Nonlinear multiobjective optimization, 12. Springer Science & Business Media, 2012.
- 56.K. M. Miettinen, Nonlinear multiobjective optimization, 12. Springer Science {&} Business Media, 1998.
- 57.Rasmussen C. E. and Williams C. K. I., Gaussian processes for machine learning. 2006, 38, (2). 2006. [Google Scholar]
- 58.Kleijnen J. P. C., Kriging metamodeling in simulation: A review, European Journal of Operational Research, 192, (3), pp. 707–716, 2009. [Google Scholar]
- 59.J. P. C. C. Kleijnen, Design and analysis of simulation experiments (2nd). Springer, 2015.
- 60.Simpson T. W., Mauery T. M., Korte J., and Mistree F., Kriging models for global approximation in simulation-based multidisciplinary design optimization, AIAA Journal, 39, (12), pp. 2233–2241, 2001. [Google Scholar]
- 61.Parnianifard A., Azfanizam A. S., Ariffin M. K., Ismail M. I., Maghami M. R., and Gomes C., Kriging and Latin Hypercube Sampling Assisted Simulation Optimization in Optimal Design of PID Controller for Speed Control of DC Motor, Journal of Computational and Theoretical Nanoscience, 15, (5), pp. 1471–1479, 2018. [Google Scholar]
- 62.Eberhart R. and Kennedy J., Particle swarm optimization, in Proceedings of the IEEE international conference on neural networks, 1995, 4, pp. 1942–1948. [Google Scholar]
- 63.Ab Wahab M. N., Nefti-Meziani S., and Atyabi A., A Comprehensive Review of Swarm Optimization Algorithms, PLOS ONE, 10, (5), p. e0122827, May 2015. 10.1371/journal.pone.0122827 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Del Valle Y., Venayagamoorthy G. K., Mohagheghi S., Hernandez J.-C., and Harley R. G., Particle swarm optimization: basic concepts, variants and applications in power systems, IEEE Transactions on evolutionary computation, 12, (2), pp. 171–195, 2008. [Google Scholar]
- 65.Zamani M., Karimi-Ghartemani M., Sadati N., and Parniani M., Design of a fractional order PID controller for an AVR using particle swarm optimization, Control Engineering Practice, 17, (12), pp. 1380–1387, 2009. [Google Scholar]
- 66.Yu H., Tan Y., Zeng J., Sun C., and Jin Y., Surrogate-assisted hierarchical particle swarm optimization, Information Sciences, 454–455, pp. 59–72, 2018. [Google Scholar]
- 67.Dutta S., A sequential metamodel-based method for structural optimization under uncertainty, Structures, 26, (July 2019), pp. 54–65, 2020. [Google Scholar]
- 68.Regis R. G., Particle swarm with radial basis function surrogates for expensive black-box optimization, Journal of Computational Science, 5, (1), pp. 12–23, 2014. [Google Scholar]
- 69.G. Dellino, P. C. Kleijnen, Jack, and C. Meloni, Metamodel-Based Robust Simulation-Optimization: An Overview, in In Uncertainty Management in Simulation-Optimization of Complex Systems, Springer US, 2015, pp. 27–54.
- 70.Parnianifard A., Azfanizam A. S., Ariffin M. K. A., and Ismail M. I. S., Crossing weighted uncertainty scenarios assisted distribution-free metamodel-based robust simulation optimization, Engineering with Computers, 36, (1), pp. 139–150, 2019. [Google Scholar]
- 71.S. Park and J. Antony, Robust design for quality engineering and six sigma. World Scientific Publishing Co Inc, 2008.
- 72.M. S. Phadke, Quality Engineering Using Robust Design. Prentice Hall PTR, 1989.
- 73.F. Jurecka, Robust Design Optimization Based on Metamodeling Techniques, PhD Thesis, 2007.
- 74.Havinga J., van den Boogaard A. H., and Klaseboer G., Sequential improvement for robust optimization using an uncertainty measure for radial basis functions, Structural and Multidisciplinary Optimization, 55, (4), pp. 1345–1363, 2017. [Google Scholar]
- 75.Drira N. et al. , Convergence rates of the efficient global optimization algorithm for improving the design of analog circuits, Analog Integrated Circuits and Signal Processing, 103, (1), pp. 143–162, 2020. [Google Scholar]
- 76.K. Rutten, Methods For Online Sequential Process Improvement, PhD Thesis, 2015.
- 77.Kleijnen J. P. C., Van Beers W., and Van Nieuwenhuyse I., Expected improvement in efficient global optimization through bootstrapped kriging, Journal of Global Optimization, 54, (1), pp. 59–73, 2012. [Google Scholar]
- 78.Quenouille M. H., Approximate tests of correlation in time-series 3, in Mathematical Proceedings of the Cambridge Philosophical Society, 1949, 45, (3), pp. 483–484. [Google Scholar]
- 79.Tukey J., Bias and confidence in not quite large samples, Ann. Math. Statist., 29, p. 614, 1958. [Google Scholar]
- 80.R. Nisbet, J. Elder, and G. Miner, Handbook of statistical analysis and data mining-2nd. Academic Press., 2017.
- 81.McKay M. D., Beckman R. J., and Conover W. J., Comparison of three methods for selecting values of input variables in the analysis of output from a computer code, Technometrics, 21, (2), pp. 239–245, 1979. [Google Scholar]
- 82.Iman R. L. and Conover W. J., A distribution-free approach to inducing rank correlation among input variab, Communications in Statistics—Simulation and Computation, 11, (3). pp. 311–334, 1982. [Google Scholar]
- 83.Viana F. A. C., A Tutorial on Latin Hypercube Design of Experiments, Quality and Reliability Engineering International, 32, (5), pp. 1975–1985, 2016. [Google Scholar]
- 84.R. C.. Cheng, Resampling methods, Handbooks in operations research and management science, 13, pp. 415–453, 2006.
- 85.Dellino G., Kleijnen J. P. C., and Meloni C., Robust optimization in simulation: Taguchi and Krige combined, INFORMS Journal on Computing, 24, (3), pp. 471–484, 2012. [Google Scholar]
- 86.Kleijnen J. P. C. and van Beers W. C. M., Monotonicity-preserving bootstrapped Kriging metamodels for expensive simulations, Journal of the Operational Research Society, 64, (5), pp. 708–717, 2013. [Google Scholar]
- 87.A. T. Azar, J. Kumar, V. Kumar, and K. P. S. Rana, Control of a two link planar electrically-driven rigid robotic manipulator using fractional order SOFC, in International Conference on Advanced Intelligent Systems and Informatics, 2017, pp. 57–68.
- 88.T. Kathuria, V. Kumar, K. P. S. Rana, and A. T. Azar, Control of a Three-Link Manipulator Using Fractional-Order PID Controller, in Fractional Order Systems, Elsevier Inc., 2018, pp. 477–510.
- 89.Krishan G. and Singh V. R., Motion control of five bar linkage manipulator using conventional controllers under uncertain conditions, International Journal of Intelligent Systems and Applications, 8, (5), pp. 34–40, 2016. [Google Scholar]
- 90.M. W. Spong, S. Hutchinson, and M. Vidyasagar, Robot modeling and control, (Apr 13). 2020.
- 91.Tepljakov A., Petlenkov E., and Belikov J., FOMCON: a MATLAB Toolbox for Fractional-order System Identification and Control, International Journal of Microelectronics and Computer Science, 2, (2), pp. 51–62, 2011. [Google Scholar]
- 92.Badamchizadeh M. A., Hassanzadeh I., and Abedinpour Fallah M., Extended and unscented kalman filtering applied to a flexible-joint robot with jerk estimation, Discrete Dynamics in Nature and Society, 2010, 2010. [Google Scholar]
- 93.Lophaven S. N., Nielsen H. B., Søndergaard J., and Nielsen H. B., DACE—A Matlab Kriging Toolbox (Version 2.0), IMM Informatiocs and Mathematical Modelling, pp. 1–34, 2002. [Google Scholar]
- 94.Mirjalili S., Mirjalili S. M., and Lewis A., Grey Wolf Optimizer, Advances in Engineering Software, 69, pp. 46–61, 2014. [Google Scholar]
- 95.Mirjalili S., The ant lion optimizer, Advances in Engineering Software, 83, pp. 80–98, 2015. [Google Scholar]
- 96.Verma S. K., Yadav S., and Nagar S. K., Optimization of Fractional Order PID Controller Using Grey Wolf Optimizer, Journal of Control, Automation and Electrical Systems, 28, (3), pp. 314–322, 2017. [Google Scholar]
- 97.Pradhan R., Majhi S. K., Pradhan J. K., and Pati B. B., Optimal fractional order PID controller design using Ant Lion Optimizer, Ain Shams Engineering Journal, (xxxx), 2019. [Google Scholar]
- 98.J. D. Gibbons and S. Chakraborti, Nonparametric Statistical Inference: Revised and Expanded. CRC press, 2014.
- 99.W. J. Conover, Practical nonparametric statistics, 350. John Wiley & Sons, 1998.
- 100.Derrac J., García S., Molina D., and Herrera F., A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms, Swarm and Evolutionary Computation, 1, (1), pp. 3–18, 2011. [Google Scholar]
- 101.Runarsson T. P. and Yao X., Stochastic ranking for constrained evolutionary optimization, IEEE Transactions on evolutionary computation, 4, (3), pp. 284–294, 2000. [Google Scholar]
- 102.Bagheri S., Konen W., Emmerich M., and Bäck T., Self-adjusting parameter control for surrogate-assisted constrained optimization under limited budgets, Applied Soft Computing Journal, 61, pp. 377–393, 2017. [Google Scholar]
- 103.Regis R. G., Constrained optimization by radial basis function interpolation for high-dimensional expensive black-box problems with infeasible initial points, Engineering Optimization, 46, (2), pp. 218–243, 2014. [Google Scholar]
- 104.Moré J. J. and Wild S. M., Benchmarking Derivative-Free Optimization Algorithms, SIAM Journal on Optimization, 20, (1), pp. 172–191, 2009. [Google Scholar]
- 105.Wolpert D. H. and Macready W. G., No free lunch theorems for optimization, IEEE transactions on evolutionary computation, 1, (1), pp. 67–82, 1997. [Google Scholar]













