Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2012 Jul 11;2012:721867. doi: 10.1155/2012/721867

A New Stochastic Technique for Painlevé Equation-I Using Neural Network Optimized with Swarm Intelligence

Muhammad Asif Zahoor Raja 1,2,3,*, Junaid Ali Khan 1,2,3, Siraj-ul-Islam Ahmad 2,4, Ijaz Mansoor Qureshi 1,2
PMCID: PMC3418643  PMID: 22919371

Abstract

A methodology for solution of Painlevé equation-I is presented using computational intelligence technique based on neural networks and particle swarm optimization hybridized with active set algorithm. The mathematical model of the equation is developed with the help of linear combination of feed-forward artificial neural networks that define the unsupervised error of the model. This error is minimized subject to the availability of appropriate weights of the networks. The learning of the weights is carried out using particle swarm optimization algorithm used as a tool for viable global search method, hybridized with active set algorithm for rapid local convergence. The accuracy, convergence rate, and computational complexity of the scheme are analyzed based on large number of independents runs and their comprehensive statistical analysis. The comparative studies of the results obtained are made with MATHEMATICA solutions, as well as, with variational iteration method and homotopy perturbation method.

1. Introduction

The history of Painlevé equations is more than one century old. These are six second-order nonlinear irreducible equations that define new transcendental functions known as Painlevé Transcendents. These functions describe different physical processes and have been extensively used in both pure [1] and applied mathematics [2], along with theoretical physics [3]. For instance, Painlevé Transcendents are used in solutions to Korteweg-de Vries (KdV), cylindrical KdV and Bussinesq equations [4, 5], bifurcations in nonlinear non-integral models [6], matrix models of quantum gravity with continuous limits [7, 8], among others. The history, importance, and applications of Painlevé transcendents can be seen elsewhere [9, 10]. In recent publications, Painlevé equation-I (PE-I) is used in Tronquée [11] and hyperasymptotic [12] solutions, as well as in modeling the viscous shocks in Hele-shaw flow and Stokes phenomena [13]. Beside this, many researchers have employed PE-I in diverse fields of applied science and engineering [1422].

In this article, our investigation is confined to find out solution of initial value problem (IVP) of nonlinear second-order PE-I, written in the following form:

y′′(t)=6y(t)2+t,t(0,1),y(0)=0,y(0)=1. (1)

In the present study, the strength of feed forward artificial neural networks (ANNs) is exploited for approximate mathematical model of PE-I. The real strength of such model to solve the differential equations can be achieved by using modern stochastic solvers for optimization of weights based on particle swarm optimization (PSO) technique hybrid with local search methods. For example, the solution of linear and nonlinear differential equations and their systems is reported extensively by many researchers [2326]. Recently, solution of well-known fractional-order systems in engineering based on Bagley-Torvik and Riccati equations are other illustrative examples of such approaches [27, 28].

In this paper, ANNs supported with PSO and active set algorithm (ASA) are used to find the solution of IVP of PE-I. Monte-Carlo simulations of proposed scheme are performed and analyzed statistically to verify and validate its effectiveness. Comparison with standard MATHEMATICA, as well as homotopy perturbation method (HPM) [29, 30] and variational iteration method (VIM) [29, 30] results is also carried out for the given scheme.

The organization of the paper is as follows: the objective function creation with neural network mathematical modeling and its training methodology are introduced in Section 2. A brief introduction to particle swarm optimization algorithm is revealed in Section 3. A detailed application of the designed method along with some discussion on the results is presented in Section 4. Comparative analysis of the results is presented in Section 5. Last section concludes the findings along with some directions for future research.

2. Mathematical Modeling

The Feed-forward ANNs is well known to be used as a universal function approximator for any continuous function along with its derivatives on a compact set. In the modeling of differential equation through ANN methodology, following continuous mapping is used for the solution y(t), first-order derivative y′(t) and second-order derivative y′′(t), respectively, given as [3134]

y^(t)=i=1mαif(wit+βi),y^(t)=i=1mαif(wit+βi),y^′′(t)=i=1mαif′′(wit+βi), (2)

where α i, w i, and β i are real-valued bounded adaptive parameters, m is the number of neurons, and f being the activation function taken as log sigmoid for hidden layers:

f(x)=11+ex. (3)

The linear combination of the networks represented by (2) can arbitrarily model the nonlinear PE-I. It is called differential equation neural network (DE-NN) of the equation and its architecture is shown in Figure 1.

Figure 1.

Figure 1

DE-NN architecture for Painlevé equation I.

The objective or fitness function e for solving PE-I has been formulated by defining an unsupervised error, given as the sum of mean square errors:

e=e1+e2|j,j=1,2,3,,M, (4)

where j is the iteration number, M represents the total number of iterations, and the error term e 1 is related with the differential equation and is written as

e1=1N+1m=0N(y^m′′6y^m2tm)2,t(0,1),N=1h,y^m=y^(tm),tm=mh. (5)

Here, the interval t ∈ (0, 1) is divided into N number of steps t ∈ (t 0 = 0, t 1, t 2,…, t N = 1) with step size h and y^(t) and y^′′(t) are DE-NN given in set of (2).

Similarly, the error term e 2 is due to initial conditions and it is given as

e2=12(y^02+(y^01)2). (6)

It is quite clear that, with the availability of weights α i, w i, and β i in DE-NN model for which the values of functions e 1 and e 2 approach zero, the value of unsupervised error e also approaches zero. Therefore, the solution y(t) of the PE-I is approximated by y^(t) as given in (2).

3. Learning Methodology

The learning procedures are introduced here for finding the unknown weights of DE-NN networks representing PE-I using PSO and ASA algorithms.

The swarm intelligence techniques often referred to as PSO algorithm was first introduced by Kennedy and Eberhart [35]. It is a kind of global optimization technique which was based on the model of social behavior of bird flocking and fish schooling. Its discrete and continuous versions have been developed and used in different optimization problems of applied science and engineering. Few examples include the application to mobile communications, sensor networks, inventory control, multiprocessor scheduling, controls, stock market prediction, and steganographic methods [3639] and so forth.

In PSO algorithm, each single candidate solution to an optimization problem is taken as a particle in the search space. The exploration of a problem space in PSO algorithm is made with a population of particles called a swarm. All particles in the swarm have own fitness values associated with problem specific objective function. The PSO algorithm is initialized with a swarm of particles randomly and is used to search for optimal solution iteratively. The broader spread of initial swarm results in optimal performance of the algorithm. The position and the velocity of each particle are updated according to its known previous local best position P lb n−1 and the global best position of all particles P gb n−1 in the swarm so far. The updating formula for each particle velocity and position in continuous standard PSO is written as

vin=ωvin1+c1r1(Plbn1xin1)+c2r2(Pgbn1xin1)xin=xin1+vin1, (7)

where x i is vector to represent ith particle of the swarm, i = 1, 2, …, p,  p is an integer giving the total number of particles in a swarm. The v i is the velocity vector associated with ith particle, c 1 and c 2 are the local and global social acceleration constants, respectively, ω is the inertia weight linearly decreasing over the course of search between 0 and 1, and r 1 and r 2 are random vectors with elements distributed between 0 and 1.

The elements of velocity are bounded as v i ∈ [−v max⁡, v max⁡], where v max⁡ is maximum velocity. If the velocity goes beyond its maximum value, it will be set to v max⁡. This parameter has its own importance and controls the convergence rate of the algorithm. The execution of the algorithm is terminated on the criterion to maximum flights/cycles is completed or a specific value of the fitness is achieved. The generic flowchart showing the process of proposed algorithm is presented in Figure 2.

Figure 2.

Figure 2

Flowchart of particle swarm optimization aided with actives set algorithm.

The algorithm is given in the following steps.

Step 1 —

Initialization. The initial swarm of particle is generated randomly using bounded real values. Each particle represents the candidate solution with as many members as number of unknown parameters in DE-NN networks. The scattering in values of initial swarm is maintained for better search space for the algorithm. Initialize the values of parameters as given in Table 1 before execution of algorithm.

Table 1.

Parameter settings for ASA and PSO algorithms.

PSO ASA
Parameters Setting Parameters Setting
Swarm Size 160 Start Point Best Particle of PSO
Particle size 30 No. of variable 30
Flights 2000 Iteration 1000
c 1 Linear decreasing (2.5 to 0.5) Maximum function Evaluations (MaxFunEvals) 50000
c 2 Linear increasing (0.5 to 2.5) Function tolerance (TolFun) 10-18
ω Linearly decreasing (0.9 to 0.4) Nonlinear Constraints tolerance (TolCon) 10-18
v max⁡ 02 Derivative approximate Finite forward difference
Population Span (−50, 50) X-Tolerance (TolX) 10-12
Velocity Span (−2, 2) Bounds (−50, 50)

Step 2 —

Fitness Evaluation. Calculate fitness for each particle by using the fitness function given in (4), (5), and (6).

Step 3 —

Termination Criteria. Terminate the algorithm if any of following criteria satisfies the following:

  1. predefined value of fitness achieved that is, e ≤ 10−10,

  2. number of maximum flights/cycles is executed.

If yes, then go to Step 6.

Step 4 —

Ranking. Rank each particle on the basis of minimum of the fitness function values.

Step 5 —

Renewal. Update the velocity and position are updated using the equations given in (7). Repeat the algorithm from Step 2 to Step 5 until total number of flights is reached.

Step 6 —

Refinement. MATLAB built-in formulation is used for running active set algorithm (ASA) for fine tuning of the results. The global best particle of PSO is used as a start point of the algorithm and other parameter settings for ASA is also provided in Table 1.

Step 7 —

Statistical Analysis. Store the value of global best particle along with its fitness value and time of executions for algorithm. Repeat Step 1 to 6 for sufficient large number of runs for reliable statistical analysis.

4. Simulation and Results

The well-known analytic solvers like adomian decomposition method (ADM), variational iteration method (VIM), homotopy perturbation method (HPM), homotopy analysis method (HAM), and their modified versions are used extensively to provide the solutions of strong nonlinear Painlevé equations [29, 30, 4042]. The generic solution is provided in all these mentioned methods in the form of infinite series, but the results are determined using finite number of terms, whose accuracy can be enhanced with the increase in number of terms.

Before describing the designed approach, a brief description of VIM and HPM has been provided for PE-I. The approximate solution of y(t) for PE-I (1) in case of both VIM and HPM is represented in the form of a function u(t). To describe the VIM, consider the following nonlinear differential equation:

L(u)+N(u)=g(t), (8)

where L and N are linear and nonlinear operators, respectively, and g(t) is an inhomogeneous term. In standard VIM, the correct functional can be constructed as

un+1(t)=un(t)+0tλ(τ)(Lun(t)+Nu~n(τ)g(τ))dτ, (9)

here, λ(τ) represents a general Lagrangian multiplier, n and n + 1 are the present and next approximation, ũ represents restricted variation, that is, δũ = 0. With suitable choice of u 0, the fixed point of the correction functional (9) can be considered as an approximate solution of the equation and successive approximation u n+1 of the solution u will be readily obtained upon by determining the Lagrange multiplier. Consequently, the solution is given by u = lim⁡nu n. For interesting readers, the detail description of VIM along with its applicability to various kinds of differential equations can be seen in [4346].

The governing iterative formulation for solving the PE-I using VIM with the optimal value for λ(τ) = τt can be written as

un+1(t)=un(t)+0t(τt)(un′′(τ)6u~n2(τ)τ)dτ, (10)

The approximate solution u(t) by VIM for first three iterations using u 0(t) = t + (1/6)t 3 can be written as

u1(t)=t+16t3+12t4+115t6+1336t8,u2(t)=t+16t3+12t4+115t6+17t7++187764400t14+1100800t16+15757696t18,u3(t)=t+16t3+12t4+115t6+17t7++11871210325276160000t14+160939454466400t36+17768399149858816t38. (11)

On the other hand to explain the HPM, let us consider the following general equation of type:

A(u)g(t)=0,tΩ, (12)

where A represents the differential operator, Ω is the domain and g(t) is a known analytical function. The operator A can consist of linear part L and a nonlinear part N. In standard HPM, the Hopotopy v(r, p) : Ω × [0, 1] → R can be constructed which satisfies

H(v,p)=(1p)(L(v)L(u0))+p(A(v)g(t))=0, (13)

here, p ∈ [0, 1] is an embedding parameter and u 0 is an initial approximation of (12). By using small embedding parameter p, the vth solution of (12) can be given as power series in p, that is,

v=v0+pv1+p2v2+, (14)

for p = 1, the approximate solution for (12) is given as

u=limp1v=v0+v1+v2+. (15)

The method provides enough convergence in most of the cases and the rate of convergence depends on the nonlinear operator A(v). The detail description of the HPM and its applications can be seen in [4750].

Using HPM, the solution of PE-I (1) starting from u 0(t) = t + (1/6)t 3 can be given as

u=u(0)+tu(0)+16t3+60tu2dtdt. (16)

using the initial conditions (1), the homotopy constructed by the method can be given as

H(u,p)=ut16t3+6p0tu2dtdt=0. (17)

Inserting ∑i=0 p i u i instead of u and comparing the coefficients of 1, 2, 3, 4, and 5 powers of p, the approximate solutions by HPM for PE-I are, respectively, given as

u1(t)=t+16t3+12t4+115t6+1336t8,u2(t)=t+16t3+12t4+115t6+17t7++140t9+1746200t11+126208t13,u3(t)=t+16t3+12t4+115t6+17t7++52198408400t14+3551144144000t16+95224550144t18,u4(t)=t+16t3+12t4+115t6+17t7++16346914378364000t19+163451491203440000t21+14580637139501513648000000t23,u5(t)=t+16t3+12t4+115t6+17t7++404709982723087434930560000t24+14580637139501513648000000t26+748491737058836344832t28. (18)

Along with the analytical solutions given by VIM and HPM, the results are also calculated with MATHEMATICA y M(t). These results are used for comparison with the solution provided by the proposed design scheme.

Mathematical model of the equation is made with DE-NN networks given by (2) by taking 10 numbers of neurons, resulting in 30 unknown weights in DE-NN networks. The weights are taken as bounded real numbers between −50 to 50. The initial swarm consists of 160 particles, each with 30 numbers of elements, which is equivalent to the number of weights for DE-NN networks. Input of the training set is taken as t ∈ (0,1) with a step of h = 0.1. The fitness function provided in (4) can be formulated as

e=111m=010(y^m′′6y^m2tm)2+12(y^02+(y^01)2)|j,j=1,2,3,,2000. (19)

The PSO algorithm runs iteratively in order to compute the minimum of fitness function (19) and the best particle of PSO technique is passed to ASA algorithm as a start point for rapid local search. The parameter settings for the execution of algorithms are provided in Table 1.

Similarly, hundred independent runs of PSO and PSO hybrid with ASA (PSO-ASA) algorithms have been performed for finding the optimal weights. The weights obtained by the algorithms are used in (2) to obtained the proposed solutions y^(x) of PE-I. One set of weights for which the value of fitness function 1.3699 × 10−04 and 9.1210 × 10−08 for PSO and PSO-ASA, respectively, are provided in Table 2. Results calculated by these weights are provided in Tables 3 and 4 for inputs between 0 and 1 with a step of 0.1. The solutions of 3rd order VIM, 5th order HPM for the same inputs are also given in Table 3 while detailed information has been provided in Table 4. It can be seen that the value of mean absolute error (MAE) for 3rd order VIM, 5th order HPM, PSO, and PSO-ASA techniques are 4.6098 × 10−04, 7.6946 × 10−05, 9.8594 × 10−04, 1.1198 × 10−05, respectively.

Table 2.

A set of weights trained for DE-NN networks.

PSO PSO-ASA PSO PSO-ASA
w 1 11.003223968159700 −7.607757117408950 w 6 2.212628719834520 2.474582479870870
w 2 1.215553756976230 −1.490826886546450 w 7 −3.435984020017740 1.576108544205200
w 3 −0.105724097992031 −0.545895962495148 w 8 3.162425319231580 −3.958973407625250
w 4 −0.631338164295177 7.264468183878910 w 9 0.783489331068287 1.998216516692480
w 5 5.142485647745030 −8.830221260895670 w 10 −1.369123625736370 0.021432666153674

α 1 2.212942902261790 1.227208825263930 α 6 1.400872158822050 1.316114527930020
α 2 8.952236157349910 −3.632183934646160 α 7 −0.420235540017917 1.323004693243890
α 3 0.211094325564814 4.680032352849500 α 8 0.771318701716261 −2.560496480052050
α 4 −0.918096490077305 9.995894668412700 α 9 −1.394533732778610 1.004134898951920
α 5 34.739272824112700 −0.083663885403684 α 10 −0.343093197542097 0.887019181878860

β 1 −14.756692394364600 8.897738213117890 β 6 1.201391820838980 1.100068363593600
β 2 −3.441636802273880 2.675860058162530 β 7 3.935692917406700 0.587872446048014
β 3 2.708026281856620 −0.220415897788109 β 8 −3.847728945588120 5.075428578185270
β 4 1.986093010073140 −9.999999632285810 β 9 −1.648848649331190 −2.429219273541780
β 5 −9.056219460761080 8.881048710700930 β 10 −0.531535793135294 2.174973763274050

Table 3.

Comparison of the results for solution of Painlevé I.

T y M(t) Results for u(t) Proposed Results y^(t)
VIM HPM PSO PSO-ASA
0.1 0.1002167469 0.1002167477 0.1002167477 0.101267314 0.100211733
0.2 0.2021394539 0.2021394527 0.2021394527 0.202944631 0.202137722
0.3 0.3086307548 0.3086307489 0.3086307492 0.309301843 0.308629917
0.4 0.4239862999 0.4239862788 0.4239862895 0.424625376 0.423985302
0.5 0.5543401416 0.5543399112 0.5543401182 0.555018673 0.554341815
0.6 0.7084621313 0.7084596600 0.7084620603 0.709234588 0.708467300
0.7 0.8992500131 0.8992296942 0.8992493803 0.900159983 0.899254787
0.8 1.1465318491 1.1463982509 1.1465240263 1.147598102 1.146541253
0.9 1.4825246345 1.4817789520 1.4824439956 1.483815424 1.482531252
1.0 1.9631285609 1.9594210423 1.9624483064 1.965104083 1.963204324

Table 4.

Comparison of the results for solution of Painlevé I.

T u k(t) for VIM u k(x) for HPM DE-NN
k = 1 k = 2 k = 3 k = 2 k = 3 k = 4 k = 5 PSO PSO-ASA
0.1 1.35E − 08 8.00E − 10 8.00E − 10 7.96E − 10 8.00E − 10 8.00E − 10 8.00E − 10 1.05E − 03 5.01E − 06
0.2 1.85E − 06 1.21E − 09 1.19E − 09 4.88E − 09 1.19E − 09 1.19E − 09 1.19E − 09 8.05E − 04 1.73E − 06
0.3 3.20E − 05 9.34E − 09 5.85E − 09 2.22E − 07 6.96E − 09 5.62E − 09 5.61E − 09 6.71E − 04 8.38E − 07
0.4 2.45E − 04 1.44E − 07 2.11E − 08 3.94E − 06 6.87E − 08 1.12E − 08 1.04E − 08 6.39E − 04 9.98E − 07
0.5 1.20E − 03 2.23E − 06 2.30E − 07 3.79E − 05 1.12E − 06 5.31E − 08 2.34E − 08 6.79E − 04 1.67E − 06
0.6 4.50E − 03 2.23E − 05 2.47E − 06 2.45E − 04 1.24E − 05 6.38E − 07 7.10E − 08 7.72E − 04 5.17E − 06
0.7 1.40E − 02 1.62E − 04 2.03E − 05 1.21E − 03 9.75E − 05 7.55E − 06 6.33E − 07 9.10E − 04 4.77E − 06
0.8 3.84E − 02 9.20E − 04 1.34E − 04 4.97E − 03 5.98E − 04 6.89E − 05 7.82E − 06 1.07E − 03 9.40E − 06
0.9 9.63E − 02 4.39E − 03 7.46E − 04 1.78E − 02 3.05E − 03 5.02E − 04 8.06E − 05 1.29E − 03 6.62E − 06
1.0 2.27E − 01 1.84E − 02 3.71E − 03 5.74E − 02 1.36E − 02 3.07E − 03 6.80E − 04 1.98E − 03 7.58E − 05

Mean 3.82E − 02 2.39E − 03 4.61E − 04 8.16E − 03 1.73E − 03 3.65E − 04 7.69E − 05 9.86E − 04 1.12E − 05

The values of minimum (MIN), maximum (MAX), mean, and standard deviation (STD), for absolute error (AE) |yM(t)-y^(t)|, are calculated on the basis hundred independent runs for PSO and PSO-ASA methods. Results are provided in the Table 5 for inputs 0 and 1 with step of 0.1 for PE-I. It can be seen that values of AE on the basis of MIN are of the order 10−03 to 10−04 and 10−06 to 10−08 for PSO and PSO-ASA algorithms, while the values of AE based on mean lies in the range 10−01 to 10−02 and 10−04 to 10−05, respectively, for said solvers. The lowest values of statistical parameters are found by hybrid technique PSO-ASA than that of PSO algorithm.

Table 5.

Statistical analysis of results based on values of absolute error.

t PSO PSO-ASA
MIN MAX MEAN STD MIN MAX MEAN STD
0.1 1.0506E − 03 2.2340E − 01 9.2274E − 02 5.1432E − 02 3.1813E − 08 2.1601E − 03 9.6880E − 05 2.7723E − 04
0.2 8.0518E − 04 2.4845E − 01 1.0033E − 01 5.5966E − 02 3.4372E − 08 2.2102E − 03 9.7567E − 05 2.7860E − 04
0.3 6.7109E − 04 2.7628E − 01 1.1029E − 01 6.2147E − 02 8.3812E − 07 2.1940E − 03 1.0290E − 04 2.8024E − 04
0.4 6.3908E − 04 3.2281E − 01 1.2356E − 01 7.0441E − 02 1.2269E − 07 2.3203E − 03 1.1739E − 04 3.0450E − 04
0.5 6.7853E − 04 4.2304E − 01 1.4202E − 01 8.1644E − 02 3.0549E − 07 2.7264E − 03 1.3725E − 04 3.5327E − 04
0.6 7.7246E − 04 5.4096E − 01 1.6843E − 01 9.7069E − 02 9.4072E − 07 3.3844E − 03 1.6471E − 04 4.2624E − 04
0.7 9.0997E − 04 6.8783E − 01 2.0732E − 01 1.1879E − 01 4.7794E − 08 4.2143E − 03 2.1027E − 04 5.3421E − 04
0.8 1.0663E − 03 8.8150E − 01 2.6607E − 01 1.5005E − 01 5.0503E − 07 5.4046E − 03 2.7398E − 04 6.8757E − 04
0.9 1.2908E − 03 1.1516E + 00 3.5688E − 01 1.9625E − 01 4.0723E − 06 7.5318E − 03 3.8273E − 04 9.5844E − 04
1.0 1.9755E − 03 1.5506E + 00 5.0247E − 01 2.6814E − 01 2.9828E − 06 1.0874E − 02 5.6390E − 04 1.3640E − 03

The accuracy and convergence of the algorithms are analyzed further by calculating the values of MAE and fitness achieved (FA) for 100 independents runs. Results are shown graphically on semilog scale in order to elaborate the difference in the values. The value of FA and MAE against number of independent runs are shown graphically in Figure 3(a) for PSO and PSO-ASA methods. The numbers of independent runs are arranged on the basis of ascending order of the fitness achieved, as well as values of MAE and are plotted in Figure 3(b). On the basis of acceptability criteria of MAE ≤ 10−03, the convergence for PSO and PSO-ASA is 1% and 100%, respectively.

Figure 3.

Figure 3

Comparison of results on the basis of fitness achieved (FA) and mean absolute error (MAE).

5. Comparative Analysis of Results

In this section, the comparative study for results is presented based on values of MAE, execution time (ET), mean fitness achieved (MFA), global mean absolute error (GMAE), number of grid points, and convergence rate for PSO and PSO-ASA proposed solvers.

The effects on results by changing the number of gird points used in fitness function (19) are analyzed. By taking the number of grid points 21 instead of 11, the new fitness function is formulated using the input of the training set t ∈ (0,1) with a step of h = 0.05 and given as

e=121m=020(y^m′′6y^m2tm)2+12(y^02+(y^01)2)|j,j=1,2,3,,2000. (20)

Hundred independents runs of PSO and PSO-ASA method are executed for finding the appropriate weights and subsequently the solution of IVP of PE-1. The MFA is defined as average values of fitness achieved taken over all independent runs of the algorithm. Similarly, GMAE is average value of mean absolute error for all independents runs of solvers. The values of MFA and GMAE are calculated for PSO and PSO-ASA techniques using both 11 and 21 grids points and results are tabulated in Table 6. It is found that the values of MFA are of the order 10−02 and 10−06 for PSO and PSO-ASA, respectively. The values of MAE are matched with MATHEMATICA solution y M(t) upto 2 to 3 and 5 to 8 decimal places for PSO and PSO-ASA techniques. The value of GMAE for 11 and 21 grids points are determined and given in Table 6. It is found that values of GMAE are of the order 10−04 for both the cases for inputs between 0 and 1 with a step of h = 0.1. So, by increasing number of grid points the same level of the accuracy in results is maintained for both PSO and PSO-ASA techniques. The increase in number of grids points, that is, decrease in the mesh size h, does not guaranty improvement of results for stochastic solvers and same is observed in our simulations by changing the value of step size h = 0.1 to h = 0.05.

Table 6.

Comparative analysis of the results.

Parameters h = 0.1 h = 0.05
PSO PSO-ASA PSO PSO-ASA
GMAE 1.9576E − 01 2.0372E − 04 1.8444E − 01 6.3211E − 04
MF 1.1837E − 02 4.6577E − 06 1.0804E − 02 7.1294E − 06
MAE ≤ 10-02 21% 100% 22% 100%
MAE ≤ 10-03 01% 100% 00% 099%
MAE ≤ 10-04 00% 095% 00% 083%
FA ≤ 10-03 50% 100% 55% 100%
FA ≤ 10-04 05% 100% 05% 100%
FA ≤ 10-05 01% 099% 00% 098%
MIN-ET 075.23 s 079.29 s 132.42 s 157.14 s
MAX-ET 103.15 s 113.74 s 145.67 s 191.32 s
M-ET 076.93 s 092.17 s 142.43 s 183.57 s
STD-ET 002.93 s 007.32 s 002.02 s 007.62 s

The convergence rates of the proposed stochastic solvers are determined based on the value of MAE ≤10−02, 10−03, and 10−04, as well as the value of MFA e ≤ 10−03, 10−04, and 10−05. Results are tabulated in Table 6 for both PSO and PSO-ASA techniques. It can be seen that, with the acceptability of the results based on value of MFA e ≤ 10−05, the hybrid approach PSO-ASA achieved this criteria for 99% and 98% of the independents runs for 11 and 21 numbers of grid points, respectively. In addition to that, results are consistently convergent for larger number of independents runs for PSO-ASA method than that of PSO algorithms.

Moreover, the computational complexity is analyzed by calculating the time of execution of each algorithm for 100 numbers of independent runs for both h = 0.1 and h = 0.05. The values for minimum execution time (MIN-ET), maximum execution time (MAX-ET), mean execution time (M-ET), and standard deviation of execution time (STD-ET) are provided in the Table 6. It can be seen that values of M-ET for PSO and PSO-ASA techniques are 76.93 seconds and 92.17 seconds, respectively, for h = 0.1, and almost double values of M-ET in case for h = 0.05. The time taken for the execution for hybrid approach PSO-ASA is relatively greater than that of PSO algorithm, but this can be overshadow by its invariable dominance in accuracy and convergence. The time analysis provided in this article is carried out using Dell Latitude D630 laptop with Intel(R) Core(TM) 2 Duo CPU T9300, 2.50 GHz, and running with MATLAB version R2009b.

6. Conclusion

On the basis of the simulations and their comparative studies provided in the last sections, it can be concluded that the IVP of PE-I can be solved effectively by the designed computational intelligence algorithms using neural networks supported with PSO and ASA techniques. Results obtained using the weights for DE-NN networks of PE-I trained by PSO-ASA algorithm are found to be better than from PSO techniques. The values of absolute error from MATHEMATICA solution y M(t) of PSO-ASA lies in the range 10−04 to 10−07 and match upto 5 to 7 decimal places with state of art analytical techniques of VIM and HPM.

On the basis of 100 independent runs for each algorithm to solve PE-I using 11 and 21 grid points, the PSO-ASA technique is found to be invariable superior than PSO algorithm using the criteria based on the values of mean absolute error, global mean absolute error, mean fitness achieved, and convergence rate. However, as far as computational complexity is concerned that PSO-ASA technique is slightly more computational exhaustive, but it can be overlook due to its overall dominance in the performance. This study can also be extended to use other biological inspired computational algorithms for solving the PE-I, as well as other Painlevé transcendents.

Acknowledgment

The authors gratefully acknowledge Professor Dr. Adiqa Kausar Kiani for providing her valuable suggestions to improve the presentation of the paper.

References

  • 1.Bassom AP, Clarkson PA, Hicks AC. Numerical studies of the fourth Painlevé equation. IMA Journal of Applied Mathematics. 1993;50(2):167–139. [Google Scholar]
  • 2.He J. Variational iteration method—a kind of non-linear analytical technique: some examples. International Journal of Non-Linear Mechanics. 1999;34(4):699–708. [Google Scholar]
  • 3.He JH. Some asymptotic methods for strongly nonlinear equations. International Journal of Modern Physics B. 2006;20(10):1141–1199. [Google Scholar]
  • 4.Ablowitz MJ, Segur H. Solitons and the Inverse Scattering Transform. Philadelphia, Pa, USA: SIAM; 1981. (Studies in Applied Mathematics 4). [Google Scholar]
  • 5.Tajiri M, Kawamoto S. Reduction of KdV and cylindrical KdV equations to Painlevé equation. Journal of the Physical Society of Japan. 1982;51(5):1678–1681. [Google Scholar]
  • 6.Haberman R. Slowly varying jump and transition phenomena associated with algebraic bifurcation problems. SIAM Journal on Applied Mathematics. 1979;37(1):69–106. [Google Scholar]
  • 7.Fokas AS, Its AR, Kitaev AV. Discrete Painlevé equations and their appearance in quantum gravity. Communications in Mathematical Physics. 1991;142(2):313–344. [Google Scholar]
  • 8.Fokas AS, Its AR, Kitaev AV. Matrix models of two-dimensional quantum gravity and isomonodromic solutions of “Discrete Painlevé”equations. Journal of Mathematical Sciences. 1995;73(4):415–429. [Google Scholar]
  • 9.Ablowitz MJ, Clarkson PA. Solitons, Nonlinear Evolution Equation and Inverse Scattering. Cambridge University Press; 1991. [Google Scholar]
  • 10.Gromak VI, Laine I, Shimomura S. Painleve Differential Equations in the Complex Plane. New York, NY, USA: Walter de Gruyter; 2002. [Google Scholar]
  • 11.Dai D, Zhang L. On tronquée solutions of the first Painlevé hierarchy. Journal of Mathematical Analysis and Applications. 2010;368(2):393–399. [Google Scholar]
  • 12.Daalhuis ABO. Hyperasymptotics for nonlinear odes ii. The first Painlevé equation and a second-order riccati equation. Proceedings of the Royal Society A. 2005;461(2062):3005–3021. [Google Scholar]
  • 13.Lee SY, Teodorescu R, Wiegmann P. Viscous shocks in hele-shaw flow and stokes phenomena of the Painlevé I transcendent. Physica D. 2011;240(13):1080–1091. [Google Scholar]
  • 14.Kudryashov NA. The first and second Painlevé equations of higher order and some relations between them. Physics Letters, Section A. 1997;224(6):353–360. [Google Scholar]
  • 15.Kudryashov NA. Amalgamations of the Painlevé equations. Journal of Mathematical Physics. 2003;44(12):6160–6178. [Google Scholar]
  • 16.Cresswell G, Joshi N. The discrete first, second and thirty-fourth Painlevé hierarchies. Journal of Physics A. 1999;32(4):655–669. [Google Scholar]
  • 17.Gordoa PR, Pickering A. Nonisospectral scattering problems: a key to integrable hierarchies. Journal of Mathematical Physics. 1999;40(11):5749–5786. [Google Scholar]
  • 18.Muğan U, Jrad F. Painlevé test and the first Painlevé hierarchy. Journal of Physics A. 1999;32(45):7933–7952. [Google Scholar]
  • 19.Mugan U, Jrad F. Painlevé test and higher order differential equations. Journal of Nonlinear Mathematical Physics. 2002;9(3):282–310. [Google Scholar]
  • 20.Fornberg B, Weideman JAC. A Numerical methodology for the Painlevé equations. Journal of Computational Physics. 2011;230(15):5957–5973. [Google Scholar]
  • 21.Kudryashov NA. Fourth-order analogies to the Painlevé equations. Journal of Physics A. 2002;35(21):4617–4632. [Google Scholar]
  • 22.Kudeyashov NA. One method for finding exact solutions of nonlinear differential equations. Communications in Nonlinear Science and Numerical Simulations. 2012;17(6):2248–2253. [Google Scholar]
  • 23.Saidi H, Khelil N, Hassouni S, Zerarka A. Energy spectra of the Schrödinger equation and the differential quadrature method: improvement of the solution using particle swarm optimization. Applied Mathematics and Computation. 2006;182(1):559–566. [Google Scholar]
  • 24.Lee ZY. Method of bilaterally bounded to solution blasius equation using particle swarm optimization. Applied Mathematics and Computation. 2006;179(2):779–786. [Google Scholar]
  • 25.Raja MAZ, Qureshi IM, Khan JA. Swarm Intelligent optimized neural networks for solving fractional differential equations. International Journal of Innovative Computing. 2011;7(11):6301–6318. [Google Scholar]
  • 26.Raja MAZ, Khan JA, Qureshi IM. Heuristic computational approach using swarm intelligence in solving fractional differential equations. Proceedings of the 12th Annual Genetic and Evolutionary Computation Conference (GECCO'10); July 2010; Portland, Oregon. pp. 2023–2026. [Google Scholar]
  • 27.Raja MAZ, Khan JA, Qureshi IM. A new stochastic approach for solution of riccati differential equation of fractional order. Annals of Mathematics and Artificial Intelligence. 2010;60(3):229–250. [Google Scholar]
  • 28.Raja MAZ, Khan JA, Qureshi IM. Swarm intelligence optimized neural network for solving fractional order system of bagely-tervik equation. Engineering Intelligent System. 2011;19(1):41–51. [Google Scholar]
  • 29.Hesameddini E, Peyrovi A. The use of variational iteration method and homotopy perturbation method for Painlevé equation I. Applied Mathematical Sciences. 2009;3(37–40):1861–1871. [Google Scholar]
  • 30.Behzadi SS. Convergence of iterative methods for solving Painlevé equation. Applied Mathematical Sciences. 2010;4(30):1489–1507. [Google Scholar]
  • 31.Parisi DR, Mariani MC, Laborde MA. Solving differential equations with unsupervised neural networks. Chemical Engineering and Processing: Process Intensification. 2003;42(8-9):715–721. [Google Scholar]
  • 32.Shirvany Y, Hayati M, Moradian R. Multilayer perceptron neural networks with novel unsupervised training method for numerical solution of the partial differential equations. Applied Soft Computing Journal. 2009;9(1):20–29. [Google Scholar]
  • 33.Khan JA, Raja MAZ, Qureshi IM. Numerical treatment of nonlinear emden-fowler equation using stochastic technique. Annals of Mathematics and Artificial Intelligence. 2011;63(2):185–207. [Google Scholar]
  • 34.Khan JA, Raja MAZ, Qureshi IM. Novel approach for van der pol oscillator on the continuous time domain. Chinese Physics Letters. 2011;28(11)110205 [Google Scholar]
  • 35.Kennedy J, Eberhart R. Particle swarm optimization. Proceedings of the IEEE International Conference on Neural Networks; 1995; Perth, Australia. pp. 1942–1948. [Google Scholar]
  • 36.Sivanandam SN, Visalakshi P. Multiprocessor scheduling using hybrid particle swarm optimization with dynamically varying inertia. International Journal of Computer Science & Applications. 2007;4(3):95–106. [Google Scholar]
  • 37.Li GD, Masuda S, Yamaguchi D, Nagai M. The optimal GNN-PID control system using particle swarm optimization algorithm. International Journal of Innovative Computing, Information and Control. 2009;5(10):3457–3469. [Google Scholar]
  • 38.Araújo RDA. Swarm-based translation-invariant morphological prediction method for financial time series forecasting. Information Sciences. 2010;180(24):4784–4805. [Google Scholar]
  • 39.Li X, Wang J. A steganographic method based upon JPEG and particle swarm optimization algorithm. Information Sciences. 2007;177(15):3099–3109. [Google Scholar]
  • 40.Ellahi R, Abbasbandy S, Hayat T, Zeeshan A. On comparison of series and numerical solutions for second Painlevé equation. Numerical Methods for Partial Differential Equations. 2010;26(5):1070–1078. [Google Scholar]
  • 41.Dehghan M, Shakeri F. The numerical solution of the second Painlevé equation. Numerical Methods for Partial Differential Equations. 2009;25(5):1238–1259. [Google Scholar]
  • 42.Hesameddini E. Homotopy perturbation method for second Painlevé equation and comparisons with analytic continuation extension and chebishev series method. International Mathematical Forum. 2010;5(13):629–637. [Google Scholar]
  • 43.He JH. A short remark on fractional variational iteration method. Physics Letters. 2011;375(38):3362–3364. [Google Scholar]
  • 44.Dehghan M, Yousefi SA, Lotfi A. The use of He’s variational iteration method for solving the telegraph and fractional telegraph equations. International Journal for Numerical Methods in Biomedical Engineering. 2011;27(2):219–231. [Google Scholar]
  • 45.Wazwaz AM, Rach R. Comparison of the adomian decomposition method and the variational iteration method for solving the lane-emden equations of the first and second kinds. Kybernetes. 2011;40(9):1305–1318. [Google Scholar]
  • 46.He JH. Notes on the optimal variational iteration method. Applied Mathematics Letters. 2012;25(10):1579–1581. [Google Scholar]
  • 47.He JH. Homotopy perturbation method with an auxiliary term. Abstract and Applied Analysis. 2012;2012:7 pages.857612 [Google Scholar]
  • 48.Ji-Huan HE. A note on the homotopy perturbation method. Thermal Science. 2010;14(2):565–568. [Google Scholar]
  • 49.Öziş T, Yildirim A. Comparison between adomian’s method and He’s homotopy perturbation method. Computers and Mathematics with Applications. 2008;56(5):1216–1224. [Google Scholar]
  • 50.Feng X, He Y, Meng J. Application of homotopy perturbation method to the bratu-type equations. Topological Methods in Nonlinear Analysis. 2008;31(2):243–252. [Google Scholar]

Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES