Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2024 Oct 12;14:23874. doi: 10.1038/s41598-024-74600-4

Data-driven solutions and parameter estimations of a family of higher-order KdV equations based on physics informed neural networks

Jiajun Chen 1, Jianping Shi 1,, Ao He 1, Hui Fang 1
PMCID: PMC11470932  PMID: 39396058

Abstract

Physics informed neural network (PINN) demonstrates powerful capabilities in solving forward and inverse problems of nonlinear partial differential equations (NLPDEs) through combining data-driven and physical constraints. In this paper, two PINN methods that adopt tanh and sine as activation functions, respectively, are used to study data-driven solutions and parameter estimations of a family of high order KdV equations. Compared to the standard PINN with the tanh activation function, the PINN framework using the sine activation function can effectively learn the single soliton solution, double soliton solution, periodic traveling wave solution, and kink solution of the proposed equations with higher precision. The PINN framework using the sine activation function shows better performance in parameter estimation. In addition, the experiments show that the complexity of the equation influences the accuracy and efficiency of the PINN method. The outcomes of this study are poised to enhance the application of deep learning techniques in solving solutions and modeling of higher-order NLPDEs.

Keywords: Data-driven solutions, Parameter estimations, Higher-order KdV equations, Physical informed neural network, Sine activation function

Subject terms: Applied mathematics, Computational science, Computer science, Mathematics and computing, Physics

Introduction

Nonlinear partial differential equations (NLPDEs) constitute an important branch of modern mathematics. Many significant problems in natural sciences and engineering can be formulated as the study of nonlinear partial differential equations, both in theoretical research and practical applications. NLPDEs play a crucial role in fields such as mechanics, control processes, ecosystems, economics and finance, chemical cycle systems, and epidemiology. As research progresses, scientists have discovered that nonlinear effects must also be taken into account for some problems previously treated with linear differential equations.

The study of NLPDEs can generally be categorized into three major types: quantitative research, qualitative research, and numerical solution research. Quantitative research aims to find exact analytic solutions of equations with certain scientific significance. Due to the complexity caused by nonlinearity and other factors, there is no universally effective method for the exact analytic solution of NLPDEs. Common methods currently used include inverse scattering transform, Bäcklund transformation, Darboux transformation, Hirota bilinear method, Painlevé finite expansion method, Lie group method, and so on16. In recent years, there have been many new and valuable achievements in the field of analytical solutions of NLPDEs711. Qualitative research analyzes the existence, uniqueness, periodic branching, chaos, and other characteristics of the models without obtaining analytic solutions. Numerical solution research, on the other hand, verifies the correctness of theoretical analysis through numerical solutions when no analytical solution is available, making it another important research branch of NLPDEs. Commonly used methods include finite difference methods12, finite element methods13, and spectral methods14, etc.

In the mid-1980s, the back-propagation algorithm and its development caused the second wave of research in the field of artificial neural networks(ANN)15. ANN was applied in many fields and showed their powerful capabilities in the last half century1620. In 1981, a paper published by Rall proposed the automatic differential method21, which can use the chain rule to accurately calculate the derivative and can differentiate the entire neural network model according to the input coordinates and network parameters of the neural network, thus replacing the complex gradient calculation in the partial differential equation, and laying the foundation for the solution of the partial differential equation based on artificial neural network. Since the 1990s, scholars began to study the mathematical theories and methods of solving partial differential equations using neural networks and achieved a series of valuable results2226. The data-driven artificial neural network solving algorithm relies on data labels, leading to some problems such as poor generalization ability. Scholars have begun to explore network frameworks that combine data-driven approaches with the structured prior information implicit in partial differential equations, known as the method of physics-informed constraints. The deep Galerkin method proposed by Sirignamo et al.27 is the first attempt in this direction, and they provide the theorem for neural network approximation under physical constraints. Raissi et al.28 introduced the physics informed neural network (PINN) under physical constraints, combining data-driven and physical constraint methods to address the weak generalization ability of data-driven approaches and proposing a new approach for modeling and solving partial differential equations.

The PINN algorithm has attracted wide attention from researchers since it was proposed2931. For example, Chen’s team focuses on the local wave phenomenon in nonlinear integrable systems, and they successfully obtained data-driven solutions for local waves using the PINN algorithm3234. Yan’s team uses the PINN algorithm to study several variants of the nonlinear Schrödinger equation (NLS), including the Hirota equation, and explored the data-driven solutions of these equations and the associated inverse problems3537. Mishra et al. gave data-driven solutions for some higher-order nonlinear dispersion equations using the PINN algorithm and provided a theoretical analysis of the error estimates of the equations38. Some research on the improvement of the PINN algorithm has also received attention39.

The Korteweg-de Vries (KdV) family of equations is an important NLPDE. The original equation was derived by Korteweg and de Vries in 1895 to describe shallow water waves with long wavelengths and small amplitudes40. The KdV equation simulates various important finite amplitude dispersion wave phenomena in nonlinear evolution equations. It is also used to describe some important physical phenomena, such as acoustic waves in harmonic crystals and ion-acoustic waves in plasmas. As research progresses, a batch of new KdV equations has been obtained by modifying and extending the third-order original KdV equation. Additionally, higher-order KdV equations of order 5 and above have appeared in scientific applications and practical problems. These models have been favored by scholars, and researchers in the field of mathematics and physics have carried out in-depth research on these models and obtained a large number of valuable results4147.

The main work in this paper is to obtain predictive solutions and predictive parameters of several higher-order KdV equations by constructing a suitable PINN framework and obtaining a reliable network through data training. Previous studies have found that using the sine function as the activation function can solve some problems that exist in PINN with the traditional tanh function, for example, the asymptotic flattening output, and can also improve the convergence speed to a certain extent48. Inspired by this, an improvement strategy is proposed to evaluate the performance of the standard PINN, i.e., replacing tanh with sine as activation function in the PINN algorithm. A family of KdV equations, including 5 equations, is studied on two PINN frameworks with sine and tanh activation functions, respectively. Seven distinguishing solutions and six parameters are obtained successfully on the two proposed PINN frameworks, and the experimental results are mainly presented in the form of graphs and tables. Furthermore, we explore the effect of PINN on fitting the solutions and the unknown parameters of NLPDEs based on these experiment results. These results indicate that the sine activation function really has better performance in both predictive solutions and parameter estimation problems. We also analyze the influence of the complexity of equations and solutions on the data-driven problem by the experimental results.

The structure of this paper is as follows: In “Main ideas of the methods” section, we present a detailed description of the PINN algorithm, discussing its application in solving forward and inverse problems of NLPDEs. We also analyze and discuss three different activation functions and account for the details of the computational platform. In “Data-driven solution” section, by comparing the standard PINN algorithm and the PINN algorithm using sine function as activation function, we conduct a comprehensive study on five high-order Korteweg-de Vries (KdV) equations: a three-order KdV equation, a K(2, 2) equation, a mKdV equation, a five-order KdV equation (CDG) and a seven-order KdV equation (Lax’s). In “Data-driven parameter estimation of NLPDEs” section, the PINN algorithm with sine activation function shows its obvious advantage in the inverse problem of the NLPDEs. We also analyzed the possible cause of the special case. Finally, in “Discussion and Conclusion” section, we provide a summary of the work in this paper and give future research directions.

Main ideas of the methods

We discuss the following general form of Inline graphic dimensional partial differential equation

graphic file with name M2.gif 1

together with the initial and boundary conditions

graphic file with name M3.gif 2

where u(xt) represents the undetermined solution, Inline graphic is a nonlinear differential operator, and Inline graphic is a subset of Inline graphic.

PINN for solving forward and inverse problems of NLPDEs

PINN is a type of neural network whose central idea is to encode a model, such as a PDE, as a component of the neural network itself. A L layers feed-forward neural network consisting of an input layer, L-1 hidden layers, and an output layer is considered, where the lth layer (l= 1,2,... L − 1) has Inline graphic neurons. The connection between the two layers is established by a linear transformation Inline graphic and nonlinear activation function Inline graphic, then the input of layer L can be shown as:

graphic file with name M10.gif 3

where the input data Inline graphic is the transpose of the vector Inline graphic, Inline graphic represents the output of layer l, Inline graphic represents the input of layer l, and the weight matrix Inline graphic and bias vector Inline graphic are trainable parameters. Let Inline graphic represents the set of parameters to be learned in all the weight matrices and bias vectors of the network. Therefore, the connection relationship between input Inline graphic and output Inline graphic is

graphic file with name M20.gif 4

In “Data-driven solution” section, the standard PINN method and the improved method are used to compare the accuracy and efficiency in solving five KdV equations. The two methods use the hyperbolic tangent (Inline graphic) function and sine function as activation function respectively. The weight matrix and bias vector are initialized by Xavier49. The automatic differentiation21 is used to find the derivative of u(xt) with respect to time t and space x. The Adam algorithm50 or the L-BFGS-B algorithm51 is mainly used to update the parameters, and minimize the loss function of the mean square error (Inline graphic)

graphic file with name M23.gif 5

where

graphic file with name M24.gif 6

and

graphic file with name M25.gif 7

The loss Inline graphic corresponds to the initial and boundary data while Inline graphic enforces the structure imposed by Eq. (1) at a finite set of collocation points. Inline graphic denote the initial and boundary training data on u(xt) and Inline graphic specify the collocations points for f(xt), which are usually obtained by the Latin hypercube sampling52 and automatic differentiation techniques, respectively. To clarify the whole procedure of PINN, we state the main steps of PINN in Table 1 and depict the schematic diagram in Fig. 1.

Table 1.

PINN algorithm.

Algorithm1: PINN
Step1: Specification of training set
Training data: Inline graphic network Inline graphic.
Residual training points: f network Inline graphic.
Step2: Construct neural network Inline graphic with random initialization of parameters Inline graphic.
Step3: Construct the residual neural network f by substituting surrogate Inline graphic into
the governing equations using automatic differentiation and other arithmetic operations.
Step4: Specification of loss function:
Inline graphic
Step5: Find the best parameters using suitable optimization method for minimizing the loss function
Inline graphic

Fig. 1.

Fig. 1

Flowchart of PINN.

On the other hand, for the inverse problem, the parameterized NLPDEs is presented as the general form

graphic file with name M38.gif 8

where u(xt) represents the undetermined (or latent) solution, Inline graphic is a nonlinear operator with unknown parameter Inline graphic (Inline graphic is a parameter or a vector consisting of several parameters), and Inline graphic is a subset of Inline graphic. The loss function for the inverse problem of NLPDEs is given by

graphic file with name M44.gif 9

where

graphic file with name M45.gif 10

and

graphic file with name M46.gif 11

Here, Inline graphic is the dataset generated by u(xt) using the LHS technique in the specified region and used as the training dataset for Inline graphic and Inline graphic .The network structure and approaches of every calculation step is same as forward problem.

Activation function

The activation function introduces a nonlinear mapping to the neural network, which enables the neural network to learn and represent more complex functional relationships. In the absence of an activation function, the combination of multiple linear layers still produce only linear transformations, which limits the expressive power of neural networks. Common activation functions include the Sigmoid function, the ReLu function, the Tanh function, etc53. Depending on the specific task and the requirements of the network structure, selecting a suitable activation function can have a significant impact on the performance and effectiveness of the neural network. This section discusses the distinction among the three common activation functions: the Sigmoid function, the Tanh function, and the Sine function.

Sigmoid(x)

The sigmoid function, also known as the Logistic function, is one of the most commonly used activation functions. It is often used in binary classification tasks and non-mutually exclusive multi-category prediction. The sigmoid function has the nice property of mapping the input from negative infinity to positive infinity to a range between 0 and 1. Since the probability values lie exactly between 0 and 1, the sigmoid function is often used in tasks that require the output of probability values. The specific mathematical expression for the function is given in Eq. (12)

graphic file with name M50.gif 12

from this expression and the image of the function shown in Fig. 2, it is clear that the sigmoid function (abbreviated as S(x)) is continuously smooth and derivable. Its derivative can easily be expressed directly using the function itself without additional calculations,i.e., Inline graphic. However, as can be seen from the image and the mathematical expression of the derivative, when the function value is close to 0 or 1, the derivative value is close to 0, which leads to the problem of “vanishing gradient.” In deep networks, since the chain rule is essentially a multiplication of derivatives (gradients), when there are one or more small values in the multiplication, the gradient of the entire network approaches 0. This results in the network not being able to learn new knowledge efficiently or learning very slowly.

Fig. 2.

Fig. 2

The images of three activation functions and their derivatives.

Tanh(x)

The tanh function and the sigmoid function have similar S-shaped curves, which means that the tanh function may also suffer from the “vanishing gradient” problem when used as an activation function. The mathematical expression of the tanh function is:

graphic file with name M52.gif 13

It can be observed that the tanh function is a linear transformation and scaling of the sigmoid function. Unlike the sigmoid function, which has a range of (0, 1), the tanh function has a range of Inline graphic. The graph of this function is shown in Fig. 2.

Compared to the sigmoid function, the gradients produced by the tanh function are steeper. According to the chain rule, the derivative of the tanh function is four times the derivative of the sigmoid function. Therefore, when a larger gradient is needed to optimize the model or when there are restrictions on the output range, the tanh function can be chosen as the activation function.

Sin(x)

In standard PINN algorithms, using the tanh activation function and the Xavier initialization strategy can lead to asymptotic flattening of the solution outputs. This flattening of the solution outputs often limits the model’s performance as it tends to get trapped in trivial local minima in many practical scenarios. The sine function (Inline graphic) is chosen as the activation function to overcome this problem, for the sine function offers the following advantages:

Overcoming asymptotic flattening output problem: Sigmoid and tanh functions saturate at extreme input values, leading to the vanishing gradient problem and causing the network’s output to flatten. In contrast, using the sine function can maintain relatively smooth outputs throughout the input space and is less prone to saturation. Therefore, using the sine function can effectively overcome the asymptotic flattening output problem and enhance the network’s expressive power.

Symmetry: The sine function is symmetrical at the origin, i.e., Inline graphic. This symmetry enhances the stability and symmetry of the neural network model. It helps the model respond similarly to positive and negative input values, reducing model asymmetry and further improving the model’s expressive power.

Improved optimization and convergence speed: The sine function is a continuously differentiable function with smooth properties. This allows the sine activation function to better propagate gradients during the training process, facilitating information flow and optimization convergence. Compared to some discontinuous or non-differentiable activation functions, the smoothness of the sine function makes the network easier to learn and adjust parameters, accelerating the optimization process and improving the model’s training efficiency and performance.

Increased nonlinear fitting capability: The sine function is a nonlinear function, introducing the sine activation function adds richer nonlinear features and increases the neural network’s representation capacity. This is important for solving problems with complex nonlinear relationships, such as image, audio, or time-series data analysis. The periodic nature of the sine function allows it to better capture periodic patterns or variations in the data, further enhancing the model’s fitting capability.

In conclusion, using the sine function as the activation function offers advantages such as overcoming asymptotic flattening output, periodicity, symmetry, and increased nonlinear fitting capability. These characteristics enable the sine activation function to improve the performance of models in solving real-world problems and avoid getting stuck in trivial local minima.

Error reduction rate

When examining the differences in prediction accuracy between the PINN with the sine function and the standard PINN, a metric noted as ’the Error Reduction Rate (ERR)’54 is adopted. The ERR is designed to quantify the improvement in prediction precision that a method offers over another method. ERR is a dimensionless quantity that indicates the percentage reduction in relative error when using a method compared to another method. Specifically, the formula for calculating the ERR is as follows:

graphic file with name M56.gif 14

In this paper, Inline graphic represents the relative error of the results obtained by the standard PINN, and Inline graphic represents the relative error of the results obtained by the PINN with the sine function. A positive ERR value indicates that the sine function has improved the prediction accuracy, with higher values signifying greater improvements. With this measure, the effectiveness of the ’Sine’ activation function in enhancing PINN’s predictive power can be evaluated more accurately.

Details of the computational platform

In order to ensure the reliability and comparability of experimental results, all experiments were conducted in one experimental environment. We give the specific information of the computing platform to provide a reference for readers to do related research or reproduce the experiments in this paper. All the experiments in this article were completed on a Lenovo computer with an Intel(R) Core(TM) i5-8250U CPU @ 1.60GHz 1.80 GHz processor and 8.00 memory. The environment in which the codes are written and run is based on Python 3.6 and TensorFlow 1.14.

Data-driven solution

In this paper, we consider the following Inline graphic-dimensional NLPDE:

graphic file with name M60.gif 15

where the subscripts denote the partial derivatives, e.g., Inline graphic, Inline graphic, Inline graphic. Inline graphic is some nonlinear function of related variables such that Eq. (15) contains many famous nonlinear dispersion equations. In this paper, we focus on the following five equations:

  1. The third-order dispersion KdV equation55,56
    graphic file with name M65.gif 16
    where a denotes the strength of the nonlinear term or the magnitude of the interaction force. In nonlinear fluctuations, the term describes wave interactions and nonlinear correction effects. b denotes the strength of the dispersion term.
  2. The mKdV equation57,58
    graphic file with name M66.gif 17
    where a is the free parameter. Eq. (17) can be used as an isolated sub-model for describing the multiplicity of plasma and phonon interactions in a non-tonal lattice, and is generally referred to by researchers as the modified KdV equation.
  3. The general K(2, 2) equation59
    graphic file with name M67.gif 18
    where a is a parameter indicating the strength or force of the nonlinear term. Eq. (18) is also known as the two-dimensional KdV equation of the second kind.
  4. The fifth-order Caudrey-Dodd-Gibbon (CDG) equation60
    graphic file with name M68.gif 19
    where abc are arbitrary non-zero constants and Inline graphic denote dispersion terms. Eq. (19) is a generalized fifth order dispersion KdV studied primordially by Caudrey et al61,62.
  5. The seventh-order KdV equation(Lax’s equation)63
    graphic file with name M70.gif 20
    It was introduced by Pomeau et al., in order to discuss the stability of the structure of KdV equations in cases of abnormal disorder64.

In this section, we use the PINN algorithm to learn the solutions of the higher-order KdV equations described above. A 5-layer feed-forward neural network (4 hidden layers of 20 neurons each) is constructed, using tanh and sine functions as the activation functions for comparison. The L-BFGS-B algorithm is adopted to optimize the loss function and use the relative Inline graphic norm to measure the norm error in the solution predictions. In addition, each equation has the same number of randomly selected points Inline graphic at the initial boundaries and the same number of randomly selected configuration points Inline graphic in the space-time domain.

The KdV equation

The general expression for the KdV equation is shown in Eq. (16). In this paper we consider only the parameters are Inline graphic, Inline graphic55,56, then the equation is

graphic file with name M76.gif 21

The single soliton and double soliton solutions are discussed separately in this subsection.

Case 1

In55, the authors used the sine-cosine method to obtain a single soliton solution of Eq. (21), and the expression was given as follows:

graphic file with name M77.gif 22

The initial value and boundary value conditions are considered as

graphic file with name M78.gif 23

The residual term of Eq. (21) is:

graphic file with name M79.gif 24

To obtain the training dataset, we divide the spatial region Inline graphic and the temporal region Inline graphic into Inline graphic and Inline graphic discrete equidistant points, respectively. Thus, the training dataset is the discrete 262,144 points in the given spatio-temporal domain [− 2,2]Inline graphic[− 2,2]. The training data used for the loss function(MSE) consists of Inline graphic points randomly selected from the initial boundary data and Inline graphic points randomly selected from the spatio-temporal domain(The following positive problem generates training data in the same way as here and will not be redefined). The experiment results are illustrated in Figs. 3, 4, 5 and 6.

Fig. 3.

Fig. 3

The dynamical evolution of the predictive single soliton solution of the Eq. (21).

Fig. 4.

Fig. 4

The exact single soliton solution of the Eq. (21).

Fig. 5.

Fig. 5

Error dynamics.

Fig. 6.

Fig. 6

Loss curve.

The top panel of Fig. 3 shows the heat map predicted by the PINN algorithm with sine function. The left of the bottom of Fig. 3 shows the more specific comparison between the exact solution and the prediction solution at four different times t = − 0.5, − 0.25, 0.25, and 0.5, and the spatiotemporal dynamics behavior of the prediction single soliton solution of the Eq. (21) is shown on the right. The spatiotemporal dynamics of the exact single soliton solution of the Eq. (21) is shown in Fig. 4 for comparison.

Figure 5 shows the error dynamics between the exact soliton solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 6 illustrates the loss curve in the process of solving the single soliton solution.

As shown in Table 2, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 73 seconds.

Table 2.

KdV equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 18.18%
Time cost 50 s 73 s

Case 2

In this subsection, we focus on the double soliton solution of Eq. (21). The expression is as follows56

graphic file with name M92.gif 25

The initial value and boundary value conditions are considered as

graphic file with name M93.gif 26

The residual term is same as Case 1, we omit it.

The experiment results are illustrated in Figs. 7, 8, 9 and 10.

Fig. 7.

Fig. 7

The dynamical evolution of the predictive double soliton solution of the Eq. (21).

Fig. 8.

Fig. 8

The exact double soliton solution of the Eq. (21).

Fig. 9.

Fig. 9

Error dynamics.

Fig. 10.

Fig. 10

Loss curve.

The top panel of Fig. 7 shows the heat map predicted by the PINN algorithm with the sine function. The left of the bottom of Fig. 7 shows the more specific comparison between the exact solution and the prediction solution at four different times Inline graphic, and the spatiotemporal dynamics behavior of the prediction double soliton solution of the Eq. (21) is shown on the right. The spatiotemporal dynamics of the exact double soliton solution of the Eq. (21) is shown in Fig. 8 for comparison.

Figure 9 shows the error dynamics between the exact soliton solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 10 illustrates the loss curve in the process of solving the prediction double soliton solution.

As shown in Table 3, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 679 seconds.

Table 3.

KdV equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 76.11%
Time cost 379 s 679 s

mKdV equation

We investigate the mKdV equation Eq. (17) in this subsection.

Case 1

When Inline graphic, Eq. (17) is

graphic file with name M101.gif 27

An solitary wave solution of the Eq. (27) is obtained in57:

graphic file with name M102.gif 28

and the initial value and boundary value conditions are

graphic file with name M103.gif 29

The residual term can be obtained by Eq. (27), we omit it. The experiment results are illustrated in Figs. 11, 12, 13 and 14:

Fig. 11.

Fig. 11

The dynamical evolution of the predictive solitary wave solution of the Eq. (27).

Fig. 12.

Fig. 12

The exact solitary wave solution of the Eq. (27).

Fig. 13.

Fig. 13

Error dynamics.

Fig. 14.

Fig. 14

Loss Curve.

The top panel of Fig. 11 shows the heat map predicted by the PINN algorithm with the sine function. The left of the bottom of Fig. 11 shows the more specific comparison between the exact solution and the prediction solution at four different times Inline graphic, and the spatiotemporal dynamics behavior of the prediction solitary wave solution of the Eq. (27) is shown on the right. The spatiotemporal dynamics of the exact solitary wave solution of the Eq. (27) is shown in Fig. 12 for comparison.

Figure 13 shows the error dynamics between the exact soliton solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 14 illustrates the loss curve in the process of solving the solitary wave solution.

As shown in Table 4, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 11 seconds.

Table 4.

MKdV equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 50.57%
Time cost 10 s 11 s

Case 2

When Inline graphic, Eq. (17) is

graphic file with name M111.gif 30

A kink solution of the Eq. (30) is obtained in58:

graphic file with name M112.gif 31

and the initial value and boundary value conditions is

graphic file with name M113.gif 32

The residual term can be obtained by Eq. (30). The experiment results are illustrated in Figs. 15, 16, 17 and 18:

Fig. 15.

Fig. 15

The dynamical evolution of the predictive kink solution of the Eq. (30).

Fig. 16.

Fig. 16

The exact kink solution of the Eq. (30).

Fig. 17.

Fig. 17

Error dynamics.

Fig. 18.

Fig. 18

Loss Curve.

The top panel of Fig. 15 shows the heat map predicted by the PINN algorithm with the sine function. The left of the bottom of Fig. 15 shows the more specific comparison between the exact solution and the prediction solution at four different times Inline graphic, and the spatiotemporal dynamics behavior of the prediction kink solution of the Eq. (30) is shown on the right. The spatiotemporal dynamics of the exact kink solution of the Eq. (30) is shown in Fig. 16 for comparison.

Figure 17 shows the error dynamics between the exact kink solution and the prediction solution learned by the PINN algorithm with the sine function. Figure 18 illustrates the loss curve in the process of solving the kink solution.

As shown in Table 5, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 7 seconds.

Table 5.

MKdV equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 50.37%
Tme cost 12 s 7 s

The K(2, 2) equation

We consider the following specific form of the K(2, 2) equation when Inline graphic:

graphic file with name M121.gif 33

In this section, we focus on the period travelling wave solution of Eq. (33). Through the travelling wave transformation method, the period travelling wave solution of the Eq. (33) is obtained in59:

graphic file with name M122.gif 34

where c is wave velocity, g and c1 are integral constants. In59, the parameters are selected as Inline graphic, then the initial value and boundary value conditions for the Eq. (33) are

graphic file with name M124.gif 35

The residual term can be obtained by Eq. (33), we omit it. The experiment results are illustrated in Figs. 19, 20, 21 and 22:

Fig. 19.

Fig. 19

The dynamical evolution of the predictive periodic travelling solution of the Eq. (33).

Fig. 20.

Fig. 20

The exact periodic travelling wave solution of the Eq. (33).

Fig. 21.

Fig. 21

Error dynamics.

Fig. 22.

Fig. 22

Loss curve.

The top panel of Fig. 19 shows the heat map predicted by the PINN algorithm with the sine function. The left bottom of Fig. 19 shows the more specific comparison between the exact solution and the prediction solution at four different spaces Inline graphic, and the spatiotemporal dynamics behavior of the prediction periodic traveling wave solution of the Eq. (33) is shown on the right. The spatiotemporal dynamics of the exact periodic traveling wave solution of the Eq. (33) is shown in Fig. 20 for comparison.

Figure 21 shows the error dynamics between the exact solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 22 illustrates the loss curve in the process of solving the periodic traveling wave solution.

As shown in Table 6, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 93 seconds.

Table 6.

K(2, 2) equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 82.85%
Time cost 106 s 93 s

The CDG equation

When Inline graphic, Inline graphic, Inline graphic, the CDG equation60 is

graphic file with name M134.gif 36

Zhou et al obtained the solitary wave solution of the Eq. (36) by using the Sine-Cosine expansion method60 as follows:

graphic file with name M135.gif 37

and the initial value and boundary value conditions is

graphic file with name M136.gif 38

The residual term can be obtained by Eq. (36), we omit it. The experiment results are illustrated in Figs. 23, 24, 25 and 26:

Fig. 23.

Fig. 23

The dynamical evolution of the predictive solitary wave solution of the Eq. (36).

Fig. 24.

Fig. 24

The exact solitary wave solution of the Eq. (36).

Fig. 25.

Fig. 25

Error dynamics.

Fig. 26.

Fig. 26

Loss curve.

The top panel of Fig. 23 shows the heat map predicted by the PINN algorithm with the sine function. The left of the bottom of Fig. 23 shows the more specific comparison between the exact solution and the prediction solution at four different times Inline graphic, and the spatiotemporal dynamics behavior of the prediction solitary wave solution of the Eq. (36) is shown on the right. The spatiotemporal dynamics of the exact solitary wave solution of the Eq. (36) is shown in Fig. 24 for comparison.

Figure 25 shows the error dynamics between the exact solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 26 illustrates the loss curve in the process of solving the solitary wave solution.

As shown in Table 7, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 312 seconds.

Table 7.

CDG equation: Inline graphic relative error and time cost of PINN(tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 72.03%
Time cost 185 s 312 s

The Lax’s equation

We discuss the following solitary wave solution63 to Eq. (20)

graphic file with name M143.gif 39

The initial value and boundary value conditions for the Eq. (20) is

graphic file with name M144.gif 40

The residual term can be obtained by Eq. (20), we omit it. The experiment results are illustrated in Figure 27, 28, 29 and 30:

Fig. 27.

Fig. 27

The dynamical evolution of the predictive solitary wave solution of the Lax’s equation.

Fig. 28.

Fig. 28

The exact solitary wave solution of the Lax’s equation.

Fig. 29.

Fig. 29

Error dynamics.

Fig. 30.

Fig. 30

Loss curve.

The top panel of Fig. 27 shows the heat map predicted by the PINN algorithm with the sine function. The left of the bottom of Fig. 27 shows the more specific comparison between the exact solution and the prediction solution at four different times Inline graphic, and the spatiotemporal dynamics behavior of the prediction solitary wave solution of Eq. (20) is shown on the right. The spatiotemporal dynamics of the exact solitary wave solution (39) is shown in Fig. 28 for comparison.

Figure 29 shows the error dynamics between the exact solution and the predictive solution learned by the PINN algorithm with the sine function. Figure 30 illustrates the loss curve in the process of solving the solitary wave solution.

As shown in Table 8, the PINN method with the sine activation function can improve the accuracy of the prediction solution. The relative Inline graphic error of size Inline graphic is reached when the model runs for about 3961 seconds.

Table 8.

Lax equation: Inline graphic relative error and time cost of PINN (tanh) and PINN(sine).

Solution Method
PINN(tanh) PINN(sine) ERR
u(xt) Inline graphic Inline graphic 66.56%
Tme cost 2521 s 3961 s

Under a fixed framework (4 hidden layers and 20 neurons each), 7 different solutions of 5 equations are learned successfully by the PINN method. Tanh and sine as activation functions are used in the PINN method, respectively, aiming to explore the influence of different activation functions on the accuracy and efficiency of learning results. Experiments show that in the problem of learning the solution of the KdV equation family, the sine function can improve the accuracy, but the degree of improvement is not very obvious; especially in the case of simple solution forms (kink solution, single soliton solution), the original PINN (tanh) can obtain an acceptable learning effect. When the form of the solution is more complex (double soliton solution, periodic solution), the PINN framework using sine as the activation function can improve the accuracy by 1 order of magnitude.

The experiment also shows that the more complex the form of the differential equation, such as higher order differential terms, more nonlinear terms, are obvious factors that affect the learning accuracy and efficiency. We can see that both the five-order and seven-order equations are less accurate than the third-order equations, and the time required increases sharply with the complexity of the equations.

Data-driven parameter estimation of NLPDEs

In this section, the PINN method with tanh and sine as activation functions, respectively, is used to study the inverse problem of nonlinear partial differential equations: that is, to learn the unknown parameters of the above five KdV equations. In order to obtain the estimated parameters and investigate the effects of the activation functions, all examples in this section continue to use the same structural framework: 3 hidden layers with 20 neurons each. The L-BFGS-B algorithm is used to optimize the loss function, the relative Inline graphic norm is used to measure the error between the predicted and actual, and the Latin hypercubic sampling method is used to get the training dataset N.

Data-driven parameter estimation in the KdV equation

We consider the following three-order KdV equation with an unknown real-valued parameter Inline graphic:

graphic file with name M153.gif 41

then the residual term is

graphic file with name M154.gif 42

The PINN algorithm learns u(xt) and the parameter Inline graphic by minimizing the mean square error loss

graphic file with name M156.gif 43

where Inline graphic represents the training data come from the exact double soliton solution of Eq. (41) with Inline graphic and Inline graphic:

graphic file with name M160.gif

The spatial region Inline graphic and the temporal region Inline graphic are divided into Inline graphic and Inline graphic discrete equidistant points, respectively. Thus, the dataset is a discrete set of 262144 points in the given time-space region [-2,2]Inline graphic[-2,2]. We use Latin hypercubic sampling to select Inline graphic points out of the 262144 points in Inline graphic as the training dataset. (The training data used in the following four subsections are generated in the same way as here and are omitted.)

Table 9 shows the performance comparison between the PINN algorithm with the sine function and the standard PINN algorithm to solve the double soliton solution and the parameter Inline graphic of the Eq. (41). It can be easily found from Table 9 that the performance of the PINN algorithm with the sine function is better than the standard PINN algorithm in both the accuracy of learning solutions and the accuracy of identifying unknown parameter Inline graphic.

Table 9.

Performance comparison of the predictive solution and predictive parameter of Eq. (41).

Case PINN(Inline graphic) PINN(Inline graphice)
Loss 3.8809 × 10 −4 6.7632 × 10−5
Inline graphic Error u 9.5180 × 10−3 2.3708 × 10−3
Error Inline graphic 1.1878% 0.1151%
Inline graphic 1.0119 0.9988

Data-driven parameter estimation in the mKdV equation

We consider the following three-order mKdV equation with an unknown real-valued parameter Inline graphic:

graphic file with name M176.gif 44

then the residual term is

graphic file with name M177.gif 45

The PINN algorithm learns u(xt) and the parameter Inline graphic by minimizing the mean square error loss

graphic file with name M179.gif 46

where Inline graphic represents the training data on the exact kink solution of Eq. (44) with Inline graphic, and Inline graphic:

graphic file with name M183.gif

Table 10 shows the performance comparison between the PINN algorithm with the sine function and the standard PINN algorithm to learn the parameter Inline graphic of the Eq. (44). It is obvious that the performance of the PINN algorithm with the sine activation function is better than the standard PINN algorithm in the accuracy of the learning solution. However, the accuracies of the values of unknown parameter Inline graphic obtained by the two methods is very high, and there are no obvious difference.

Table 10.

Performance comparison of the predictive solution and predictive parameter of Eq. (44).

Case PINN(Inline graphic) PINN(Inline graphice)
Loss 9.7835 × 10−7 2.7736 × 10−7
Inline graphic Error u 2.8395 × 10−4 9.6103 × 10−5
Error Inline graphic 0.0338% 0.0371%
Inline graphic 1.00034 1.00037

Data-driven parameter estimation in the K(2, 2) equation

We consider the following K(2, 2) equation with two unknown real-valued parameter Inline graphic and Inline graphic:

graphic file with name M193.gif 47

then the residual term is

graphic file with name M194.gif 48

The PINN algorithm learns u(xt) and the parameter Inline graphic and Inline graphic by minimizing the mean square error loss

graphic file with name M197.gif 49

where Inline graphic represents the training data which comes from the exact double soliton solution of Eq. (47) with Inline graphic and Inline graphic, in Inline graphic:

graphic file with name M202.gif

Table 11 shows the performance comparison between the PINN algorithm with the sine function and the standard PINN algorithm to solve the unknown parameters Inline graphic and Inline graphic of the K(2,2) equation. It can be easily found from Table 11 that the performance of the PINN algorithm with the sine function is better in both the accuracy of learning solutions and the accuracy of identifying unknown parameters Inline graphic and Inline graphic.

Table 11.

Performance comparison of the predictive solution and predictive parameter of Eq. (47).

Case PINN(Inline graphic) PINN(Inline graphice)
Loss 4.3772 × 10−4 1.8681 × 10−5
Inline graphic Error u 1.6164 × 10−3 4.1269 × 10−4
Error Inline graphic 0.0766% 0.0117%
Error Inline graphic 0.1217% 0.0106%
Inline graphic 0.9992 0.9998
Inline graphic 0.9987 0.9998

Data-driven parameter estimation in the CDG equation

In this section, we consider the CDG equation with an unknown parameter Inline graphic:

graphic file with name M215.gif 50

then the residual term is

graphic file with name M216.gif 51

The PINN algorithm learns u(xt) and the parameter Inline graphic by minimizing the mean square error loss

graphic file with name M218.gif 52

where Inline graphic represents the training data on the exact solitary wave solution of Eq. (50) with Inline graphic,in Inline graphic:

graphic file with name M222.gif

Table 12 shows the performance comparison between the PINN algorithm with the sine function and the standard PINN algorithm to learn the unknown parameter Inline graphic of the CDG equation. It can be seen from Table 12, that the value of the loss function is small, and the precisions of the solution and the unknown parameter are acceptable. The learning effects of the two activation functions are not much different.

Table 12.

Performance comparison of the predictive solution and predictive parameter of Eq. (50).

Case PINN(Inline graphic) PINN(Inline graphice)
Loss 7.3111 × 10−7 9.7328 × 10−7
Inline graphic Error u 5.7212 × 10−3 5.07647 × 10−3
Error Inline graphic 8.1902% 3.7163%
Inline graphic 29.9181 30.037

Data-driven parameter estimation in the Lax’s equation

In this section, we consider the Lax’s equation with coefficient Inline graphic:

graphic file with name M230.gif 53

Then the residual term is

graphic file with name M231.gif 54

The PINN algorithm learns u(xt) and the parameter Inline graphic by minimizing the mean square error loss

graphic file with name M233.gif 55

where Inline graphic represents the training data on the exact solitary solution of Eq. (53) with Inline graphic and Inline graphic:

graphic file with name M237.gif

Table 13 shows the performance comparison between the PINN algorithm with the sine function and the standard PINN algorithm to learn the unknown parameter Inline graphic of the CDG equation. It can be seen from Table 13, that the value of the loss function is small, and the precisions of the solution and the unknown parameter are acceptable. The learning effect of PINN with the tanh activation function is better than the other.

Table 13.

Performance comparison of the predictive solution and predictive parameter of Eq. (53).

Case PINN(Inline graphic) PINN(Inline graphice)
Loss 2.3444 × 10−7 1.1147 × 10−6
Inline graphic Error u 6.8690 × 10−4 1.8833 × 10−3
Error Inline graphic 0.0025% 0.2154%
Inline graphic 0.99997 1.00215

In this section, we learn the solutions and unknown parameters of 5 equations using two different activation functions on the same PINN framework. Experiments show that PINN can learn the values of unknown parameters very well. In addition to the 7-order Lax’s equation, the PINN method using the sine function as the activation function can obtain relatively higher precision.

Discussion and conclusion

Through the tireless research of mathematical researchers, a number of effective methods have been obtained in both exact analytical solutions and numerical solutions. However, for various NLPDEs, there are still a lot of problems to be solved. The excellent performance of artificial neural networks in many research fields provides a new way for studying nonlinear partial differential equations. The PINN framework imparts the model structure of PDE into the learning iterative process of the artificial neural network, which guarantees the reliability of the method theoretically and has been validated effectively in practice. However, can the robustness of PINN be maintained for NLPDEs with complex structures, such as high-order partial derivatives and nonlinear terms of various forms? Even for equations with simple structures, it still has special solutions with complex forms such as multi-soliton solutions, periodic solutions, etc.. Can PINN learn solutions with high precision based on the data-driven principle? These are many questions worth investigating.

In this paper, a series of high-order NLPDEs are studied to get predicted solutions and parameters and explore the effects of different activation functions on learning accuracy and efficiency. At the same time, we hope to analyze the influence of the structure of the model on the reliability of the PINN method through the experimental results. The PINN algorithm with the sine activation function and the standard PINN algorithm(with the tanh activation function) are applied to solving the forward problem and inverse problem of a family of higher order dispersive KdV equations. The results show that both algorithms can learn data-driven single soliton solutions, double soliton solutions, period traveling wave solutions, and kink solutions. Experiments show that PINN with sine activation function can obtain better accuracy than the standard PINN in the problem. For the parameter estimation of these higher-order NLPDEs, the same is true in most cases, and the PNN algorithm using the sine activation function can obtain higher accuracy.

The experiment also shows that when the form of the solution is relatively simple, the PINN algorithm can learn the solution with high precision in a short time, while when the form of the solution is more complex, the PINN algorithm needs more time to get the solution with acceptable precision. For the solution of the same form, the more complex the structure of the equation, the more learning time is required, which is consistent with the structure of the loss function. These results show that it is necessary to improve the PINN algorithm for complex equations or complex solutions.

In order to make a comparison, this paper uses one framework to learn the solutions of different equations for the forward problem and another framework to learn the solutions and unknown parameters for the inverse problem. In practice, the number of PINN layers and the number of neurons in each layer can be adjusted to fit different equations. A typical example is the 7th-order Lax’s equation, in the learning of the positive problem, using the PINN method with 5 hidden layers, the accuracy of the learned solution obtained by the two activation functions can only reach Inline graphic. In the inverse problem, four hidden layers are used, and the learning accuracy of the standard PINN can be improved by an order of magnitude, and the performance of the standard PINN exceeds that of the PINN using the sine activation function.

It should be emphasized that the research in this paper is not limited to various KdV equations. We also performed experiments on other types of higher order partial differential equations (not included in this paper). A large number of experimental results show that for problems with rich physical backgrounds, it is effective to use the PINN framework and known information to train models to solve or estimate model parameters(modeling). When the complexity of the problem increases (high derivative or complex nonlinearity, etc.), it is necessary to improve the PINN, such as the selection of activation function, the correction of the loss function and the improvement of optimization measures. The results of these experiments also show that the PINN algorithm still has the potential to explore depth. Most of the NLPDEs can be used to describe the complex phenomena in the real world, which contains rich physical information and mechanisms. Therefore, the research in this paper provides a research idea for exploring deep learning in the aspects of physical problems (optics, electromagnetism, microwave, quantum, etc.) or problems based on the intrinsic mechanism of physics (econophysics, biophysics). Future research will continue to focus on the PINN algorithm, exploring algorithm improvements and applications in more areas.

Acknowledgements

The authors thanks the referees for their valuable suggestions. This work is supported by the National Natural Science Foundation of China (No.11561034) and Yunnan Fundamental Research Projects (No.202401AT070395).

Author contributions

Writing-original draft, C.J.; Methodology, C.J.; Software, C.J.; Writing-Review and Editing, S.J; Resources,S.J.; Supervision, H.A.; Supervision, F.H. All authors reviewed the manuscript.

Data availability

The datasets used and/or analyzed during the current study available from the corresponding author on reasonable request.

Declarations

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Ablowitz, M. J. & Clarkson, P. A. Solitons, Nonlinear Evolution Equations and Inverse Scattering (Cambridge University Press, 1991). [Google Scholar]
  • 2.Hu, Y. Geometry of Bäcklund transformations I: Generality. Trans. Am. Math. Soc.373(2), 1181–1210 (2020). [Google Scholar]
  • 3.Gu, C., Hu, H. & Zhou, Z. Darboux transformations in integrable systems: Theory and their applications to geometry (Springer, 2004). [Google Scholar]
  • 4.Hirota, R. The Direct Method in Soliton Theory (Cambridge University Press, 2004). [Google Scholar]
  • 5.Shao, C. et al. Periodic, n-soliton and variable separation solutions for an extended (3+1)-dimensional KP-Boussinesq equation. Sci. Rep.13(1), 15826 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Junaid-U-Rehman, M., Kudra, G. & Awrejcewicz, J. Conservation laws, solitary wave solutions, and lie analysis for the nonlinear chains of atoms. Sci. Rep.13(1), 11537 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Zhang, H. et al. N-lump and interaction solutions of localized waves to the (2+ 1)-dimensional generalized KP equation. Results Phys.25, 104168 (2021). [Google Scholar]
  • 8.Chen, H. et al. Behavior of analytical schemes with non-paraxial pulse propagation to the cubic-quintic nonlinear Helmholtz equation. Math. Comput. Simul.220, 341–356 (2024). [Google Scholar]
  • 9.Zhang, M. et al. Characteristics of the new multiple rogue wave solutions to the fractional generalized CBS-BK equation. J. Adv. Res.38, 131–142 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Gu, Y. et al. Variety interaction between k-lump and k-kink solutions for the (3+ 1)-D Burger system by bilinear analysis. Results Phys.43, 106032 (2022). [Google Scholar]
  • 11.Zhou, X. et al. N-lump and interaction solutions of localized waves to the (2+ 1)-dimensional generalized KDKK equation. J. Geom. Phys.168, 104312 (2021). [Google Scholar]
  • 12.Liang, C. Finite Difference Methods for Solving Differential Equations (National Taiwan University, 2009). [Google Scholar]
  • 13.Dhatt, G., Lefrancois, E. & Touzot, G. Finite element method (John Wiley and Sons, 2012). [Google Scholar]
  • 14.Shen, J., Tang, T. & Wang, L. L. Spectral Methods: Algorithms, Analysis and Applications (Springer, 2011). [Google Scholar]
  • 15.Rumelhart, D. E., Hinton, G. E. & Williams, R. J. Learning representations by back-propagating errors. Nature323(6088), 533–536 (1986). [Google Scholar]
  • 16.Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inform. Process. Syst.25, (2012).
  • 17.Enab, K. & Ertekin, T. Artificial neural network based design for dual lateral well applications. J. Petrol. Sci. Eng.123, 84–95 (2014). [Google Scholar]
  • 18.LeCun, Y., Bengio, Y. & Hinton, G. Deep learning. Nature521(7553), 436–444 (2015). [DOI] [PubMed] [Google Scholar]
  • 19.Shi, S., Han, D. & Cui, M. A multimodal hybrid parallel network intrusion detection model. Connect. Sci.35(1), 2227780 (2023). [Google Scholar]
  • 20.Fei, R. et al. An improved BPNN method based on probability density for indoor location. IEICE Trans. Inf. Syst.106(5), 773–785 (2023). [Google Scholar]
  • 21.Baydin, A. G. et al. Automatic differentiation in machine learning: A survey. J. Mach. Learn. Res.18(153), 1–43 (2018). [Google Scholar]
  • 22.Hornik, K., Stinchcombe, M. & White, H. Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks. Neural Netw.3(5), 551–560 (1990). [Google Scholar]
  • 23.Li, X. Simultaneous approximations of multivariate functions and their derivatives by neural networks with one hidden layer. Neurocomputing12(4), 327–343 (1996). [Google Scholar]
  • 24.Lagaris, I. E., Likas, A. & Fotiadis, D. I. Artificial neural networks for solving ordinary and partial differential equations. IEEE Trans. Neural Netw.9(5), 987–1000 (1998). [DOI] [PubMed] [Google Scholar]
  • 25.Aarts, L. P. & Van Der Veer, P. Neural network method for solving partial differential equations. Neural Process. Lett.14, 261–271 (2001). [Google Scholar]
  • 26.Ramuhalli, P., Udpa, L. & Udpa, S. S. Finite-element neural networks for solving differential equations. IEEE Trans. Neural Netw.16(6), 1381–1392 (2005). [DOI] [PubMed] [Google Scholar]
  • 27.Sirignano, J. & Spiliopoulos, K. DGM: A deep learning algorithm for solving partial differential equations. J. Comput. Phys.375, 1339–1364 (2018). [Google Scholar]
  • 28.Raissi, M., Perdikaris, P. & Karniadakis, G. E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys.378, 686–707 (2019). [Google Scholar]
  • 29.Pang, G., Lu, L. & Karniadakis, G. E. fPINNs: Fractional physics-informed neural networks. SIAM J. Sci. Comput.41(4), A2603–A2626 (2019). [Google Scholar]
  • 30.Kharazmi, E., Zhang, Z., Karniadakis, G. E. Variational physics-informed neural networks for solving partial differential equations. arXiv:1912.00873 (2019).
  • 31.Zhang, D., Guo, L. & Karniadakis, G. E. Learning in modal space: Solving time-dependent stochastic PDEs using physics-informed neural networks. SIAM J. Sci. Comput.42(2), A639–A665 (2020). [Google Scholar]
  • 32.Li, J. & Chen, Y. A physics-constrained deep residual network for solving the sine-Gordon equation[J]. Commun. Theor. Phys.73(1), 015001 (2020). [Google Scholar]
  • 33.Pu, J., Li, J. & Chen, Y. Solving localized wave solutions of the derivative nonlinear Schrödinger equation using an improved PINN method. Nonlinear Dyn.105, 1723–1739 (2021). [Google Scholar]
  • 34.Pu, J., Peng, W. & Chen, Y. The data-driven localized wave solutions of the derivative nonlinear Schrödinger equation by using improved PINN approach. Wave Motion107, 102823 (2021). [Google Scholar]
  • 35.Wang, L. & Yan, Z. Data-driven rogue waves and parameter discovery in the defocusing nonlinear Schrödinger equation with a potential using the PINN deep learning. Phys. Lett. A404, 127408 (2021). [Google Scholar]
  • 36.Zhou, Z. & Yan, Z. Solving forward and inverse problems of the logarithmic nonlinear Schrödinger equation with PT-symmetric harmonic potential via deep learning. Phys. Lett. A387, 127010 (2021). [Google Scholar]
  • 37.Zhou, Z. & Yan, Z. Deep learning neural networks for the third-order nonlinear Schrödinger equation: Bright solitons, breathers, and rogue waves. Commun. Theor. Phys.73(10), 57–65 (2021). [Google Scholar]
  • 38.Bai, G., Koley, U., & Mishra, S., et al. Physics informed neural networks(PINNs) for approximating nonlinear dispersive PDEs. arXiv:2104.05584, (2021).
  • 39.Xie, G. et al. Gradient-enhanced physics-informed neural networks method for the wave equation. Eng. Anal. Boundary Elements166, 105802 (2024). [Google Scholar]
  • 40.Shahrill, M., Chong, M. S. F. & Nor, H. N. H. M. Applying explicit schemes to the korteweg-de vries equation. Mod. Appl. Sci.9(4), 200 (2015). [Google Scholar]
  • 41.Zhang, G., Li, Z. & Duan, Y. Exact solitary wave solutions of nonlinear wave equations. Sci. China, Ser. A Math.44, 396–401 (2001). [Google Scholar]
  • 42.Wazwaz, A. M. New solitary-wave special solutions with compact support for the nonlinear dispersive K(m, n) equations. Chaos Solitons Fractals13(2), 321–330 (2002). [Google Scholar]
  • 43.Wazwaz, A. M. Analytic study on the generalized fifth-order KdV equation: New solitons and periodic solutions. Commun. Nonlinear Sci. Numer. Simul.12(7), 1172–1180 (2007). [Google Scholar]
  • 44.Salas, A. H. & Gómez, S. C. A. Application of the Cole-Hopf transformation for finding exact solutions to several forms of the seventh-order KdV equation. Math. Probl Eng2010, 1–14 (2010). [Google Scholar]
  • 45.Khan, K. et al. Electron-acoustic solitary potential in nonextensive streaming plasma. Sci. Rep.12(1), 15175 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Khan, K., Ali, A., & Irfan, M. Spatio-temporal fractional shock waves solution for fractional Korteweg-de Vries burgers equations. In Waves in Random and Complex Media 1-17 (2023).
  • 47.Khan, K. et al. Higher order non-planar electrostatic solitary potential in a streaming electron-ion magnetoplasma: Phase plane analysis. Symmetry15(2), 436 (2023). [Google Scholar]
  • 48.Li, J. & Chen, Y. A deep learning method for solving third-order nonlinear evolution equations. Commun. Theor. Phys.72(11), 115003 (2020). [Google Scholar]
  • 49.Glorot, X., & Bengio, Y. Understanding the difficulty of training deep feedforward neural networks[C]//Proceedings of the thirteenth international conference on artificial intelligence and statistics. In JMLR Workshop and Conference Proceedings 249-256 (2010).
  • 50.Kingma, D.P., & Ba, J. Adam: A method for stochastic optimization. arXiv:1412.6980 (2014).
  • 51.Byrd, R. H. et al. A limited memory algorithm for bound constrained optimization. SIAM J. Sci. Comput.16(5), 1190–1208 (1995). [Google Scholar]
  • 52.Stein, M. Large sample properties of simulations using Latin hypercube sampling. Technometrics29(2), 143–151 (1987). [Google Scholar]
  • 53.Sharma, S., Sharma, S. & Athaiya, A. Activation functions in neural networks. Towards Data Sci.6(12), 310–316 (2017). [Google Scholar]
  • 54.Zhang, Z. Y. et al. Enforcing continuous symmetries in physics-informed neural network for solving forward and inverse problems of partial differential equations. J. Comput. Phys.492, 112415 (2023). [Google Scholar]
  • 55.Wazwaz, A. M. A sine-cosine method for handling nonlinear wave equations. Math. Comput. Model.40(5–6), 499–508 (2004). [Google Scholar]
  • 56.Gardner, C. S. et al. Korteweg–Devries equation and generalizations. VI. methods for exact solution. Commun. Pure Appl. Math.27(1), 97–133 (1974). [Google Scholar]
  • 57.Shi-Kuo, L. et al. New periodic solutions to a kind of nonlinear wave equations. Acta Phys. Sin. Chin. Edit.51(1), 14–19 (2002). [Google Scholar]
  • 58.Fu, Z. et al. Jacobi elliptic function expansion method and periodic wave solutions of nonlinear wave equations. Phys. Lett. A289(1–2), 69–74 (2001). [Google Scholar]
  • 59.Dong-Bing, L. et al. On exact solution to nonlinear dispersion KdV equation. J. Southw. China Norm. Univ.45(6), 21–28 (2020). [Google Scholar]
  • 60.Lan-Suo, Z. et al. The solitary waves solution for a class of the fifth-order KdV equation. Math. Appl.32(2), 376–381 (2019). [Google Scholar]
  • 61.Caudrey, P. J., Dodd, R. K. & Gibbon, J. D. A new hierarchy of Korteweg-de Vries equations. Proc. R. Soc. Lond. Math. Phys. Sci.351(1666), 407–422 (1976). [Google Scholar]
  • 62.Dodd, R. K. & Gibbon, J. D. The prolongation structure of a higher order Korteweg-de Vries equation. Proc. R. Soc. Lond. Math. Phys. Sci.358(1694), 287–296 (1978). [Google Scholar]
  • 63.El-Sayed, S. M. & Kaya, D. An application of the ADM to seven-order Sawada–Kotara equations. Appl. Math. Comput.157(1), 93–101 (2004). [Google Scholar]
  • 64.Pomeau, Y., Ramani, A. & Grammaticos, B. Structural stability of the Korteweg-de Vries solitons under a singular perturbation. Phys. D31(1), 127–134 (1988). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets used and/or analyzed during the current study available from the corresponding author on reasonable request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES