Skip to main content
Springer Nature - PMC COVID-19 Collection logoLink to Springer Nature - PMC COVID-19 Collection
. 2022 Mar 29;8(2):87. doi: 10.1007/s40819-022-01288-3

Solving Fredholm Integral Equations Using Deep Learning

Yu Guan 1, Tingting Fang 1, Diankun Zhang 1, Congming Jin 1,
PMCID: PMC8960669  PMID: 35372640

Abstract

The aim of this paper is to provide a deep learning based method that can solve high-dimensional Fredholm integral equations. A deep residual neural network is constructed at a fixed number of collocation points selected randomly in the integration domain. The loss function of the deep residual neural network is defined as a linear least-square problem using the integral equation at the collocation points in the training set. The training iteration is done for the same set of parameters for different training sets. The numerical experiments show that the deep learning method is efficient with a moderate generalization error at all points. And the computational cost does not suffer from “curse of dimensionality” problem.

Keywords: Fredholm integral equation, High-dimensional problem, Residual neural network, Deep learning

Introduction

Integral equations have wide applications in electrical engineering [1], optics [2], mathematical biology [3] and other fields. The most popular integral equations are the Fredhom integral equations and the Volterra integral equations. The Fredholm integral equation can be considered as a reformulation of the elliptic partial differential equation and the Volterra integral equation is a reformulation of the fractional-order differential equation, which has wide applications in modeling the real problems, for instance, the chaotic system [4], the dynamics of COVID-19 [5], the motion of beam on nanowire [6], the capacitor microphone dynamical system [7], etc. Since these integral equations usually can not be solved explicitly, numerical methods are necessary to be considered.

We consider the linear Fredholm integral equation of the second kind

f(x)-Ωk(x,y)f(y)dy=g(x), 1

where x,yΩRm, the function g(x) and the kernel k(x,y) are given, and f(x) is the unknown that we want to find.

So far, many numerical methods have been proposed to solve the Fredholm integral equations, for example, the Nyström method [3, 8, 9], the Galerkin method [10], the wavelet analysis method [11], the neural network [12, 13], the collocation method [14], the maximum entropy method [15], etc. However, most of these traditional methods can only solve low-dimensional Fredholm integral equations and suffer from “curse of dimensionality".

The neural network has been successful in solving partial differential equations in mathematical modelling and the applied science, such as medical smoking model [16], nonlinear high order singular models [17], food chain system [1820], Liénard differential model [21], etc. The neural network was also used to solve the Fredholm integral equations in [12, 13], where the authors only evaluated the approximation at some fixed points without generalization. And the integral was evaluated using numerical integral method whose cost depends on the dimension exponentially.

In recent years, deep learning method has been successfully used in artificial intelligence solving high-dimensional problems, such as image recognition [22, 23], speech recognition [24, 25], natural language processing [26], and also in mathematical problems [2729] and physical problems [30].

E and his collaborators have done a series of works on solving high-dimensional differential equations based on deep learning method. In [28], a deep learning-based algorithm was proposed for solving high-dimensional semilinear parabolic partial differential equations and reverse stochastic differential equations from a relation between BSDE (backward stochastic differential equations) and reinforcement learning. In [29], the deep Ritz method for elliptic differential equations was given by numerically solving variational problems. In [27], a machine learning approximation algorithm was raised to solve high-dimensional fully nonlinear second-order partial differential equations. These works show that deep learning method provides a new idea to solve high-dimensional mathematical problems.

In this paper, a deep residual neural network method is proposed to approximate the solution of the high-dimensional linear Fredholm integral equations of the second kind. Few novel highlights of this deep learning method are briefly provided as follows:

  • A deep residual neural network is constructed to solve numerically the linear Fredholm integral equations of the second kind.

  • The proposed method can solve high-dimensional Fredholm integral equations and does not suffer from “curse of dimensionality” problem, that is the cost depends on the dimension linearly.

  • The reasonable absolute error values validate the reliability of the deep learning method.

  • The proposed method has a small generalization error in the domain.

This paper is organized as follows. In Sect. 2 we construct a deep residual neutral network for solving the Fredholm integral equations. In Sect. 3, some numerical experiments are given to show the efficiency of the numerical method. The conclusion is given in Sect. 4.

Proposed Deep Learning Method for Solving Fredhom Integral Equations

The output F(x,θ) of the neural network is a composite function of the input x, where θ denotes the parameters of the neural network including the weighs and bias. Let x be any point in the domain Ω. Now we want to train a deep neural network whose output F(x,θ) is the solution of the Fredholm integral Eq. (1), that is

F(x,θ)-Ωk(x,y)F(y,θ)dy=g(x),xΩ.

To learn the parameters θ, and so the function F(x,θ), take randomly n points {x1,x2,,xn} with a uniform distribution as the training set. Initializing the parameter vector θ, the prediction values F(xi,θ) for i=1,2,,n, can be obtained by forward propagation neural network.

Define the loss function as

lossx(θ)=1ni=1nF(xi,θ)-Ωk(xi,y)F(y,θ)dy-g(xi)2. 2

The training of the neural network is to minimize the loss function (2) by the backward propagation neural network, which is a least-square problem

minθlossx(θ). 3

In Eq. (2) the integral term Ωk(xi,y)F(y,θ)dy can be evaluated using the Monte Carlo method, leading to

Ωk(xi,y)F(y,θ)dyβnj=1nk(xi,xj)F(xj,θ), 4

where β=Ωdx is the volume of Ω.

The training can be done repeatedly for different training set until we get a stationary loss function.

As the network deepens, minimizing the loss function has great difficulties, such as vanishing gradient problem, gradient explosion, and degradation problem. The residual neural network can avoid the vanishing gradient problem and may greatly improve the solution. It also can reduce the risk of over-adapting the parameters to a specific dataset [22]. A residual block is shown in Fig. 1, where an identity shortcut connection is added to a shallow neural network, whose output is H(x)=φ(x)+x, where φ(x) is the output of the shallow neural network. Then the output of the residual block is taken as the input of the next residual block.

Fig. 1.

Fig. 1

Residual neural network block

Our algorithm of deep residual neural network for solving Fredholm integral equations is shown in algorithm 1.graphic file with name 40819_2022_1288_Figa_HTML.jpg

Numerical Experiments

In this section, several Fredholm integral equations are numerically solved using algorithm 1. In the numerical experiments, n=1000 points in Ω are randomly sampled uniformly as the training set to train the deep residual neural network, and the number of training iterations is M=2000. The neural network consists of one input layer, two blocks of residual neural network shown in Fig. 1, and one output layer. There are 30 neurons in the second layer and the forth layer and 10 neurons in the other layers. The ReLU function is used as the active function in the neural network. Minimization is realized by “AdamOptimizer” [31] built in TensorFlow (version 1.13.1 ) with a learning rate 0.001.

To measure the efficiency of the deep learning method for solving the Fredholm integral equations, we consider several examples whose exact solutions are known. Denote f(x1),f(x2),,f(xn) as the exact solutions at n points x1,x2,,xn in the test set. Define the generalization error between the exact solution f(x) of the integral equation and the approximate solution F(x,θ) obtained by using the deep residual neural network as

error=1ni=1nf(xi)-F(xi,θ)2.

The generalization error is evaluated for each example in the following numerical experiments.

Example 1

Consider the three-dimensional Fredholm integral equation

f(x,y,z)+121212k(x,y,z,s,t,v)f(s,t,v)dsdtdv=g(x,y,z), 5

where Ω=[1,2]3,

k(x,y,z,s,t,v)=e-xs-yt-zv,

and

g(x,y,z)=1-1xyz(e-2x-e-x)(e-2y-e-y)(e-2z-e-z).

The exact solution of Eq. (5) is f(x,y,z)=1.

For Example 1, the convergence of the loss function and the generalization error are shown in Fig. 2. Some typical iteration data of the loss function (loss) and the generalization error (error) are given in Table 1. The loss function converges to 10-5, and the generalization error converges to 10-3.

Fig. 2.

Fig. 2

Convergence of the loss function (left) and the generalization error (right) of Example 1

Table 1.

Partial iterative results of the loss function and the generalization error for Example 1

Number of training loss error Number of training loss error
1 0.8958 0.9446 800 3.498e-4 0.0161
60 0.0428 0.2040 1200 9.472e-5 0.0080
80 0.008 0.0804 1600 3.852e-5 0.0051
409 9.937e-4 0.0270 2000 1.917e-5 0.0035

Example 2

Consider a high-dimensional version of the three-dimensional Fredholm integral Eq. (5), that is

f(x)+Ωk(x,y)f(y)dy=g(x),

where Ω=[1,2]m, x={x1,x2,,xm}, y={y1,y2,,ym},

k(x,y)=e-x1y1--xmym

and

g(x)=1+(-1)m1x1x2xm(e-2x1-e-x1)(e-2x2-e-x2)(e-2xm-e-xm).

The exact solution is f(x)=1.

For Example 2, when the dimension m=100, the convergence of the loss function and the generalization error are shown in Fig. 3. Some typical iteration data of the loss function (loss) and the generalization error (error) are given in Table 2. The loss function converges to 10-4, and the generalization error converges to 10-3.

Fig. 3.

Fig. 3

Convergence of the loss function (left) and the generalization error (right) of Example 2

Table 2.

Partial iterative results of the loss function and the generalization error for Example 2 when the dimension m=100

Number of training loss error Number of training loss error
1 1.264 1.123 800 5.814e-4 0.0195
32 0.0131 0.1055 1100 4.195e-4 0.0162
300 0.0012 0.0277 1600 2.239e-4 0.0119
400 9.805e-4 0.0254 2000 1.417e-4 0.0094

Example 3

Consider the four-dimensional Fredholm integral equation

f(x,y,z,w)-xyzw01010101f(s,t,v,r)dsdtdvdr=1516xyzw 6

where Ω=[0,1]4,

k(x,y,z,w,s,t,v,r)=xyzw,

and

g(x,y,z,w)=1516xyzw.

The exact solution is f(x,y,z,w)=xyzw.

For Example 3, the convergence of the loss function and the generalization error are shown in Fig. 4, and some typical iteration data of the loss function (loss) and the generalization error (error) are given in Table 3. The loss function and the generalized error function synchronously converge very fast to a stable state. The loss function converges to 10-3, and the generalization error converges to 10-2.

Fig. 4.

Fig. 4

Convergence of the loss function (left) and the generalization error (right) of Example 3

Table 3.

Partial iterative results of the loss function and the generalization error for Example 3

Number of training loss error Number of training loss error
1 0.0167 0.1190 1100 0.0074 0.0621
100 0.0070 0.0602 1400 0.0079 0.0603
500 0.0075 0.0632 1700 0.0064 0.0585
800 0.0064 0.0614 2000 0.0066 0.0606

Example 4

Consider a high-dimensional version of the four-dimensional Fredholm integral Eq. (6), that is

f(x)-x1x2xmΩf(y)dy=1-12mx1x2xm

where x[0,1]m, x={x1,x2,,xm}, y={y1,y2,,ym},

k(x,y)=x1x2xm,

and

g(x)=1-12mx1x2xm.

The exact solution of the equation is f(x)=x1x2xm.

For Example 4, when the dimension m=100, the convergence of the loss function and the generalization error are shown in Fig. 5, and some typical iteration data of the loss function (loss) and the generalization error (error) are given in Table 4. The loss function converges to 10-7, and the generalization error converges to 10-4.

Fig. 5.

Fig. 5

Convergence of the loss function (left) and the generalization error (right) of Example 4

Table 4.

Partial iterative results of the loss function and the generalization error for Example 4 when the dimension m=100

Number of training loss error Number of training loss error
1 0.6974 0.8341 1114 9.910e-6 0.0023
500 0.0010 0.0256 1243 3.728e-6 0.0010
800 2.731e-4 0.0129 1337 7.648e-7 4.484e-4
1000 4.119e-5 0.0050 2000 2.826e-7 1.718e-4

Discussion and Conclusions

In this paper, we propose a deep learning method based on the residual neural network to solve numerically the linear Fredholm integral equations of the second kind. The output of the deep residual network is used as the numerical solution. The loss function is defined using the Fredholm integral equation. The loss function is optimized by Adam method built in TensorFlow. Then the numerical results, including high-dimensional problems, confirm the efficiency of the method. The main advantage of this method is that it can solve high-dimensional Fredholm integral equations with a cost less sensitive to the dimensionality of the problem.

The accuracy of the residual neural network is not as good as that of the traditional method, such as the Galerkin method. Some error analysis of the neural network has been discussed in [3234]. But so far rigorous error analysis for neural network can not be given yet. The error of the neural network consists of three parts, that is the error between the space of the output of the neural network and the exact solution of the Fredholm integral equation, the optimization error in Eq. (3), and the approximation error in Eq. (4). The error in our numerical experiments has a good accuracy compared to the error of the Monte Carlo method 1/n in Eq. (4). In the future we will explore more techniques or theory to improve the convergent accuracy. Additionally, we will try to construct a deep residual neural network to solve the Volterra integral equations.

Acknowledgements

The research of Congming Jin was supported by the National Natural Science Foundation of China (No. 11571314).

Funding

The authors have not disclosed any funding.

Availability of data and materials

All data generated or analyzed during this study are included in this manuscript.

Declarations

Conflict of interest

On behalf of all authors, the corresponding author states that there is no conflict of interest.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Michalski KA, Mosig JR. Multilayered media Green’s functions in integral equation formulations. IEEE T. Antenn. Propag. 2002;45(3):508–519. doi: 10.1109/8.558666. [DOI] [Google Scholar]
  • 2.Yang P, Liou KN. Geometric-optics-integral-equation method for light scattering by nonspherical ice crystals. Appl. Opt. 1996;35:6568–6584. doi: 10.1364/AO.35.006568. [DOI] [PubMed] [Google Scholar]
  • 3.De Bonis MC, Stanić MP, Tomović Mladenović TV. Nystörm methods for approximating the solutions of an integral equation arising from a problem in mathematical biology. Appl. Numer. Math. 2022;171:193–211. doi: 10.1016/j.apnum.2021.09.004. [DOI] [Google Scholar]
  • 4.Baleanu D, Zibaei S, Namjoo M, Jajarmi A. A nonstandard finite difference scheme for the modelling and nonidentical synchronization of a novel fractional chaotic system. Adv. Differ. Equ. 2021;2021:308. doi: 10.1186/s13662-021-03454-1. [DOI] [Google Scholar]
  • 5.Baleanu D, Hassan Abadi M, Jajarmi A, Zarghami Vahid K, Nieto JJ. A new comparative study on the general fractional model of COVID-19 with isolation and quarantine effects. Alex. Eng. J. 2022;61(6):4779–4791. doi: 10.1016/j.aej.2021.10.030. [DOI] [Google Scholar]
  • 6.Erturk VS, Godwe E, Baleanu D, Kumar P, Asad J, Jajarmi A. Novel fractional-order Lagrangian to describe motion of beam on nanowire. Acta Phys. Pol. A. 2021;140(3):265–272. doi: 10.12693/APhysPolA.140.265. [DOI] [Google Scholar]
  • 7.Jajarmi A, Baleanu D, Zarghami Vahid K, Mohammadi Pirouz H, Asad JH. A new and general fractional Lagrangian approach: a capacitor microphone case study. Results Phys. 2021;31:104950. doi: 10.1016/j.rinp.2021.104950. [DOI] [Google Scholar]
  • 8.Atkinson K. Iterative variants of the Nyström method for the numerical solution of integral equations. Numer. Math. 1974;22(1):17–31. doi: 10.1007/BF01436618. [DOI] [Google Scholar]
  • 9.Khorrami N, Shamloo AS, Parsa Moghaddam B. Nyström method for solution of Fredholm integral equations of the second kind under interval data. J. Intell. Fuzzy Syst. 2019;36(3):2807–2816. doi: 10.3233/JIFS-181206. [DOI] [Google Scholar]
  • 10.Han G, Wang R. Richardson extrapolation of iterated discrete Galerkin solution for two-dimensional Fredholm integral equations. J. Comput. Appl. Math. 2002;139(1):49–63. doi: 10.1016/S0377-0427(01)00390-9. [DOI] [Google Scholar]
  • 11.Babolian E, Shahsavaran A. Numerical solution of nonlinear Fredholm integral equations of the second kind using Haar wavelets. J. Comput. Appl. Math. 2009;225(1):87–95. doi: 10.1016/j.cam.2008.07.003. [DOI] [Google Scholar]
  • 12.Asady B, Hakimzadegan F, Nazarlue R. Utilizing artificial neural network approach for solving two-dimensional integral equations. Math. Sci. 2014;8:117. doi: 10.1007/s40096-014-0117-6. [DOI] [Google Scholar]
  • 13.Effati S, Buzhabadi R. A neural network approach for solving Fredholm integral equations of the second kind. Neural Comput. Appl. 2012;21(5):843–852. doi: 10.1007/s00521-010-0489-y. [DOI] [Google Scholar]
  • 14.Wang K, Wang Q. Taylor collocation method and convergence analysis for the Volterra-Fredholm integral equations. J. Comput. Appl. Math. 2014;260:294–300. doi: 10.1016/j.cam.2013.09.050. [DOI] [Google Scholar]
  • 15.Jin C, Ding J. Solving Fredholm integral equations via a piecewise linear maximum entropy method. J. Comput. Appl. Math. 2016;304:130–137. doi: 10.1016/j.cam.2016.02.044. [DOI] [Google Scholar]
  • 16.Saeed T, Sabir Z, Alhodaly MS, Alsulami HH, Guerrero Sánchez Y. An advanced heuristic approach for a nonlinear mathematical based medical smoking model. Results Phys. 2022;32:105137. doi: 10.1016/j.rinp.2021.105137. [DOI] [Google Scholar]
  • 17.Sabir Z, Wahab HA, Javeed S, Baskonus HM. An efficient stochastic numerical computing framework for the nonlinear higher order singular models. Fractal Fract. 2021;5:176. doi: 10.3390/fractalfract5040176. [DOI] [Google Scholar]
  • 18.Sabir, Z.: Stochastic numerical investigations for nonlinear three-species food chain system. Int. J. Biomath. 2250005, (2021)
  • 19.Sabir Z, Wahab HA. Evolutionary heuristic with Gudermannian neural networks for the nonlinear singular models of third kind. Phys. Scr. 2021;96:125261. doi: 10.1088/1402-4896/ac3c56. [DOI] [Google Scholar]
  • 20.Sabir Z, Ali MR, Sadat R. Gudermannian neural networks using the optimization procedures of genetic algorithm and active set approach for the three-species food chain nonlinear model. J. Ambient Intell. Human. Comput. 2022 doi: 10.1007/s12652-021-03638-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wang B, Wang Y, Gómez-Aguilar JF, Sabir Z, Raja MAZ, Jahanshahi H, Alassafi MO, Alsaadi FE. Gudermannian neural networks to investigate the Liénard differential model. Fractals. 2021 doi: 10.1142/S0218348X22500505. [DOI] [Google Scholar]
  • 22.He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. IEEE Conference on Computer Vision and Pattern Recognition, (2016), Las Vegas, USA
  • 23.Sun, Y., Chen, Y., Wang, X., Tang, X.: Deep learning face representation by joint identification-verification. NIPS 27, (2014)
  • 24.Graves, A., Mohamed, A. R., Hinton, G.: Speech recognition with deep recurrent neural networks. IEEE International Conference on Acoustics, Speech and Signal Processing, May 2013, Vancouver, Canada
  • 25.Hinton G, Deng L, Yu D, Dahl GE, Mohamed A, Jaitly N, Senior A, Vanhoucke V, Nguyen P, Sainath TN, Kingsbury B. Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Proc. Mag. 2012;29(6):82–97. doi: 10.1109/MSP.2012.2205597. [DOI] [Google Scholar]
  • 26.Tom Y, Devamanyu H, Soujanya P, Erik C. Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag. 2018;13(3):55–75. doi: 10.1109/MCI.2018.2840738. [DOI] [Google Scholar]
  • 27.Beck CEW, Jentzen A. Machine learning approximation algorithms for high-dimensional fully nonlinear partial differential equations and second-order backward stochastic differential equations. J. Nonlinear Sci. 2019;29:1563–1619. doi: 10.1007/s00332-018-9525-3. [DOI] [Google Scholar]
  • 28.Han J, Jentzen A. Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations. Commun. Math. Stat. 2017;5:349–380. doi: 10.1007/s40304-017-0117-6. [DOI] [Google Scholar]
  • 29.Yu B. The deep Ritz method: a deep learning-based numerical algorithm for solving variational problems. Commun. Math. Stat. 2018;6:1–12. [Google Scholar]
  • 30.Zhang L, Han J, Wang H, Car REW. Deep potential molecular dynamics: a scalable model with the accuracy of quantum mechanics. Phys. Rev. Lett. 2018;120(14):143001. doi: 10.1103/PhysRevLett.120.143001. [DOI] [PubMed] [Google Scholar]
  • 31.Kingma, D. P., Ba, J. L.: Adam: A Method for stochastic Optimization. The International Conference on Learning Representations (ICLR), May 2015, San Diego, USA
  • 32.Barron AR. Universal approximation bounds for superpositions of a sigmoidal function. IEEE T. Inform. Theory. 1993;39(3):930–945. doi: 10.1109/18.256500. [DOI] [Google Scholar]
  • 33.Barron, A. R., Klusowski, J. M.: Approximation and estimation for high-dimensional deep learning networks. arXiv:1809.03090v2, (2018)
  • 34.Ma, C., Wu, L., Wojtowytsch, S., Wu, L.: Towards a mathematical understanding of neural network-based machine learning: what we know and what we don’t. arXiv:2009.10713v2, (2020)

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All data generated or analyzed during this study are included in this manuscript.


Articles from International Journal of Applied and Computational Mathematics are provided here courtesy of Nature Publishing Group

RESOURCES