LEGENDRE DEEP NEURAL NETWORK (LDNN) AND ITS APPLICATION FOR APPROXIMATION OF NON-LINEAR VOLTERRA-FREDHOLM-HAMMERSTEIN IN-TEGRAL EQUATIONS

Abstract

Various phenomena in biology, physics, and engineering are modeled by differential equations. These differential equations including partial differential equations and ordinary differential equations can be converted and represented as integral equations. In particular, Volterra-Fredholm-Hammerstein integral equations are the main type of these integral equations and researchers are interested in investigating and solving these equations. In this paper, we propose Legendre Deep Neural Network (LDNN) for solving nonlinear Volterra-Fredholm-Hammerstein integral equations (V-F-H-IEs). LDNN utilizes Legendre orthogonal polynomials as activation functions of the Deep structure. We present how LDNN can be used to solve nonlinear V-F-H-IEs. We show using the Gaussian quadrature collocation method in combination with LDNN results in a novel numerical solution for nonlinear V-F-H-IEs. Several examples are given to verify the performance and accuracy of LDNN.

1. INTRODUCTION

Deep neural networks are a main and beneficial part of machine learning family which are applied in various areas including speech processing, computer vision, natural language processing and image processing (LeCun et al., 2015; Krizhevsky et al., 2012) . Also, the approximation of the functions is a significant branch in scientific computational and achieving success in this area is considered by some research (Tang et al., 2019; Hanin, 2019) . Solving differential equations is the other main branch of scientific computational which neural networks and deep learning have been shown success in this area. (Lample & Charton, 2019; Berg & Nyström, 2018; Raissi et al., 2019) . Various phenomena in biology, physics, finance, neuroscience and engineering are modeled by differential equations (Courant & Hilbert, 2008; Davis, 1961) . In recent years, several researchers studied the solving differential equations via deep learning or neural networks. differential equations consists of ordinary differential equations, partial differential equations and integral equations. (Sirignano & Spiliopoulos, 2018; Lu et al., 2019; Meng et al., 2020) . It is notable that the various numerical methods are applied for solving differential equations. Homotopy analysis method (HAM) (Liao, 2012) and variational iteration method (VIM) (He & Wu, 2007) are known as analytical/semi-analytical methods. Usually, spectral methods (Canuto et al., 2012) , Runge-Kutta methods (Hairer et al., 2006) , the finite difference methods (FDM) (Smith, 1985) and the finite element methods (FEM) (Johnson, 2012) are considered as the popular numerical methods. When the complexity of the model does not allow us to obtain the solution explicitly, numerical methods are a proper selection for finding the approximate solution for the models. Recently, some of the machine learning methods are applied for solving differential equations. Chakraverty & Mall (2017) introduced orthogonal neural networks which used orthogonal polynomials in the structure of the network. Raja et al. ( 2019) applied meta-heuristic optimization algorithm to neural network for obtaining the solution of differential equations. Moreover, other methods of machine learning such as support vector machine (Vapnik, 2013) are used to approximate the solution of the models. Least squares support vector machines are considered in these researches (Hajimohammadi et al., 2020; Mehrkanoon & Suykens, 2015) . In this paper, we concern nonlinear Volterra-Fredholm-Hammerstein integral equations (V-F-H-IEs) and try to obtain the solution of them via deep neural network. We present a new numerical approach of machine learning which is a combination of deep neural network and Legendre collocation method. This approach is useful for solving the differential equations and we applied it for solving nonlinear V-F-H-IEs. We used Legendre collocation method to our network for perfect the numerical computations and enhancement the performance the network.

2. LEGENDRE DEEP NEURAL NETWORK (LDNN)

The main purpose of introducing LDNN is to apply it for solving differential models. Indeed, this purpose is to expand the utilization of deep learning networks in the field of scientific computing, especially the solution of differential equations. Moreover, this network has the advantages of solving equations by deep learning as well as numerical methods such as collocation method used to achieve better solution to the equations. LDNN presents a combination of a deep neural network and Legendre collocation method. In fact, our network consists of two networks which have connected consecutive to each other. The first network is a feed forward neural network which has an orthogonal Legendre layer. The second network includes operation nodes to create the desired computational model. In recent decades, numerical methods especially collocation method are popular methods for solving differential equations. In the collocation method, first an approximation of the solution is expanded by using the sum of the basic functions. The basic functions consists of the orthogonal polynomials such as Legendre polynomials.Then this approximation is placed in the differential equation. By considering the appropriate set of candidate points, an attempt is made to obtain the unknown coefficients of the basic functions so that the solution satisfies the equation in a set of candidate points. The first network is applied to creat the approximation of the solution. This approximation can be known as the scattered data interpolation problem. The second network is used to obtain the desired equation so that the solution satisfies it. The structure of LDNN is described in detail at the following rest. Consider that the first network has a M-layer which defined as follows: H 0 = x, x ∈ R d , H 1 = L(W (1) H 0 + b (1) ), H i = f (W (i) H i-1 + b (i) ), 2 ≤ i ≤ M -1, H M = W (M) H M-1 + b (M) . where H 0 is the input layer with d dimension. H i , 1 ≤ i ≤ M -1 are hidden layers, L = [L 0 , L 1 , ...L n ] T which L i are i-th degrees of Legendre orthogonal polynomials, H 1 is an orthogonal layer, f is the hyperbolic tangent activation function or other commonly used activation functions. W (i) , i = 1, • • • , M are the weight parameters and b (i) , 1 ≤ i ≤ M are the bias parameters. H M is the output layer. It is notable that the second network is applied to obtain the desired differential model. This aim is possible by using operation nodes including integrals, derivatives, and etc. These nodes are applied to the output of the first network. Moreover, automatic differentiation (AD) (Baydin et al., 2017) and Legendre Gaussian integration (Shen et al., 2011) have been used in network computing to obtain more accurate and fast calculations. How to train the network and set the parameters are also important points. Supervised learning method is used to train network. The cost function for setting parameters is defined as follows: CostFun = min(y t -y p ) + min(R m ). (1) where y t is an exact value of the model and y p is a predicted value of the LDNN. The definition of R m is explained in section 3.The minimization of CostFun is obtained by performing Adam algorithm (Kingma & Ba, 2015) and the L-BFGS method (Liu & Nocedal, 1989) on mean squared errors of training data set.



Baker et al. (2019) selected deep neural networks for solving the differential equations. Pang et al. (2019) introduced a new network to find the solution of the different equations. Han et al. (2018) solved high-dimensional problems via deep networks. Also, Long et al. (2018) and Raissi et al. (2019) introduced a group of the equations which solved by deep learning. Furthermore, He et al. (2018) and Molina et al. (2019) investigated the effect of the activation function on networks.

