2018
DOI: 10.48550/arxiv.1807.03973
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ReLU Deep Neural Networks and Linear Finite Elements

Juncai He,
Lin Li,
Jinchao Xu
et al.

Abstract: In this paper, we investigate the relationship between deep neural networks (DNN) with rectified linear unit (ReLU) function as the activation function and continuous piecewise linear (CPWL) functions, especially CPWL functions from the simplicial linear finite element method (FEM). We first consider the special case of FEM. By exploring the DNN representation of its nodal basis functions, we present a ReLU DNN representation of CPWL in FEM. We theoretically establish that at least 2 hidden layers are needed i… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
93
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 62 publications
(93 citation statements)
references
References 33 publications
0
93
0
Order By: Relevance
“…During training, multilayer perceptrons (MLPs) construct data-driven bases with no reference to underlying geometry [Cyr et al, 2020, He et al, 2018. It has been proven that for MLPs of increasing width and depth, weights and biases exist for which the Sobolev norm of approximation error converge algebraically.…”
Section: Introductionmentioning
confidence: 99%
“…During training, multilayer perceptrons (MLPs) construct data-driven bases with no reference to underlying geometry [Cyr et al, 2020, He et al, 2018. It has been proven that for MLPs of increasing width and depth, weights and biases exist for which the Sobolev norm of approximation error converge algebraically.…”
Section: Introductionmentioning
confidence: 99%
“…The universal approximation theorem ( [17,33]) clarifies that every continuous function on a compact domain can be uniformly approximated by shallow neural networks with continuous, non-polynomial activation functions. The relationship between ReLU-DNN and linear finite element function was studied in [29]. More results have been established in [4,50,60,11,66] for activation functions with a certain regularity, and these approximation errors were given in the sense of L p norm.…”
Section: Neural Networkmentioning
confidence: 99%
“…DNNs produce a large class of nonlinear functions through compositional construction. Due to their powerful universal approximation ability, in recent years, DNNs have been applied for solving partial differential equations (PDEs), and several DNNbased methods ( [21,5,28,29,22,56,16,23,37,45,42,44]) were proposed to overcome the difficulty so-called the "curse of dimensionality" of the traditional PDE solvers such as finite element method (FEM), which requires a discretization of the interested domain, while the number of the mesh points will increase exponentially fast with respect to the problem dimension and make it quickly become computationally intractable. In such a situation, the generation of meshes is very time-consuming as well.…”
Section: Introductionmentioning
confidence: 99%
“…For instance, in finite-deformation simulations using finite elements, the optimal nodal locations and the solution coefficients have both been simultaneously treated as unknowns in the minimization of the potential energy functional [38]. Since the PINN approximation that is composed by the ReLU activation function can exactly represent piecewise affine functions (Delaunay basis functions) [39], one can view the ReLU network solution as a variational r-adaptive finite element solution procedure. Instead of refining elements in h-adaptive finite elements, adaptive solutions can be realized via a basis refinement strategy that has advantages (for example, 'hanging nodes' are a nonissue), which was put forth by Grinspun [40], and a similar basis refinement perspective can be associated with a multilayer neural network [41].…”
Section: Introductionmentioning
confidence: 99%