Loading…
A non-gradient method for solving elliptic partial differential equations with deep neural networks
Deep learning has achieved wide success in solving Partial Differential Equations (PDEs), with particular strength in handling high dimensional problems and parametric problems. Nevertheless, there is still a lack of a clear picture on the designing of network architecture and the training of networ...
Saved in:
Published in: | Journal of computational physics 2023-01, Vol.472, p.111690, Article 111690 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Deep learning has achieved wide success in solving Partial Differential Equations (PDEs), with particular strength in handling high dimensional problems and parametric problems. Nevertheless, there is still a lack of a clear picture on the designing of network architecture and the training of network parameters. In this work, we developed a non-gradient framework for solving elliptic PDEs based on Neural Tangent Kernel (NTK): 1. ReLU activation function is used to control the compactness of the NTK so that solutions with relatively high frequency components can be well expressed; 2. Numerical discretization is used for differential operators to reduce computational cost; 3. A dissipative evolution dynamics corresponding to the elliptic PDE is used for parameter training instead of the gradient-type descent of a loss function. The dissipative dynamics can guarantee the convergence of the training process while avoiding employment of loss functions with high order derivatives. It is also helpful for both controlling of kernel property and reduction of computational cost. Numerical tests have shown excellent performance of the non-gradient method.
•A non-gradient method is developed to solve elliptic PDEs using DNNs without introducing loss functions.•The residual of the PDEs are used for training and the training kernel is the neural tangent kernel.•ReLU activation function is used to control the locality of the training kernel.•Numerical discretization is used to reduce computational cost. |
---|---|
ISSN: | 0021-9991 1090-2716 |
DOI: | 10.1016/j.jcp.2022.111690 |