
arXiv: 2211.03627
handle: 10810/61476 , 20.500.11824/1540
Residual minimization is a widely used technique for solving Partial Differential Equations in variational form. It minimizes the dual norm of the residual, which naturally yields a saddle-point (min-max) problem over the so-called trial and test spaces. In the context of neural networks, we can address this min-max approach by employing one network to seek the trial minimum, while another network seeks the test maximizers. However, the resulting method is numerically unstable as we approach the trial solution. To overcome this, we reformulate the residual minimization as an equivalent minimization of a Ritz functional fed by optimal test functions computed from another Ritz functional minimization. We call the resulting scheme the Deep Double Ritz Method (D$^2$RM), which combines two neural networks for approximating trial functions and optimal test functions along a nested double Ritz minimization strategy. Numerical results on different diffusion and convection problems support the robustness of our method, up to the approximation properties of the networks and the training capacity of the optimizers.
28 pages
variational formulation, FOS: Computer and information sciences, Ritz Method, Computer Science - Machine Learning, optimal test functions, Partial Differential Equations, Numerical Analysis (math.NA), neural networks, Optimal test functions, Ritz method, Machine Learning (cs.LG), residual minimization, Variational formulation, Residual minimization, partial differential equations, FOS: Mathematics, Mathematics - Numerical Analysis, Neural networks
variational formulation, FOS: Computer and information sciences, Ritz Method, Computer Science - Machine Learning, optimal test functions, Partial Differential Equations, Numerical Analysis (math.NA), neural networks, Optimal test functions, Ritz method, Machine Learning (cs.LG), residual minimization, Variational formulation, Residual minimization, partial differential equations, FOS: Mathematics, Mathematics - Numerical Analysis, Neural networks
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 1 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
