
arXiv: 2406.12672
Classical model reduction techniques project the governing equations onto a linear subspace of the original state space. More recent data-driven techniques use neural networks to enable nonlinear projections. Whilst those often enable stronger compression, they may have redundant parameters and lead to suboptimal latent dimensionality. To overcome these, we propose a multistep algorithm that induces sparsity in the encoder-decoder networks for effective reduction in the number of parameters and additional compression of the latent space. This algorithm starts with sparsely initialized a network and training it using linearized Bregman iterations. These iterations have been very successful in computer vision and compressed sensing tasks, but have not yet been used for reduced-order modelling. After the training, we further compress the latent space dimensionality by using a form of proper orthogonal decomposition. Last, we use a bias propagation technique to change the induced sparsity into an effective reduction of parameters. We apply this algorithm to three representative PDE models: 1D diffusion, 1D advection, and 2D reaction-diffusion. Compared to conventional training methods like Adam, the proposed method achieves similar accuracy with 30% less parameters and a significantly smaller latent space.
27 pages, 20 figures, 4 tables
FOS: Computer and information sciences, Numerical optimization and variational techniques, Nonlinear dimensionality reduction, math.NA, Computer Science - Artificial Intelligence, G.1.6, Multidimensional problems, Machine Learning (stat.ML), Statistics - Machine Learning, 41A63 (Secondary), Linearized Bregman iterations, FOS: Mathematics, nonlinear dimensionality reduction, Mathematics - Numerical Analysis, 65D99, cs.NA, Artificial neural networks and deep learning, Scientific machine learning, scientific machine learning, 65K10 (Primary) 68T07, I.2.6, sparsity, G.1.6; I.2.6, Numerical Analysis (math.NA), cs.AI, stat.ML, Numerical approximation and computational geometry (primarily algorithms), neural architecture search, Artificial Intelligence (cs.AI), linearized Bregman iterations, 65K10 (Primary) 68T07, 65D99, 41A63 (Secondary), 2025 OA procedure, Sparsity, Neural architecture search
FOS: Computer and information sciences, Numerical optimization and variational techniques, Nonlinear dimensionality reduction, math.NA, Computer Science - Artificial Intelligence, G.1.6, Multidimensional problems, Machine Learning (stat.ML), Statistics - Machine Learning, 41A63 (Secondary), Linearized Bregman iterations, FOS: Mathematics, nonlinear dimensionality reduction, Mathematics - Numerical Analysis, 65D99, cs.NA, Artificial neural networks and deep learning, Scientific machine learning, scientific machine learning, 65K10 (Primary) 68T07, I.2.6, sparsity, G.1.6; I.2.6, Numerical Analysis (math.NA), cs.AI, stat.ML, Numerical approximation and computational geometry (primarily algorithms), neural architecture search, Artificial Intelligence (cs.AI), linearized Bregman iterations, 65K10 (Primary) 68T07, 65D99, 41A63 (Secondary), 2025 OA procedure, Sparsity, Neural architecture search
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
