
arXiv: 2212.11481
The modeling of probability distributions, specifically generative modeling and density estimation, has become an immensely popular subject in recent years by virtue of its outstanding performance on sophisticated data such as images and texts. Nevertheless, a theoretical understanding of its success is still incomplete. One mystery is the paradox between memorization and generalization: In theory, the model is trained to be exactly the same as the empirical distribution of the finite samples, whereas in practice, the trained model can generate new samples or estimate the likelihood of unseen samples. Likewise, the overwhelming diversity of distribution learning models calls for a unified perspective on this subject. This paper provides a mathematical framework such that all the well-known models can be derived based on simple principles. To demonstrate its efficacy, we present a survey of our results on the approximation error, training error and generalization error of these models, which can all be established based on this framework. In particular, the aforementioned paradox is resolved by proving that these models enjoy implicit regularization during training, so that the generalization error at early-stopping avoids the curse of dimensionality. Furthermore, we provide some new results on landscape analysis and the mode collapse phenomenon.
fixed typos
FOS: Computer and information sciences, Artificial intelligence, Computer Science - Machine Learning, generative modeling, memorization, generalization error, Machine Learning (stat.ML), Nonparametric inference, Machine Learning (cs.LG), Statistics - Machine Learning, density estimation, implicit regularization, Linear function spaces and their duals, 68T07, 62G05, 60-08
FOS: Computer and information sciences, Artificial intelligence, Computer Science - Machine Learning, generative modeling, memorization, generalization error, Machine Learning (stat.ML), Nonparametric inference, Machine Learning (cs.LG), Statistics - Machine Learning, density estimation, implicit regularization, Linear function spaces and their duals, 68T07, 62G05, 60-08
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 1 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
