Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Robust training of implicit generative models for multivariate and heavy-tailed distributions with an invariant statistical loss

Created by
  • Haebom

Author

Jos e Manuel de Frutos, Manuel A. Vazquez, Pablo Olmos, Joaqu in M iguez

Outline

To address the unstable learning dynamics and mode-loss issues of existing implicit generative models, this paper proposes Pareto-ISL, an extension of the invariant statistical loss (ISL) method that accurately models the tails of a distribution along with the central features. To overcome the limitation of existing ISL, which is limited to one-dimensional data, we propose a generator using the Generalized Pareto Distribution (GPD) and a novel loss function suitable for multidimensional data using random projections. Experiments demonstrate its performance in multidimensional generative modeling and demonstrate its potential as a pre-training technique for GANs to prevent mode collapse. In particular, we focus on effectively handling heavy-tailed distributions encountered in real-world phenomena.

Takeaways, Limitations

Takeaways:
We present a new method that effectively addresses the limitations of existing implicit generative models, such as unstable learning and mode loss.
Presenting the possibility of effective generative modeling for multidimensional data with heavy-tailed distributions.
Suggesting the possibility of improving performance and preventing mode collapse by using it as a pre-training technique for GAN.
It shows robust performance across various hyperparameter settings.
Limitations:
Further research is needed on the efficiency and accuracy of multidimensional expansion methods using random projections.
Additional experiments and comparative analyses on various types of datasets are needed.
A more detailed analysis of the computational complexity of the proposed method is needed.
👍