Periodic Reporting for period 4 - GTBB (General theory for Big Bayes)
Periodo di rendicontazione: 2023-09-01 al 2025-03-31
The main objectives of the project are (briefly):
1. Asymptotic analysis of the posterior distribution of complex high dimensional models
2. Asymptotic analysis of approximate computational Bayes approaches – Interractions between asymptotics and computations.
1. Complex modelling and inference. We studied 4 families of complex models: Point processes, Networks, Diffusion models and Deep Neural networks. We also investigated Bayesian manifold learning.
Point processes: We developed theory and methodology for Bayesian nonparametric methods for multivariate nonlinear Hawkes processes and estimation of the graph of interactions: published in Bernoulli and under minor revision at JMLR. We developed and studied Bayesian nonparametric methods for spatial point processes with covariates, published in Bernoulli. We also used discrete times Hawkes processes to model the early stage of the Covid epidemic, published in PLoSone and for conflict data, submitted on arxiv.
Networks. We developed new generative models for networks, we studied the asymptotic behaviour of a wide class of models called the graphex models and the asymptotic behaviour of a Bayesian approach. These have been published in Bernoulli, Advances in applied probability, EJS and ECML PKDD.
diffusion models: We studied Bayesian nonparametric methods for estimating the drift vector field in a multi-dimensional, published in the Annals of Statistics.
Deep neural networks: We studied the impact of initialisation in the large width large depth regimes, leading to 3 publications in AISTAT, Neurips workshop, TMLR and 1 preprint. We also studied the interplay between penalties and priors in the context of DNN, submitted on arxiv. We compared Gaussian processes and deep learning approaches for Gaussian regression, showing the superiority of the latter in some situations, published in Neurips. We studied convergence rates of score based diffusion generative models based on DNN under the manifold hypothesis, submitted on arxiv..
Bayesian manifold learning. We developed and studied a Bayesian procedure for estimating the density living near a manifold. We also studied regression under the manifold hypothesis, under both extrinsic an intrinsic Gaussian process priors. We obtained minimax rates of convergence for Bayesian nonparametric regression using based graph Laplacians in large dimension,. These works have been published respectively in Annals of Statistics, Neurips and (minor revision) JMLR.
2.Semiparametric inference
(i) Non standard losses: We studied concentration of the posterior distribution under local losses; published in JASA and Bernoulli. We also studied the multivariate deconvolution problem in max-sliced Wasserstein, from a frequentist a Bayesian perspective, published in the Annals of Statistics.
(ii) Bernstein von Mises theorems (BvM): we proposed two different Bayesian quasi posterior methods to ensure that the resulting distributions behave well for multiple targets, published in EJS and in JRSS B. Finally we derived a BvM for a class on nonparametric mixture models, submitted on arxiv.
(iii) Empirical Bayes. We characterised the behaviour of the Maximum marginal likelihood estimator in various irregular scenarios
3. Approximate Bayesian approaches:
(i) We derived robust Bayesian methods using Generalized Bayesian inference, published in AISTAT. We also generalised the notion of minimax convergence rates under random sample sizes,submitted on arxiv.
(ii) We studied 3 families of approximate Bayesian computations: coresets, ABC and variational algorithms, published in Neurips, arxiv and JMLR (under minor revision) respectively.
(ii) We studied efficient algorithms to compute the marginal likelihoods in mixture models investigating also distributed inference, published in a thesis.
Bayesian Manifold learning. We developped a mathematical framework to describe anisotropic smoothness along a manifold and proved minimax rates for Bayesian nonparametric Gaussian mixtures. We also obtained adaptive minimax rates for graph Laplacian approaches, constructing a new approximation strategy for regression functions in this context.
Deep neural network and Bayesian inference: We showed for various architectures that initialisation can have a dramatic effect on the learning ability of the network and we proposed a rescaling which allows for better stabilisation. We showed the inability of Gaussian processes to learn compositional functions. Finally we proved that diffusion generative models did not suffer the curse of the ambiant dimensional under the manifold hypothesis.
semi-parametric inference. We have proposed two alternative methods to construct posterior distributions in semiparametric problems to obtain efficient estimation and correct uncertainty for many functionals of interest and models
We obtained an inversion inequality which allows to understand better Wasserstein neighbourhood of mixing densities in deconvolution problems.
We also developed a theory to study posterior distributions for functions whose regularity is spatially varying.
We obtained an asymptotic description of the posterior distribution for a class of sparse networks which allows for uncertainty quantification, covering also the misspecified case.
We developped variational methods for large non linear Hawkes processes. We improved on existing coresets approaches to approximate Bayesian methods for large data sets studied more complex scenarios for ABC algorithms.