Periodic Reporting for period 1 - PrSc-HDBayLe (Provable Scalability for high-dimensional Bayesian Learning)
Período documentado: 2023-05-01 hasta 2025-10-31
Throughout the project, the research team involved in the project is tackling several open problems in Bayesian computation. In the first scientific theme, we develop a rigorous mathematical theory of the computational behavior of Bayesian hierarchical models, analyzing algorithmic performance in sparse, structured settings and advancing convergence diagnostics via dimensionality reduction techniques for MCMC algorithms. In the second theme, we focus on the computational cost of high-dimensional Bayesian inference, linking Bayesian asymptotics with Markov chain theory to design provably scalable algorithms. Finally, we study robustness of commonly used algorithms to data heterogeneity, model misspecification, and tail behavior, leading to the development of more stable and generalizable computational techniques.
The results have direct implications on the design of novel and more scalable computational schemes, as well as on the optimization of existing ones. Focus is given to develop algorithms with provably linear overall cost both in the number of datapoints and unknown parameters. The project contributes to significantly reduce the gap between theory and practice in Bayesian computation and allow practitioners to fully benefit of the huge potential of the Bayesian and probabilistic modeling in popular data science pipelines.
- New interdisciplinary links between Bayesian asymptotics and MCMC complexity theory, under random data-generating assumptions. This allows to simplify and enable novel complexity results, for example in the context of large hierarchical models and latent variable models.
- Connections between Bayesian computation and random graph theory, used to analyze coordinate-wise algorithms for hierarchical models and answer fundamental questions regarding computational aspects of Bayesian hierarchical models, such as: how do sampling and optimization coordinate-wise algorithms perform on for increasingly larger and sparser models? How does the observation pattern affect convergence speed? Which algorithms perform well on average-case random designs? These results offer practical insights into the design and optimization of Gibbs Samplers, optimization methods based on backfitting, and Coordinate Ascent Variational Inference methods, especially for crossed and nested models.
- The first fully explicit, non-asymptotic analysis of the convergence rate of the popular and classical Gibbs sampler algorithm under log-concavity, in the form of an entropy contraction results, with interesting and unexpected implications for the complexity theory of log-concave sampling methods.
- Novel results on zero-order parallel sampling methods, including results on the fundamental limitations of multiproposal MCMC methods, showing that (under appropriate assumptions) they can only achieve a logarithmic speed-up in the number of parallel workers; and the development of novel methodologies based on parallel-in-time integrators that provably achieve polynomial speed-ups.
- We developed various novel mathematical techniques, or improved existing ones, in order to analyze and improve Bayesian computational algorithms. For example, we introduced a new comparison techniques for Markov chains, overcoming limitations of classical methods like Peskun ordering. Our approach, based on “conditional capacitance,” enables rigorous comparison of hybrid samplers (e.g. Metropolis-within-Gibbs) with ideal counterparts, providing precise assessments of performance trade-offs.
- We developed scalable Bayesian computation tools, such as: conjugate gradient samplers for generalized linear mixed models (GLMMs); novel partially-factorized variational inference algorithms for improved uncertainty quantification; novel algorithms for zeroth-order parallel sampling; and mixture importance sampling estimators for Bayesian cross-validation criteria; and others. The corresponding codes are freely available online (e.g. through Github repositories, R software packages, etc).