Research

Bayesian modeling plays a central role in many scientific disciplines and branches of machine learning, where it is of interest to quantify uncertainty. A persistent bottleneck in Bayesian analysis is the computation of the posterior distribution. This problem is exacerbated by the increasing complexity of models and the growing size of data. To address this challenge, my research program centers on developing algorithms for Bayesian inference, which can take advantage of modern computation.

A selection of research papers is available on my home page. Here, I provided a more exhaustive list which includes conference posters and other technical reports. I ordered these into four broad themes:

  • Markov chain Monte Carlo
  • Variational inference
  • Bayesian workflow and scientific applications
  • Automatic differentiation

Markov chain Monte Carlo

MCMC is the workhorse of Bayesian inference. In recent years, it has become possible to efficiently parallelize MCMC by running hundreds, even thousands, of Markov chains on hardware accelerators such as GPUs. My work investigates shifts in the practice and analysis of MCMC as we transition from CPU- to GPU-based hardware.

Papers:

  • (2024) Nested $\widehat R$: Assessing the convergence of Markov chain Monte Carlo when running many short chains. C. Margossian, M. Hoffman, P. Sountsov, L. Riou-Durand, A. Vehtari and A. Gelman. Bayesian Analysis [article, pdf, code]
  • (2024) Listening to the Noise: Blind Denoising with Gibbs Diffusion. D. Heurtel-Depeiges, C. Margossian, R. Ohana and B. Regaldo-Saint Blancard. PMLR: International Conference on Machine Learning [article, pdf]
  • (2024) For how many iterations should we run Markov chain Monte Carlo? C. Margossian and A. Gelman. Handbook of Markov chain Monte Carlo, (upcoming) 2nd Edition. [pdf]
  • (2022) Adaptive Tuning for Metropolis Adjusted Langevin Trajectories. L. Riou-Durand, P. Sountsov, J. Vogrinc, C. Margossian and S. Power. PMLR: Artificial Intelligence and Statistics [article, pdf, code]
  • (2020) Hamiltonian Monte Carlo using an adjoint-differentiated Laplace approximation: Bayesian inference for latent Gaussian models and beyond. C. Margossian, A. Vehtari, D. Simpson and R. Agrawal. Advances in Neural Information Processing Systems. [article, poster, pdf, code]

Posters and technical reports:

  • (2024) Monitoring Nonstationary Variance to Assess Convergence of MCMC. E. Mokel and C. Margossian. International Society of Bayesian Analysis (ISBA) world meeting. Best poster award.
  • (2023) Parallelization for Markov chains with heterogeneous runtimes. S. du Ché and C. Margossian. BayesComp. [poster]

Variational Inference

VI is another workhorse of Bayesian analysis. The modus operandi of VI is to find a parametric approximation of the posterior by solving an optimization problem. We can obtain fast optimization by restricting the family of approximations, for instance by using factorized and/or amortized distributions. VI is then suitable for the constrained computational budget imposed by large problems. On the other hand, a restricted approximation cannot perfectly match the posterior. A major axis of my research studies whether an imperfect approximation can still capture key features of the posterior, such as its mean, variance or entropy, even if this approximation is misleading in other ways. Concurently, I am interested in developing VI with highly expressive families of approximations, and building tractable optimization problems for such families.

Papers:

  • (2024) Variational Inference in Location-Scale Families: Exact Recovery of the Mean and Correlation Matrix. C.Margossian and L. Saul. [pdf]
  • (2024) Variational Inference for Uncertainty Quantification: an Analysis of Trade-Offs. C. Margossian, L. Pillaud-Vivien and L. Saul. [pdf, Code]
  • (2024) EigenVI: score-based variational inference with orthogonal function expansions. D. Cai, C. Modi, C. Margossian, R. Gower, D. Blei and L. Saul. Advances in Neural Information Processing Sysmtes (spotlight)
  • (2024) Batch and match: black-box variational inference with a score-based divergence. D. Cai, C. Modi, L. Pillaud-Vivien, C. Margossian, R. Gower, D. Blei and L. Saul. PMLR: International Conference on Machine Learning (spotlight) [article, pdf, code]
  • (2024) Amortized Variational Inference: when and why? C. Margossian and D. Blei. PMLR: Uncertainty in Artificial Intelligence [article,pdf, code, talk]
  • (2023) Variational Inference with Gaussian Score Matching. C. Modi, C. Margossian, Y. Yao, R. Gower, D. Blei and L. Saul. Advances in Neural Information Processing Systems [article, pdf, code]
  • (2023) The Shrinkage-Delinkage Trade-off: An Analysis of Factorized Gaussian Approximations for Variational Inference. C. Margossian and L. Saul. PMLR: Uncertainty in Artificial Intelligence (Oral) [paper, pdf, code, talk]

Bayesian workflow and scientific applications

Much of my work is motivated by Bayesian modeling problems in pharmacometrics, epidemiology, and (astro)physics. A defining characteristic of these models is that they integrate a scientific (forward) model, often based on a system of differential equations, inside a statistical model.

Papers:

  • (2022) Flexible and efficient Bayesian pharmacometrics modeling using Stan and Torsten, Part I. C. Margossian, Y. Zhang and W. Gillespie. CPT: Pharmacometrics & Systems Pharmacology. [article, pdf, code, software]
  • (2021) Bayesian Workflow for disease transmission modeling in Stan. L. Grinsztajn, E. Semenova, C. Margossian and J. Riou. Statistics in medicine. [article, pdf, code, talk]
  • (2021) Simulating Ising and Potts models at critical and cold temperatures using auxiliary Gaussian variables. C. Margossian and S. Mukherjee. [pdf]
  • (2020) Estimation of SARS-CoV-2 mortality during the early stages of an epidemic: a modelling study in Hubei, China and six regions of Europe. J. Riou et al. PLOS Medicine. [article, preprint]
  • (2020) Bayesian Workflow. A. Gelman et al. [pdf]
  • (2014) Planet Hunters. VII. Discovery of a new low-mass, low-density (PH3 C) orbiting Kepler-289 with mass measurements of two additional plamets of two additional planets (PH3 B and D). J. Schmitt et al. Astrophysical Journal. [article]

Posters and technical reports:

  • (2021) Developping a model of SARS-CoV-2 viral dynamics under monoclonal antibody treatment. A. Marc, M. Kerioui, C. Margossian, J. Bertrand, P. Maisonnasse, Y. Aldon, R. Sanders, M. Van Gils, R. Le Grand and J. Guedj. Population Approach Group in Europe.
  • (2021) Solving ODEs in a Bayesian context: challenges and opportunities. C. Margossian, L. Zhang, S. Weber and A. Gelman. Population Approach Group in Europe. [poster]
  • (2020) Bayesian model of planetary motion: exploring ideas for a modeling workflow when dealing with ordinary differential equations and multimodality. C. Margossian and A. Gelman. Stan Case Studies 7. [article, code]
  • (2018) Computing Steady states with Stan’s nonlinear algebraic solver. C. Margossian. Stan Con Asilomar 2018. [article, code, talk]
  • (2017) Gaining Efficiency by combining analytical and numerical methods to solve ODEs: implementation in Stan and application to Bayesian PK/PD modeling. C. Margossian and W. Gillespie. Journal of Pharmacokinetics and Pharmacodynamics, presented at the American Conference on Pharmacometrics 8. [poster]
  • (2017) Differential equation based models in Stan. C. Margossian and W. Gillespie. Stan Con 2017. [article, code, talk]
  • (2016) Stan functions for pharmacometrics. C. Margossian and W. Gillespie. Journal of Pharmacokinetics and Pharmacodynamics, presented at the American Conference on Pharmacometrics 7. [poster]

Automatic differentiation

Fitting models in high dimension, whether it be through sampling or optimization, requires gradient-based algorithms. Automatic differentiation (AD) provides a seamless and scalable manner to compute gradients (and higher-order derivatives). Reverse-mode AD has played a defining role over the last decade, enabling much of the progress in deep learning, AI, and probabilistic programming. Many popular software, including Stan, JAX, and PyTorch, are all AD libraries at heart.

I have contributed extensively to Stan’s AD library in C++, in particular the propagation of derivatives through implicit functions. These implicit functions arise in scientific forward models (e.g. ODEs, algebraic equations) and marginalization techniques (e.g. embedded Laplace approximation, hidden Markov models).

Papers:

  • (2019) A Review of automatic differentiation and its efficient implementation. C. Margossian. Wiley interdisciplinary reviews: data mining and knowledge discovery. [article, pdf]

Technical reports:

  • (2023) General adjoint-differentiated Laplace approximation. C. Margossian. [pdf]
  • (2022) Approximate Bayesian inference for latent Gaussian models in Stan – two years later. C. Margossian, S. Bronder, A. Vehtari, D. Simpson and R. Agrawal. [notebook, code]
  • (2022) Efficient Automatic Differentation of Implicit Functions. C. Margossian and M. Betancourt [pdf]
  • (2020) The Discrete adjoint method: efficient derivatives for functions of discrete sequences. M. Betancourt, C. Margossian and V. Leos-Barajas. [pdf]