Papers
For natural-gradient (NG) methods:
-
Improving optimizers Shampoo and SOAP via Riemannian/proximal gradient descent (ICLR 2026): Paper, Code
-
Structured NG descent for deep learning (ICML 2023, ICML 2024): Manifold View, Code 2023; Bayesian View, Code 2024
-
NG descent for exponential-family mixtures (ICML 2019): Paper, Code
-
NG descent for Bayesian deep learning (ICML 2018): Paper, Code
-
NG variational inference for non-conjugate models (AI&Stats 2017): Paper, Code
For an introduction to NG methods, see my Blog Posts.
For more publications, see my Google Scholar page.