Back

Variational AI: Training Better Models at Lower Cost

KAKENHI Grant-in-Aid for Scientific Research (A), FY 2026-2031. Project no. 26H02541.

About the Project

Contemporary AI models are highly capable but with increasing scale they become costly to train, maintain and update. In this project, we develop new foundational learning algorithms for neural networks to drastically reduce AI's training cost by improving its ability to adapt and continually learn from experiences in an open world. A focus is on variational Bayesian-like training methods, which can address such issues and our recent research shows them to be effective to large neural networks. This project aims to further advance variational learning methods for large deep networks and to demonstrate their application towards sustainable AI training.

Research and Open Positions

The project focuses on the following research directions:

  • Effective New Variational Learning Algorithms for Large Deep Networks, (e.g., pre-training)
  • Variational Methods for Adaptive, Continual, Distributed and Reinforcement Learning in Deep Networks
  • Mechanistic Interpretability, Sensitivity Analysis and Influence Functions
  • Theoretical Foundations of Variational Learning (PAC-Bayes, Optimization in Spaces of Measures, etc.)

I am looking to hire two interns (~6 months fully funded internship) to work with me at RIKEN AIP in Tokyo in FY2026. If you are interested or looking to collaborate, please get in touch by email:

Tutorials, Teaching and Other Materials

  • The Improved Variational Online Newton (IVON) Optimizer, github link.

Publications