Stein’s Method for Practical Machine Learning
Stein's method, due to
Charles M. Stein,
is a set of
remarkably powerful
theoretical techniques for
proving approximation and limit theorems in probability theory.
It has been mostly known to theoretical statisticians.
Recently, however,
it has been shown that some of the key ideas from Stein's method
can be naturally adopted to solve computational and statistical challenges in practical machine learning.
This project aims to harness Stein’s method for practical purposes,
with a focus on
develop new and efficient practical algorithms for
learning, inference and model evaluation of highly complex probabilistic graphical models and deep learning models.
Kernelized Stein Discrepancy
Kernelized Stein discrepancy (KSD), based on combining the classical Stein discrepancy with reproducing kernel Hilbert space (RKHS), allows us to access the compatibility between empirical data and probabilistic distributions, and provides a powerful tool for developing algorithms for model evaluation (goodness-of-fit test), as well as learning and inference in general. Unlike the traditional divergence measures (such as KL, Chi-square divergence), KSD does not require to evaluate the normalization constant of the distribution, and can be applied even for the intractable, unnormalized distributions widely used in modern machine learning.
- A Kernelized Stein Discrepancy for Goodness-of-fit Tests and Model Evaluation
Liu, Lee, Jordan; ICML, 2016,
[A short note],
[code: matlab, R ]
[See more details here>>].
Stein Variational Gradient Descent
Based on exploiting an interesting connection between Stein discrepancy and KL divergence,
we derive a new form of variational inference algorithm, called Stein variational gradient descent (SVGD),
that mixes the advantages of variational inference, Monte Carlo, quasi Monte Carl and gradient descent (for MAP).
SVGD provides a new powerful tool for attacking the inference and learning challenges in graphical models and probabilistic deep learning, especially when there is a need for getting diverse outputs
to capture the posterior uncertainty in the Bayesian framework.
- Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm
Liu, Wang; NeurIPS, 2016 [code]
- Stein Variational Gradient Descent as Gradient Flow
Liu; NeurIPS, 2017
- Stein Variational Gradient Descent as Moment Matching
Liu; NeurIPS, 2018
[See more details here>>].
Slides
Probabilistic Learning and Inference Using Stein's Method [slides, slides]
A Kernelized Stein Discrepancy for Goodness-of-fit Tests and Model Evaluation
[ICML2016 slides]
Papers
- A Kernelized Stein Discrepancy for Goodness-of-fit Tests and Model Evaluation
Liu, Lee, Jordan; ICML, 2016
[code: matlab, R ]
- Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm
Liu, Wang; NeurIPS, 2016 [code]
- Learning to Draw Samples: With Application to Amortized MLE for Generative Adversarial Learning
Wang, Liu; preprint 2016
[code]
- Two methods for Wild Variational Inference
Liu, Feng; preprint, 2016
- Black-box Importance Sampling
Liu, Lee; AISTATS, 2017
- Learning Deep Energy Models: Contrastive Divergence vs. Amortized MLE
Liu, Wang, 2017
- Learning to Draw Samples with Amortized Stein Variational Gradient Descent
Feng et al. UAI 2017
- Stein Variational Gradient Descent as Gradient Flow
Liu; NeurIPS, 2017
- Stein Variational Policy Gradient
Yang et al. UAI. 2017
- Stein Variational Gradient Descent as Moment Matching
Liu; NeurIPS, 2018
- Stein Variational Gradient Descent Without Gradient
Han, Liu; ICML, 2018
- Stein variational gradient descent with matrix-valued kernels
Wang, Tang, Bajaj, Liu, NeurIPS, 2019
- Nonlinear Stein Variational Gradient Descent for Learning Diversified Mixture Models
Wang, Liu, ICML, 2019
- Learning Self-Imitating Diverse Policies
Gangwani et al. ICLR. 2019
- Stein Variational Inference for Discrete Distributions
Han, et al. AISTATS, 2020
- Profiling Pareto Front With Multi-Objective Stein Variational Gradient Descent
Liu et al. NeurIPS 2021
- Sampling with Trustworthy Constraints: A Variational Gradient Framework
Liu et al. NeurIPS 2021
- Sampling in Constrained Domains with Orthogonal-Space Variational Gradient Descent
Zhang et al. NeurIPS 2022
- Goodness-of-Fit Testing for Discrete Distributions via Stein Discrepancy
Yang et al. ICML 2018
- Stein’s Method Meets Computational Statistics: A Review of Some Recent Developments
Anastasiou et al. 2022
Informal Notes & Misc
- A Short Note on Kernelized Stein Discrepancy
Liu, 2016
- Stein Variational Gradient Descent: Theory and Applications
Liu, NeurIPS Workshop on Advances in Approximate Bayesian Inference, 2016
- Learning to Sample Using Stein Discrepancy
Wang, Feng, Liu, NeurIPS Workshop on Bayesian Deep Learning, 2016
|