You are here |
francisbach.com | ||
| | | |
akosiorek.github.io
|
|
| | | | Machine learning is all about probability.To train a model, we typically tune its parameters to maximise the probability of the training dataset under the mo... | |
| | | |
fa.bianp.net
|
|
| | | | Backtracking step-size strategies (also known as adaptive step-size or approximate line-search) that set the step-size based on a sufficient decrease condition are the standard way to set the step-size on gradient descent and quasi-Newton methods. However, these techniques are typically not used for Frank-Wolfe-like algorithms. In this blog post I discuss a backtracking line-search for the Frank-Wolfe algorithm. | |
| | | |
fa.bianp.net
|
|
| | | | The Langevin algorithm is a simple and powerful method to sample from a probability distribution. It's a key ingredient of some machine learning methods such as diffusion models and differentially private learning. In this post, I'll derive a simple convergence analysis of this method in the special case when the ... | |
| | | |
francisbach.com
|
|
| |