Explore >> Select a destination


You are here

www.jeremykun.com
| | francisbach.com
4.2 parsecs away

Travel
| | [AI summary] This text discusses the scaling laws of optimization in machine learning, focusing on asymptotic expansions for both strongly convex and non-strongly convex cases. It covers the derivation of performance bounds using techniques like Laplace's method and the behavior of random minimizers. The text also explains the 'weird' behavior observed in certain plots, where non-strongly convex bounds become tight under specific conditions. The analysis connects theoretical results to practical considerations in optimization algorithms.
| | jeremykun.com
2.1 parsecs away

Travel
| | Hard to believe Sanjeev Arora and his coauthors consider it"a basic tool [that should be] taught to all algorithms students together with divide-and-conquer, dynamic programming, and random sampling."Christos Papadimitriou calls it"so hard to believe that it has been discovered five times and forgotten." It has formed the basis of algorithms inmachine learning, optimization, game theory,
| | fa.bianp.net
4.2 parsecs away

Travel
| | The Langevin algorithm is a simple and powerful method to sample from a probability distribution. It's a key ingredient of some machine learning methods such as diffusion models and differentially private learning. In this post, I'll derive a simple convergence analysis of this method in the special case when the ...
| | tiao.io
17.6 parsecs away

Travel
| An in-depth practical guide to variational encoders from a probabilistic perspective.