Explore >> Select a destination


You are here

windowsontheory.org
| | hackmd.io
3.7 parsecs away

Travel
| |
| | francisbach.com
2.0 parsecs away

Travel
| | [AI summary] This text discusses the scaling laws of optimization in machine learning, focusing on asymptotic expansions for both strongly convex and non-strongly convex cases. It covers the derivation of performance bounds using techniques like Laplace's method and the behavior of random minimizers. The text also explains the 'weird' behavior observed in certain plots, where non-strongly convex bounds become tight under specific conditions. The analysis connects theoretical results to practical considerations in optimization algorithms.
| | iclr-blogposts.github.io
3.8 parsecs away

Travel
| | Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle
| | goodfire.ai
15.9 parsecs away

Travel
| Goodfire is an AI research company building practical interpretability tools for safe and reliable generative models.