Explore >> Select a destination


You are here

teddykoker.com
| | blog.research.google
1.8 parsecs away

Travel
| | [AI summary] This blog post introduces Stochastic Re-weighted Gradient Descent (RGD), a novel optimization algorithm that improves deep neural network performance by re-weighting data points during training based on their difficulty, enhancing generalization and robustness against data distribution shifts.
| | bdtechtalks.com
1.8 parsecs away

Travel
| | Gradient descent is the main technique for training machine learning and deep learning models. Read all about it.
| | justindomke.wordpress.com
2.4 parsecs away

Travel
| | In 2012, I wrote a paper that I probably should have called "truncated bi-level optimization". I vaguely remembered telling the reviewers I would release some code, so I'm finally getting around to it. The idea of bilevel optimization is quite simple. Imagine that you would like to minimize some function $latex L(w)$. However, $latex L$...
| | trishagee.com
4.7 parsecs away

Travel
| Find out where to catch Trisha Gee in the autumn of 2024