Explore >> Select a destination


You are here

blog.evjang.com
| | teddykoker.com
2.9 parsecs away

Travel
| | Gradient-descent-based optimizers have long been used as the optimization algorithm of choice for deep learning models. Over the years, various modifications to the basic mini-batch gradient descent have been proposed, such as adding momentum or Nesterovs Accelerated Gradient (Sutskever et al., 2013), as well as the popular Adam optimizer (Kingma & Ba, 2014). The paper Learning to Learn by Gradient Descent by Gradient Descent (Andrychowicz et al., 2016) demonstrates how the optimizer itself can be replac...
| | questionableengineering.com
5.7 parsecs away

Travel
| | John W Grun AbstractIn this paper, a manually implemented LeNet-5 convolutional NN with an Adam optimizer written in Numpy will be presented. This paper will also cover a description of the data use
| | kevinlynagh.com
5.6 parsecs away

Travel
| | [AI summary] The author discusses their experience developing a simple neural network for sensor data processing on a microcontroller, highlighting challenges with quantization and inference optimization.
| | cset.georgetown.edu
49.9 parsecs away

Travel
| Place to find CSET's publications, reports, and people