|
You are here |
jaykmody.com | ||
| | | | |
www.johndcook.com
|
|
| | | | | The most obvious way to compute the soft maximum can easily fail due to overflow or underflow. The soft maximum of x and y is defined by g(x, y) = log( exp(x) + exp(y) ). The most obvious way to turn the definition above into C code would be double SoftMaximum(double x, double y) { | |
| | | | |
www.codeproject.com
|
|
| | | | | Describes why the most obvious way of evaluating functions may be bad and how to do better | |
| | | | |
jxmo.io
|
|
| | | | | A primer on variational autoencoders (VAEs) culminating in a PyTorch implementation of a VAE with discrete latents. | |
| | | | |
blog.paperspace.com
|
|
| | | Follow this tutorial to learn what attention in deep learning is, and why attention is so important in image classification tasks. We then follow up with a demo on implementing attention from scratch with VGG. | ||