|
You are here |
dmitryulyanov.github.io | ||
| | | | |
blog.google
|
|
| | | | | Neural networks can train computers to learn in a way similar to humans. Googler Maithra Raghu explains how they work. | |
| | | | |
programmathically.com
|
|
| | | | | Sharing is caringTweetIn this post, we develop an understanding of why gradients can vanish or explode when training deep neural networks. Furthermore, we look at some strategies for avoiding exploding and vanishing gradients. The vanishing gradient problem describes a situation encountered in the training of neural networks where the gradients used to update the weights [] | |
| | | | |
coen.needell.org
|
|
| | | | | In my last post on computer vision and memorability, I looked at an already existing model and started experimenting with variations on that architecture. The most successful attempts were those that use Residual Neural Networks. These are a type of deep neural network built to mimic specific visual structures in the brain. ResMem, one of the new models, uses a variation on ResNet in its architecture to leverage that optical identification power towards memorability estimation. M3M, another new model, ex... | |
| | | | |
marcospereira.me
|
|
| | | In this post we summarize the math behind deep learning and implement a simple network that achieves 85% accuracy classifying digits from the MNIST dataset. | ||