|
You are here |
siboehm.com | ||
| | | | |
dennybritz.com
|
|
| | | | | All the code is also available as an Jupyter notebook on Github. | |
| | | | |
www.paepper.com
|
|
| | | | | New blog series: Deep Learning Papers visualized This is the first post of a new series I am starting where I explain the content of a paper in a visual picture-based way. To me, this helps tremendously to better grasp the ideas and remember them and I hope this will be the same for many of you as well. Today's paper: Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour by Goyal et al. The first paper I've chosen is well-known when it comes to training deep learning models on multiple GPUs. Here is the link to the paper of Goyal et al. on arxiv. The basic idea of the paper is this: when you are doing deep learning research today, you are using more and more data and more complex models. As the complexity and size rises, of course also the computational... | |
| | | | |
www.nicktasios.nl
|
|
| | | | | In the Latent Diffusion Series of blog posts, I'm going through all components needed to train a latent diffusion model to generate random digits from the MNIST dataset. In this first post, we will tr | |
| | | | |
saturncloud.io
|
|
| | | By combining Dask and PyTorch you can easily speed up training a model across a cluster of GPUs. But how much of a benefit does that bring? This blog post finds out! | ||