|
You are here |
www.jeremymorgan.com | ||
| | | | |
vxlabs.com
|
|
| | | | | I have recently become fascinated with (Variational) Autoencoders and with PyTorch. Kevin Frans has a beautiful blog post online explaining variational autoencoders, with examples in TensorFlow and, importantly, with cat pictures. Jaan Altosaar's blog post takes an even deeper look at VAEs from both the deep learning perspective and the perspective of graphical models. Both of these posts, as well as Diederik Kingma's original 2014 paper Auto-Encoding Variational Bayes, are more than worth your time. | |
| | | | |
www.nicktasios.nl
|
|
| | | | | In the Latent Diffusion Series of blog posts, I'm going through all components needed to train a latent diffusion model to generate random digits from the MNIST dataset. In this first post, we will tr | |
| | | | |
blog.paperspace.com
|
|
| | | | | Follow this tutorial to learn what attention in deep learning is, and why attention is so important in image classification tasks. We then follow up with a demo on implementing attention from scratch with VGG. | |
| | | | |
blog.evjang.com
|
|
| | | This is a tutorial on common practices in training generative models that optimize likelihood directly, such as autoregressive models and ... | ||