|
You are here |
www.depthfirstlearning.com | ||
| | | | |
jaketae.github.io
|
|
| | | | | In this short post, we will take a look at variational lower bound, also referred to as the evidence lower bound or ELBO for short. While I have referenced ELBO in a previous blog post on VAEs, the proofs and formulations presented in the post seems somewhat overly convoluted in retrospect. One might consider this a gentler, more refined recap on the topic. For the remainder of this post, I will use the terms "variational lower bound" and "ELBO" interchangeably to refer to the same concept. I was heavily inspired by Hugo Larochelle's excellent lecture on deep belief networks. | |
| | | | |
colinraffel.com
|
|
| | | | | ||
| | | | |
www.v7labs.com
|
|
| | | | | What are Generative Adversarial Networks and how do they work? Learn about GANs architecture and model training, and explore the most popular generative models variants and their limitations. | |
| | | | |
sirupsen.com
|
|
| | | [AI summary] The article provides an in-depth explanation of how to build a neural network from scratch, focusing on the implementation of a simple average function and the introduction of activation functions for non-linear tasks. It discusses the use of matrix operations, the importance of GPUs for acceleration, and the role of activation functions like ReLU. The author also outlines next steps for further exploration, such as expanding the model, adding layers, and training on datasets like MNIST. | ||