Explore >> Select a destination


You are here

tiao.io
| | www.nicktasios.nl
0.3 parsecs away

Travel
| | In the Latent Diffusion Series of blog posts, I'm going through all components needed to train a latent diffusion model to generate random digits from the MNIST dataset. In the second post, we will bu
| | jxmo.io
0.4 parsecs away

Travel
| | A primer on variational autoencoders (VAEs) culminating in a PyTorch implementation of a VAE with discrete latents.
| | blog.keras.io
0.5 parsecs away

Travel
| | [AI summary] The text discusses various types of autoencoders and their applications. It starts with basic autoencoders, then moves to sparse autoencoders, deep autoencoders, and sequence-to-sequence autoencoders. The text also covers variational autoencoders (VAEs), explaining their structure and training process. It includes code examples for each type of autoencoder and mentions the use of tools like TensorBoard for visualization. The VAE section highlights how to generate new data samples and visualize the latent space. The text concludes with references and a note about the potential for further topics.
| | jaketae.github.io
11.5 parsecs away

Travel
| In this short post, we will take a look at variational lower bound, also referred to as the evidence lower bound or ELBO for short. While I have referenced ELBO in a previous blog post on VAEs, the proofs and formulations presented in the post seems somewhat overly convoluted in retrospect. One might consider this a gentler, more refined recap on the topic. For the remainder of this post, I will use the terms "variational lower bound" and "ELBO" interchangeably to refer to the same concept. I was heavily inspired by Hugo Larochelle's excellent lecture on deep belief networks.