Explore >> Select a destination


You are here

kvfrans.com
| | blog.otoro.net
1.1 parsecs away

Travel
| | [AI summary] This text discusses the development of a system for generating large images from latent vectors, combining Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). It explores the use of Conditional Perceptual Neural Networks (CPPNs) to create images with specific characteristics, such as style and orientation, by manipulating latent vectors. The text also covers the ability to perform arithmetic on latent vectors to generate new images and the potential for creating animations by transitioning between different latent states. The author suggests future research directions, including training on more complex datasets and exploring alternative training objectives beyond Maximum Likelihood.
| | vxlabs.com
1.8 parsecs away

Travel
| | I have recently become fascinated with (Variational) Autoencoders and with PyTorch. Kevin Frans has a beautiful blog post online explaining variational autoencoders, with examples in TensorFlow and, importantly, with cat pictures. Jaan Altosaar's blog post takes an even deeper look at VAEs from both the deep learning perspective and the perspective of graphical models. Both of these posts, as well as Diederik Kingma's original 2014 paper Auto-Encoding Variational Bayes, are more than worth your time.
| | www.nicktasios.nl
2.1 parsecs away

Travel
| | In the Latent Diffusion Series of blog posts, I'm going through all components needed to train a latent diffusion model to generate random digits from the MNIST dataset. In the second post, we will bu
| | seanzhang.me
21.2 parsecs away

Travel
| Imagine that you have 2 curves in a 2-D space, how would you measure the similarity of these 2 curves?