|
You are here |
fa.bianp.net | ||
| | | | |
www.randomservices.org
|
|
| | | | | [AI summary] The text covers various topics in probability and statistics, including continuous distributions, empirical density functions, and data analysis. It discusses the uniform distribution, rejection sampling, and the construction of continuous distributions without probability density functions. The text also includes data analysis exercises involving empirical density functions for body weight, body length, and gender-specific body weight. | |
| | | | |
yang-song.net
|
|
| | | | | This blog post focuses on a promising new direction for generative modeling. We can learn score functions (gradients of log probability density functions) on a large number of noise-perturbed data distributions, then generate samples with Langevin-type sampling. The resulting generative models, often called score-based generative models, has several important advantages over existing model families: GAN-level sample quality without adversarial training, flexible model architectures, exact log-likelihood ... | |
| | | | |
jxmo.io
|
|
| | | | | A primer on variational autoencoders (VAEs) culminating in a PyTorch implementation of a VAE with discrete latents. | |
| | | | |
blog.keras.io
|
|
| | | [AI summary] The text discusses various types of autoencoders and their applications. It starts with basic autoencoders, then moves to sparse autoencoders, deep autoencoders, and sequence-to-sequence autoencoders. The text also covers variational autoencoders (VAEs), explaining their structure and training process. It includes code examples for each type of autoencoder and mentions the use of tools like TensorBoard for visualization. The VAE section highlights how to generate new data samples and visualize the latent space. The text concludes with references and a note about the potential for further topics. | ||