Explore >> Select a destination


You are here

mertkavi.com
| | michael-lewis.com
2.8 parsecs away

Travel
| | This is a short summary of some of the terminology used in machine learning, with an emphasis on neural networks. I've put it together primarily to help my own understanding, phrasing it largely in non-mathematical terms. As such it may be of use to others who come from more of a programming than a mathematical background.
| | datadan.io
3.1 parsecs away

Travel
| | Linear regression and gradient descent are techniques that form the basis of many other, more complicated, ML/AI techniques (e.g., deep learning models). They are, thus, building blocks that all ML/AI engineers need to understand.
| | www.jeremymorgan.com
2.0 parsecs away

Travel
| | Want to learn about PyTorch? Of course you do. This tutorial covers PyTorch basics, creating a simple neural network, and applying it to classify handwritten digits.
| | yang-song.net
10.0 parsecs away

Travel
| This blog post focuses on a promising new direction for generative modeling. We can learn score functions (gradients of log probability density functions) on a large number of noise-perturbed data distributions, then generate samples with Langevin-type sampling. The resulting generative models, often called score-based generative models, has several important advantages over existing model families: GAN-level sample quality without adversarial training, flexible model architectures, exact log-likelihood ...