Explore >> Select a destination


You are here

windowsontheory.org
| | newvick.com
2.0 parsecs away

Travel
| | I've been working my way through Andrej Karpathy's 'spelled-out intro to backpropagation', and this post is my recap of how backpropagation works.
| | bytepawn.com
3.2 parsecs away

Travel
| | I will show how to solve the standard A x = b matrix equation with PyTorch. This is a good toy problem to show some guts of the framework without involving neural networks.
| | theorydish.blog
1.8 parsecs away

Travel
| | The chain rule is a fundamental result in calculus. Roughly speaking, it states that if a variable $latex c$ is a differentiable function of intermediate variables $latex b_1,\ldots,b_n$, and each intermediate variable $latex b_i$ is itself a differentiable function of $latex a$, then we can compute the derivative $latex \frac{{\mathrm d} c}{{\mathrm d} a}$ as...
| | iclr-blogposts.github.io
45.1 parsecs away

Travel
| Diffusion Models, a new generative model family, have taken the world by storm after the seminal paper by Ho et al. [2020]. While diffusion models are often described as a probabilistic Markov Chains, their underlying principle is based on the decade-old theory of Stochastic Differential Equations (SDE), as found out later by Song et al. [2021]. In this article, we will go back and revisit the 'fundamental ingredients' behind the SDE formulation and show how the idea can be 'shaped' to get to the modern form of Score-based Diffusion Models. We'll start from the very definition of the 'score', how it was used in the context of generative modeling, how we achieve the necessary theoretical guarantees and how the critical design choices were made to finally arri...