|
You are here |
www.hhyu.org | ||
| | | | |
bytepawn.com
|
|
| | | | | I will show how to solve the standard A x = b matrix equation with PyTorch. This is a good toy problem to show some guts of the framework without involving neural networks. | |
| | | | |
bdtechtalks.com
|
|
| | | | | Gradient descent is the main technique for training machine learning and deep learning models. Read all about it. | |
| | | | |
michael-lewis.com
|
|
| | | | | This is a short summary of some of the terminology used in machine learning, with an emphasis on neural networks. I've put it together primarily to help my own understanding, phrasing it largely in non-mathematical terms. As such it may be of use to others who come from more of a programming than a mathematical background. | |
| | | | |
mathematicaloddsandends.wordpress.com
|
|
| | | The function $latex f(x) = x \log x$ occurs in various places across math/statistics/machine learning (e.g. in the definition of entropy), and I thought I'd put a list of properties of the function here that I've found useful. Here is a plot of the function: $latex f$ is defined on $latex (0, \infty)$. The only... | ||