You are here |
vladfeinberg.com | ||
| | | |
www.depthfirstlearning.com
|
|
| | | | ||
| | | |
fanpu.io
|
|
| | | | An aperiodic and irreducible Markov chain will eventually converge to a stationary distribution. This is used in many applications in machine learning like Markov Chain Monte Carlo (MCMC) methods, where random walks on Markov chains are used to obtain a good estimate of the log likelihood of the partition function of a model, which is hard to compute directly as it is #P-hard (this is even harder than NP-hard). However, one common problem is that it is unclear how many steps we should take before we are guaranteed that the Markov chain has converged to the its stationary distribution. In this post, we understand how the spectral gap of the transition matrix of the Markov chain relates to its mixing time. | |
| | | |
tiao.io
|
|
| | | | One weird trick to make exact inference in Bayesian logistic regression tractable. | |
| | | |
pyimagesearch.com
|
|
| | In this tutorial, you will learn what gradient descent is, how gradient descent enables us to train neural networks, variations of gradient descent, including Stochastic Gradient Descent (SGD), and how SGD can be improved using momentum and Nesterov acceleration. |