|
You are here |
www.datagenetics.com | ||
| | | | |
www.jeremykun.com
|
|
| | | | | Machine learning is broadly split into two camps, statistical learning and non-statistical learning. The latter we've started to get a good picture of on this blog; we approached Perceptrons, decision trees, and neural networks from a non-statistical perspective. And generally "statistical" learning is just that, a perspective. Data is phrased in terms of independent and dependent variables, and statistical techniques are leveraged against the data. In this post we'll focus on the simplest example of thi... | |
| | | | |
www.code-spot.co.za
|
|
| | | | | I am playing around with generating textures and decided to post some preliminary results. The algorithm used to create these images is simple to implement, but slow. Here is how it works: 1. Gener... | |
| | | | |
aosmith.rbind.io
|
|
| | | | | I walk through an example of simulating data from a binomial generalized linear mixed model with a logit link and then exploring estimates of over/underdispersion. | |
| | | | |
nhigham.com
|
|
| | | A Householder matrix is an $latex n\times n$ orthogonal matrix of the form $latex \notag P = I - \displaystyle\frac{2}{v^Tv} vv^T, \qquad 0 \ne v \in\mathbb{R}^n. $ It is easily verified that $LATEX P$ is orthogonal ($LATEX P^TP = I$), symmetric ($LATEX P^T = P$), involutory ($LATEX P^2 = I$ that is, $LATEX P$ is... | ||