|
You are here |
rjlipton.com | ||
| | | | |
667-per-cm.net
|
|
| | | | | This post could also be subtitled "Residual deviance isn't the whole story." My favorite book on logistic regression is by Dr Joseph Hilbe, Logistic Regression Models, CRC Press, 2009, Chapman & Hill. It is a solidly frequentist text, but its discussion of models and rich examples make that besides the point. Except in one case.... | |
| | | | |
nhigham.com
|
|
| | | | | A $latex p$th root of an $latex n\times n$ matrix $LATEX A$ is a matrix $LATEX X$ such that $latex X^p = A$, and it can be written $latex X = A^{1/p}$. For a rational number $latex r = j/k$ (where $latex j$ and $latex k$ are integers), defining $latex A^r$ is more difficult: is... | |
| | | | |
nickhar.wordpress.com
|
|
| | | | | 1. Low-rank approximation of matrices Let $latex {A}&fg=000000$ be an arbitrary $latex {n \times m}&fg=000000$ matrix. We assume $latex {n \leq m}&fg=000000$. We consider the problem of approximating $latex {A}&fg=000000$ by a low-rank matrix. For example, we could seek to find a rank $latex {s}&fg=000000$ matrix $latex {B}&fg=000000$ minimizing $latex { \lVert A - B... | |
| | | | |
dsaber.com
|
|
| | | Warning: This is a love story between a man and his Python module As I mentioned previously, one of the most powerful concepts I've really learned at Zipfian has been Bayesian inference using PyMC. PyMC is currently my favorite library of any kind in any language. I dramatically italicized "learned" because I had been taught... | ||