|
You are here |
hadrienj.github.io | ||
| | | | |
stephenmalina.com
|
|
| | | | | Matrix Potpourri # As part of reviewing Linear Algebra for my Machine Learning class, I've noticed there's a bunch of matrix terminology that I didn't encounter during my proof-based self-study of LA from Linear Algebra Done Right. This post is mostly intended to consolidate my own understanding and to act as a reference to future me, but if it also helps others in a similar position, that's even better! | |
| | | | |
jrhawley.ca
|
|
| | | | | When collecting data from scientific experiments, it's often useful to compare individual samples against each other to see how similar they are. One way to do this is using the... | |
| | | | |
quomodocumque.wordpress.com
|
|
| | | | | I met Mike Freedman last week at CMSA and I learned a great metaphor about an old favorite subject of mine, random walks on groups. The Heisenberg group is the group of upper triangular matrices with 1's on the diagonal: You can take a walk on the integral or Z/pZ points of the Heisenberg group... | |
| | | | |
nhigham.com
|
|
| | | For a polynomial $latex \notag \phi(t) = a_kt^k + \cdots + a_1t + a_0, $ where $latex a_k\in\mathbb{C}$ for all $latex k$, the matrix polynomial obtained by evaluating $latex \phi$ at $latex A\in\mathbb{C}^{n \times n}$ is $latex \notag \phi(A) = a_kA^k + \cdots + a_1A + a_0 I. $ (Note that the constant term is... | ||