Explore >> Select a destination


You are here

galowicz.de
| | nhigham.com
7.8 parsecs away

Travel
| | The trace of an $latex n\times n$ matrix is the sum of its diagonal elements: $latex \mathrm{trace}(A) = \sum_{i=1}^n a_{ii}$. The trace is linear, that is, $latex \mathrm{trace}(A+B) = \mathrm{trace}(A) + \mathrm{trace}(B)$, and $latex \mathrm{trace}(A) = \mathrm{trace}(A^T)$. A key fact is that the trace is also the sum of the eigenvalues. The proof is by...
| | www.jeremykun.com
9.3 parsecs away

Travel
| | For fixed integers $ r > 0$, and odd $ g$, a Moore graph is an $ r$-regular graph of girth $ g$ which has the minimum number of vertices $ n$ among all such graphs with the same regularity and girth. (Recall, A the girth of a graph is the length of its shortest cycle, and it's regular if all its vertices have the same degree) Problem (Hoffman-Singleton): Find a useful constraint on the relationship between $ n$ and $ r$ for Moore graphs of girth $ 5$ and degree $ r$.
| | marcospereira.me
10.6 parsecs away

Travel
| | In this post we summarize the math behind deep learning and implement a simple network that achieves 85% accuracy classifying digits from the MNIST dataset.
| | www.paepper.com
51.8 parsecs away

Travel
| When you have a big data set and a complicated machine learning problem, chances are that training your model takes a couple of days even on a modern GPU. However, it is well-known that the cycle of having a new idea, implementing it and then verifying it should be as quick as possible. This is to ensure that you can efficiently test out new ideas. If you need to wait for a whole week for your training run, this becomes very inefficient.