|
You are here |
datadan.io | ||
| | | | |
mccormickml.com
|
|
| | | | | This tutorial covers the skip gram neural network architecture for Word2Vec. My intention with this tutorial was to skip over the usual introductory and abst... | |
| | | | |
golb.hplar.ch
|
|
| | | | | [AI summary] The article describes the implementation of a neural network in Java and JavaScript for digit recognition using the MNIST dataset, covering forward and backpropagation processes. | |
| | | | |
sirupsen.com
|
|
| | | | | [AI summary] The article provides an in-depth explanation of how to build a neural network from scratch, focusing on the implementation of a simple average function and the introduction of activation functions for non-linear tasks. It discusses the use of matrix operations, the importance of GPUs for acceleration, and the role of activation functions like ReLU. The author also outlines next steps for further exploration, such as expanding the model, adding layers, and training on datasets like MNIST. | |
| | | | |
sausheong.github.io
|
|
| | | I have written a lot of computer programs in my career, most of the time to solve various problems or perform some tasks (or sometimes just for fun). For most part, other than bugs, as long as I tell the computer what to do very clearly (in whichever the programming language I use) it will obediently follow my instructions. This is because computer programs are really good at executing algorithms - instructions that follow defined steps and patterns that are precise and often repetitious. | ||