Explore >> Select a destination


You are here

www.swyx.io
| | blog.vstelt.dev
1.2 parsecs away

Travel
| | [AI summary] The article explains the process of building a neural network from scratch in Rust, covering forward and backward propagation, matrix operations, and code implementation.
| | www.paepper.com
1.9 parsecs away

Travel
| | [AI summary] This article explains how to train a simple neural network using Numpy in Python without relying on frameworks like TensorFlow or PyTorch, focusing on the implementation of ReLU activation, weight initialization, and gradient descent for optimization.
| | serengil.wordpress.com
1.8 parsecs away

Travel
| | Sigmoid function (aka logistic function) is moslty picked up as activation functionin neural networks.Because its derivative is easy to demonstrate. It produces output in scale of [0 ,1] whereas input is meaningful between [-5, +5]. Out of this range produces same outputs.In this post, we'll mention the proof of the derivative calculation. Sigmoid function is...
| | sirinath.wordpress.com
19.0 parsecs away

Travel
| Welcome to WordPress.com. After you read this, you should delete and write your own post, with a new title above. Or hit Add New on the left (of the admin dashboard) to start a fresh post. Here are some suggestions for your first post. You can find new ideas for what to blog about by...