Explore >> Select a destination


You are here

speakerdeck.com
| | golb.hplar.ch
1.6 parsecs away

Travel
| | [AI summary] The article describes the implementation of a neural network in Java and JavaScript for digit recognition using the MNIST dataset, covering forward and backpropagation processes.
| | coornail.net
2.8 parsecs away

Travel
| | Neural networks are a powerful tool in machine learning that can be trained to perform a wide range of tasks, from image classification to natural language processing. In this blog post, well explore how to teach a neural network to add together two numbers. You can also think about this article as a tutorial for tensorflow.
| | sirupsen.com
2.5 parsecs away

Travel
| | [AI summary] The article provides an in-depth explanation of how to build a neural network from scratch, focusing on the implementation of a simple average function and the introduction of activation functions for non-linear tasks. It discusses the use of matrix operations, the importance of GPUs for acceleration, and the role of activation functions like ReLU. The author also outlines next steps for further exploration, such as expanding the model, adding layers, and training on datasets like MNIST.
| | blog.keras.io
10.7 parsecs away

Travel
| [AI summary] The text discusses various types of autoencoders and their applications. It starts with basic autoencoders, then moves to sparse autoencoders, deep autoencoders, and sequence-to-sequence autoencoders. The text also covers variational autoencoders (VAEs), explaining their structure and training process. It includes code examples for each type of autoencoder and mentions the use of tools like TensorBoard for visualization. The VAE section highlights how to generate new data samples and visualize the latent space. The text concludes with references and a note about the potential for further topics.