|
You are here |
transformer-circuits.pub | ||
| | | | |
www.lesswrong.com
|
|
| | | | | Text of post based on our blog post as a linkpost for the full paper which is considerably longer and more detailed. ... | |
| | | | |
thesephist.com
|
|
| | | | | [AI summary] The text provides an in-depth overview of research on sparse autoencoders (SAEs) applied to embeddings for automated interpretability. It discusses methods for analyzing and manipulating embeddings, including feature extraction, gradient-based optimization, and visualization tools. The work emphasizes the importance of understanding model representations to improve human-computer interaction with information systems. Key components include: 1) Automated interpretability prompts for generating feature labels, 2) Feature gradients implementation for optimizing embeddings to match desired feature dictionaries, and 3) Visualizations of feature spaces and embedding transformations. The text also includes FAQs addressing the use of embeddings over lan... | |
| | | | |
iclr-blogposts.github.io
|
|
| | | | | Identifying, Interpreting & Ablating the Sources of a Deep Learning Puzzle | |
| | | | |
marcospereira.me
|
|
| | | In this post we summarize the math behind deep learning and implement a simple network that achieves 85% accuracy classifying digits from the MNIST dataset. | ||