Explore >> Select a destination


You are here

harvardnlp.github.io
| | teddykoker.com
2.3 parsecs away

Travel
| | This post is the first in a series of articles about natural language processing (NLP), a subfield of machine learning concerning the interaction between computers and human language. This article will be focused on attention, a mechanism that forms the backbone of many state-of-the art language models, including Googles BERT (Devlin et al., 2018), and OpenAIs GPT-2 (Radford et al., 2019).
| | sigmoidprime.com
3.2 parsecs away

Travel
| | An exploration of Transformer-XL, a modified Transformer optimized for longer context length.
| | swethatanamala.github.io
1.9 parsecs away

Travel
| | In this paper, authors proposed a new simple network architecture, the Transformer, based solely on attention mechanisms, removing convolutions and recurrences entirely. Transformer is the first transduction model relying entirely...
| | www.khanna.law
11.2 parsecs away

Travel
| You want to train a deep neural network. You have the data. It's labeled and wrangled into a useful format. What do you do now?