|
You are here |
harvardnlp.github.io | ||
| | | | |
nlp.seas.harvard.edu
|
|
| | | | | The Annotated Transformer | |
| | | | |
sigmoidprime.com
|
|
| | | | | An exploration of Transformer-XL, a modified Transformer optimized for longer context length. | |
| | | | |
blog.eleuther.ai
|
|
| | | | | Rotary Positional Embedding (RoPE) is a new type of position encoding that unifies absolute and relative approaches. We put it to the test. | |
| | | | |
igorstechnoclub.com
|
|
| | | This week I learned something that finally made "transfer learning" click. I had always heard that you can hit strong accuracy fast by reusing a pretrain... | ||