Explore >> Select a destination


You are here

research.google
| | swethatanamala.github.io
2.1 parsecs away

Travel
| | In this paper, authors proposed a new language representation model BERT (Bidirectional Encoder Representations from Transformers) which improves fine-tuning based approaches.
| | ai.googleblog.com
2.8 parsecs away

Travel
| | [AI summary] This blog post discusses Google Research's exploration of transfer learning through the T5 model, highlighting its application in natural language processing tasks and the development of the C4 dataset.
| | bdtechtalks.com
3.6 parsecs away

Travel
| | The transformer model has become one of the main highlights of advances in deep learning and deep neural networks.
| | research.google
9.6 parsecs away

Travel
| Posted Keerthana Gopalakrishnan and Kanishka Rao, Google Research, Robotics at Google Major recent advances in multiple subfields of machine learni...