Explore >> Select a destination


You are here

sigmoidprime.com
| | bdtechtalks.com
7.5 parsecs away

Travel
| | The transformer model has become one of the main highlights of advances in deep learning and deep neural networks.
| | nlp.seas.harvard.edu
10.6 parsecs away

Travel
| |
| | blog.lambdaclass.com
12.9 parsecs away

Travel
| | TL;DR: this post addresses the paper introducing rStar-Math and the techniques for smaller language models to outperform more complex large language models on math-related tasks. You can check the code here. rStar-Math significantly improved the math reasoning abilities of SLMs. For instance, on the MATH benchmark, it enhanced Qwen2.
| | colah.github.io
46.8 parsecs away

Travel
|