You are here |
sigmoidprime.com | ||
| | | |
bdtechtalks.com
|
|
| | | | The transformer model has become one of the main highlights of advances in deep learning and deep neural networks. | |
| | | |
nlp.seas.harvard.edu
|
|
| | | | ||
| | | |
blog.lambdaclass.com
|
|
| | | | TL;DR: this post addresses the paper introducing rStar-Math and the techniques for smaller language models to outperform more complex large language models on math-related tasks. You can check the code here. rStar-Math significantly improved the math reasoning abilities of SLMs. For instance, on the MATH benchmark, it enhanced Qwen2. | |
| | | |
colah.github.io
|
|
| |