Explore >> Select a destination


You are here

favtutor.com
| | deepmind.google
15.9 parsecs away

Travel
| | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers...
| | analyticsindiamag.com
18.0 parsecs away

Travel
| | Kan-LLaMA is a 7 billion Llama 2 model which is LoRA pre-trained and fine-tuned on "Kannada" token, built by researchers from VIT.
| | www.marktechpost.com
15.2 parsecs away

Travel
| | Microsoft AI Releases Phi-3 Family of Models: A 3.8B Parameter Language Model Trained on 3.3T Tokens Locally on Your Phone
| | blog.research.google
80.9 parsecs away

Travel
|