Explore >> Select a destination


You are here

blog.moonglow.ai
| | www.alignmentforum.org
7.7 parsecs away

Travel
| | On March 29th, DeepMind published a paper, "Training Compute-Optimal Large Language Models", that shows that essentially everyone -- OpenAI, DeepMind...
| | deepmind.google
5.9 parsecs away

Travel
| | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers...
| | jack-clark.net
14.8 parsecs away

Travel
| | China releasesanotherreally good open access language model:...Baichuan2 shows us just how broadly distributed LLM training and deployment is getting...Chinese startup Baichuan has trained and release two open access language models. These models are interesting for two reasons, a) they're broadly disseminated and available for free, and b) they are 'language models with Chinese characteristics' -...
| | jan.schnasse.org
23.0 parsecs away

Travel
|