You are here |
blog.moonglow.ai | ||
| | | |
www.alignmentforum.org
|
|
| | | | On March 29th, DeepMind published a paper, "Training Compute-Optimal Large Language Models", that shows that essentially everyone -- OpenAI, DeepMind... | |
| | | |
deepmind.google
|
|
| | | | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers... | |
| | | |
jack-clark.net
|
|
| | | | China releasesanotherreally good open access language model:...Baichuan2 shows us just how broadly distributed LLM training and deployment is getting...Chinese startup Baichuan has trained and release two open access language models. These models are interesting for two reasons, a) they're broadly disseminated and available for free, and b) they are 'language models with Chinese characteristics' -... | |
| | | |
jan.schnasse.org
|
|
| |