Explore >> Select a destination


You are here

deepmind.google
| | blog.rinesi.com
6.5 parsecs away

Travel
| |
| | www.alignmentforum.org
0.9 parsecs away

Travel
| | On March 29th, DeepMind published a paper, "Training Compute-Optimal Large Language Models", that shows that essentially everyone -- OpenAI, DeepMind...
| | blog.moonglow.ai
0.9 parsecs away

Travel
| | Parameters and data. These are the two ingredients of training ML models. The total amount of computation ("compute") you need to do to train a model is proportional to the number of parameters multiplied by the amount of data (measured in "tokens"). Four years ago, it was well-known that if
| | www.jeremymorgan.com
29.3 parsecs away

Travel
| Want to run a large language model like ChatGPT on your Ubuntu machine? Here are the full instructions.