Explore >> Select a destination


You are here

www.alignmentforum.org
| | jeffreyladish.com
2.8 parsecs away

Travel
| | In this post I explain what we can and can't learn about AI sentience from conversations with Large Language Models.
| | blog.moonglow.ai
1.9 parsecs away

Travel
| | Parameters and data. These are the two ingredients of training ML models. The total amount of computation ("compute") you need to do to train a model is proportional to the number of parameters multiplied by the amount of data (measured in "tokens"). Four years ago, it was well-known that if
| | deepmind.google
0.9 parsecs away

Travel
| | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers...
| | rakuforprediction.wordpress.com
8.0 parsecs away

Travel
| This document shows how to doFunction Callingworkflows with Large Language Models (LLMs) of OpenAI.