|
You are here |
blog.moonglow.ai | ||
| | | | |
lunary.ai
|
|
| | | | | Use this tool below to understand how a piece of text might be tokenized by Anthropic models (Claude 1, Claude 2, Claude Instant, Claude 3) and the total count of tokens in that piece of text. | |
| | | | |
www.alignmentforum.org
|
|
| | | | | On March 29th, DeepMind published a paper, "Training Compute-Optimal Large Language Models", that shows that essentially everyone -- OpenAI, DeepMind... | |
| | | | |
deepmind.google
|
|
| | | | | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers... | |
| | | | |
blog.pamelafox.org
|
|
| | | Today I went on a quest to figure out the best way to use SLMs (small language models) like Phi-3 in a GitHub Codespace, so that I can... | ||