Explore >> Select a destination


You are here

blog.moonglow.ai
| | lunary.ai
2.9 parsecs away

Travel
| | Use this tool below to understand how a piece of text might be tokenized by Anthropic models (Claude 1, Claude 2, Claude Instant, Claude 3) and the total count of tokens in that piece of text.
| | www.alignmentforum.org
1.9 parsecs away

Travel
| | On March 29th, DeepMind published a paper, "Training Compute-Optimal Large Language Models", that shows that essentially everyone -- OpenAI, DeepMind...
| | deepmind.google
0.4 parsecs away

Travel
| | We ask the question: "What is the optimal model size and number of training tokens for a given compute budget?" To answer this question, we train models of various sizes and with various numbers...
| | blog.pamelafox.org
24.7 parsecs away

Travel
| Today I went on a quest to figure out the best way to use SLMs (small language models) like Phi-3 in a GitHub Codespace, so that I can...