Explore >> Select a destination


You are here

josephm.dev
| | blog.val.town
2.8 parsecs away

Travel
| | How to customize OpenAI to your liking
| | isthisit.nz
1.2 parsecs away

Travel
| | August 2024 Update: Now a solved problem. Use Structured Outputs. Large language models (LLMs) return unstructured output. When we prompt them they respond with one large string. This is fine for applications such as ChatGPT, but in others where we want the LLM to return structured data such as lists or key value pairs, a parseable response is needed. In Building A ChatGPT-enhanced Python REPL I used a technique to prompt the LLM to return output in a text format I could parse.
| | qwenlm.github.io
5.1 parsecs away

Travel
| | GITHUB HUGGING FACE MODELSCOPE DISCORD Today, we're announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder is available in multiple sizes, but we're excited to introduce its most powerful variant first: Qwen3-Coder-480B-A35B-Instruct - a 480B-parameter Mixture-of-Experts model with 35B active parameters which supports the context length of 256K tokens natively and 1M tokens with extrapolation methods, offering exceptional performance in both coding and agentic tasks. Qwen3-Coder-480B-A35B-Instruct sets new state-of-the-art results among open models on Agentic Coding, Agentic Browser-Use, and Agentic Tool-Use, comparable to Claude Sonnet 4.
| | equk.co.uk
26.5 parsecs away

Travel
| writings of a coder + sysadmin