|
You are here |
til.simonwillison.net | ||
| | | | |
blog.pamelafox.org
|
|
| | | | | Recently, as part of my work on Azure OpenAI code samples, I've been experimenting with different ways of streaming data from a server i... | |
| | | | |
isthisit.nz
|
|
| | | | | August 2024 Update: Now a solved problem. Use Structured Outputs. Large language models (LLMs) return unstructured output. When we prompt them they respond with one large string. This is fine for applications such as ChatGPT, but in others where we want the LLM to return structured data such as lists or key value pairs, a parseable response is needed. In Building A ChatGPT-enhanced Python REPL I used a technique to prompt the LLM to return output in a text format I could parse. | |
| | | | |
blog.miguelgrinberg.com
|
|
| | | | | miguelgrinberg.com | |
| | | | |
glama.ai
|
|
| | | A server that allows interaction with Google's Gemini AI through the Gemini CLI tool using the Model Context Protocol, providing a standardized interface for querying Gemini with various options and configurations. | ||