You are here |
fak3r.com | ||
| | | |
manuel.kiessling.net
|
|
| | | | Most KVM or virt-install tutorials will make you think that while you can create new virtual guests on the text console, you still have to log into them using VNC in order to actually use their OS installation tools. But in fact there is a way to completely install new guests without leaving your SSH session - as long as the guest OS does have a text-based installer, that it. I have tested this with an Ubuntu 12.04 LTS Server guest. | |
| | | |
nodogmablog.bryanhogan.net
|
|
| | | | I have been using Visual Studio Code in Ubuntu Linux for a while, but it was not an easy process to get it working the first time. I just got around to installing . | |
| | | |
newcome.wordpress.com
|
|
| | | | I wanted to spin up a Linux development environment to hack on some code that needed epoll. I could have run everything in a Docker container, but I kinda wanted to be in that environment for total hackage. I thought maybe I'd just do it in a Virtualbox instance. Then I didn't want to install... | |
| | | |
ezyang.github.io
|
|
| | When you're learning to use a new framework or library, simple uses of the software can be done just by copy pasting code from tutorials and tweaking them as necessary. But at some point, it's a good idea to just slog through reading the docs from top-to-bottom, to get a full understanding of what is and is not possible in the software. One of the big wins of AI coding is that LLMs know so many things from their pretraining. For extremely popular frameworks that occur prominently in the pretraining set, an LLM is likely to have memorized most aspects of how to use the framework. But for things that are not so common or beyond the knowledge cutoff, you will likely get a model that hallucinates things. Ideally, an agentic model would know to do a web search and find the docs it needs. However, Sonnet does not currently support web search, so you have to manually feed it documentation pages as needed. Fortunately, Cursor makes this very convenient: simply dropping a URL inside a chat message will include its contents for the LLM. |