Installing Goose-CLI with Ollama for Local AI Coding
If you're (like me) liking the idea of having an AI coding assistant that runs completely locally on your machine (privacy!), there's a good chance you've stumbled upon Goose-CLI. Combine it with Ollama for running AI models locally, and you've got yourself a pretty powerful setup that doesn't send your code to the cloud.
Let's see what we need to do in order to get both systems up and running, configure them to work together, and set up some convenient shortcuts to switch between coding and writing tasks.