Fair Warning

Make sure you are using a capable system — ideally with a powerful CPU, GPU, and adequate cooling — before running large language models locally. LLMs can consume significant resources, generate substantial heat, and may cause system instability or damage if your hardware isn’t up to the task. Please proceed with caution!

This guide walks you through setting up Ollama (with deepseek-r1-goose) and Goose CLI.

Step 1: Download & Install Ollama

curl -fsSL https://ollama.com/install.sh | sh

Or, if you prefer to download manually then checkout:https://ollama.com/download Ollama-SetUpGuide

Step 2:Pull and Run the model

The following model is optimized for the agent we are going to install next so lets pull and run the model:

ollama run michaelneale/deepseek-r1-goose

more info about the model is available here https://www.ollama.com/michaelneale/deepseek-r1-goose

Step 3: Download And configure Goose-cli

Download the goose via HomeBrew if you don’t have it installed please check this article HomeBrew-SetUpGuide

brew install block-goose-cli

now you can run the goose

goose 

Running after the first installation, the configure menu will be shown, make sure you select Ollama as the model provider, you can navigate by up and down arrow and hit return/enter to select the option.

After the model provider, next comes the model selection option, just type michaelneale/deepseek-r1-goose and hit return/enter.

Later if you want to change the model you can always run bash goose configure, i would recommend you use this model others are not working as expected this is already slow.

You can stop goose by given /exit command.

Where Offline Agents Work Best

  • The task is narrow, well-defined, and focused.
  • You want fast, private processing without sending data to the cloud.
  • You are working with small to moderate inputs and outputs, as local models may struggle with large contexts or long conversations on limited hardware.
  • Context limit required is less than 32K tokens

Again, always monitor your system’s health, and don’t hesitate to stop the model if things heat up!