This guide helps you run AI features on your own computer using Ollama, without needing any cloud API keys.
The shortest version:
- Install Ollama from ollama.com/download.
- Download a model:
ollama pull llama3.2 - Open BaoBuildBuddy > Settings > AI Providers.
- Set the endpoint to
http://localhost:11434/v1, leave the model blank for auto-detect, then test and save. - If the test fails, BaoBuildBuddy now returns the exact local failure mode instead of a generic error.
- Ollama runs the AI engine on your computer.
- BaoBuildBuddy talks to Ollama over your local network.
- Your data stays on your machine -- no cloud requests needed.
- "I want the simplest path to get AI working."
- "I don't want to start with API keys."
- "I want to test AI features on my own machine first."
Other starting points:
- Big-picture explanation: ELI5 System Walkthrough
- Full project setup: Starter Guide
Make sure:
- BaoBuildBuddy is already installed or cloned
- You can open the app locally
- You have internet for the initial Ollama install and model download
- You have enough disk space for at least one model
Download and install from the official page:
- Ollama Download
- Ollama Quickstart (vendor's getting-started guide)
macOS / Windows: Opening the Ollama app is usually enough.
Linux:
ollama serveIf Ollama is already running, you don't need to start it again.
For a safe first choice:
ollama pull llama3.2Other models that work well with BaoBuildBuddy:
granite-codemistral
ollama run llama3.2If you see a prompt and can ask a question, Ollama is working. Type /bye or press Ctrl+C to exit.
- Open Settings in BaoBuildBuddy.
- Go to AI Providers.
- Enter
http://localhost:11434/v1as the endpoint (if not already filled in). - Leave the model field blank for auto-detect, or enter
llama3.2to lock to a specific model. - Click the test button.
- Save your settings.
The app defaults already match Ollama's local OpenAI-compatible endpoint, so many users only need to install Ollama and test the connection.
The local provider readiness check now uses the same contract in both Setup and Settings:
- Endpoint reachable
/v1/modelsreturns at least one model- The selected model exists when you enter one manually
- The request does not time out
Expected diagnostic outcomes:
unreachable: the endpoint could not be contactedtimeout: the endpoint did not respond in timeempty-model-list: the server responded, but no models were availableinvalid-model: the configured model was not returned by the serverhealthy: the endpoint is ready and BaoBuildBuddy can route local AI requests to it
After saving your settings, test one of these:
- AI Chat -- ask a simple question
- Resume -- try a draft or review action
- Interview -- generate a practice prompt
If any of those works, your local AI setup is complete.
- Ollama is running
- The endpoint is
http://localhost:11434/v1 - You clicked "test" after saving
If the UI reports unreachable or timeout, verify directly:
curl -fsS http://localhost:11434/v1/modelsUsually the model isn't downloaded yet:
ollama pull llama3.2That's fine. BaoBuildBuddy auto-detects the model from Ollama when the endpoint is set and the server has models available.
Leave the model blank to use auto-detect, or update it to one of the IDs returned by:
curl -fsS http://localhost:11434/v1/models | jq '.data[].id'Normal on the first request. The model needs time to load into memory. Later requests are faster.
| Topic | Guide |
|---|---|
| First-time project setup | Starter Guide |
| Plain-English system overview | ELI5 System Walkthrough |
| Full technical reference | README.md |