This guide shows how to install and run Ollama for Audify local script generation.
Recommended starter model:
qwen3:1.7b
- At least 8 GB RAM (16 GB recommended for smoother local inference).
- ~4 GB free disk space for runtime and model files.
- Docker Desktop installed if you run Audify via Docker Compose.
Option A (Installer):
- Download from https://ollama.com/download
- Open installer and complete setup.
Option B (Homebrew):
brew install ollamaollama serveKeep this terminal running, or run Ollama as a background app from the installer.
ollama pull qwen3:1.7bollama list
curl http://localhost:11434/api/tags- Download the Windows installer from https://ollama.com/download
- Run installer and complete setup.
Ollama usually starts as a background service after installation.
If needed, start from Start Menu: Ollama.
ollama pull qwen3:1.7bollama list
curl http://localhost:11434/api/tagsIf curl aliases to Invoke-WebRequest, this also works:
Invoke-RestMethod http://localhost:11434/api/tagscurl -fsSL https://ollama.com/install.sh | shollama serveFor systemd environments, you can run as a service if configured by your setup:
sudo systemctl enable ollama
sudo systemctl start ollama
sudo systemctl status ollamaollama pull qwen3:1.7bollama list
curl http://localhost:11434/api/tagsUpdate api/llm-service/.env:
OLLAMA_BASE_URL=http://host.docker.internal:11434/v1
OLLAMA_MODEL=qwen3:1.7bNotes:
host.docker.internalworks for Docker Desktop on macOS/Windows.- On Linux, you may need to expose host networking differently or set
OLLAMA_BASE_URLto your host IP.
Run a local generation test:
curl http://localhost:11434/api/generate -d '{
"model": "qwen3:1.7b",
"prompt": "Write one sentence about document to podcast conversion.",
"stream": false
}'If you receive a JSON response with generated text, Ollama is ready.
- Reopen terminal after install.
- Confirm binary path is in
PATH.
- Check service is running:
ollama serve. - Check firewall settings.
- First pull downloads full model weights.
- Verify network connectivity and disk space.
- Confirm
OLLAMA_BASE_URLis correct for your OS. - Restart stack:
docker compose down && docker compose up --build -d.
qwen3:1.7bfor fast local testing.llama3for broader language quality.mistralas another lightweight option.
Choose based on latency, quality, and available hardware.