Use this guide to run Personal OS with Pi as the agent frontend.
If you want fully local/offline execution, run a local model server first.
brew install llama.cpp
llama-server \
-hf unsloth/GLM-4.7-Flash-GGUF:UD-Q4_K_XL \
--jinja \
--temp 0.7 --top-p 1.0 --min-p 0.01 \
--repeat-penalty 1.0 --fit on \
--port 8080Quantization guidance:
- 16 GB RAM:
UD-Q3_K_XL - 24 GB RAM:
UD-Q4_K_XL - 36 GB+ RAM:
UD-Q6_K
Fallback model for smaller machines:
llama-server -hf bartowski/Qwen2.5-Coder-7B-Instruct-GGUF:Q4_K_M --fit on --jinja --port 8080npm install -g @mariozechner/pi-coding-agentCreate ~/.pi/agent/models.json:
{
"providers": {
"llama-cpp": {
"baseUrl": "http://localhost:8080/v1",
"api": "openai-completions",
"apiKey": "none",
"models": [{ "id": "GLM-4.7-Flash" }]
}
}
}Run pi, select your local model, then continue with repo setup below.
- This workspace as working directory
- Access to
AGENTS.md(shared behavior) - Access to canonical skills in
.agents/skills/
- Open Pi in:
<repo-root>- Ensure Pi uses this repo's canonical skills path:
<repo-root>/.agents/skills- Start with:
Read AGENTS.md and run the daily standup workflow
If you run Pi with a local model backend, this repo still works unchanged. Skills and workflows are model-agnostic.
PI.mdandAGENTS.mdare wrappers/instructions only.- The source of truth for skills remains
.agents/skills/*/SKILL.md.