-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathcluster.toml.example
More file actions
56 lines (50 loc) · 1.42 KB
/
cluster.toml.example
File metadata and controls
56 lines (50 loc) · 1.42 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
[cluster]
name = "my-cluster"
domain = "example.com"
acme_email = "ops@example.com"
log_level = "info"
# Single-node setup — just omit the [[node]] sections and orca runs everything locally.
# Multi-node — list your nodes below. The first node to run `orca server` becomes the leader.
# [[node]]
# address = "10.0.0.1"
# labels = { zone = "eu-1", role = "general" }
#
# [[node]]
# address = "10.0.0.2"
# labels = { zone = "eu-1", role = "general" }
# GPU node — declare GPUs so the scheduler can place GPU workloads
# [[node]]
# address = "10.0.0.3"
# labels = { zone = "eu-1", role = "gpu" }
#
# [[node.gpus]]
# vendor = "nvidia"
# count = 2
# model = "A100"
# -- AI Operations Assistant --
# Enables: `orca ask`, conversational alerts, smart import analysis, log summarization.
# Works with any OpenAI-compatible API (LiteLLM, Ollama, vLLM, OpenAI, etc.)
# [ai]
# provider = "litellm"
# endpoint = "https://llm.example.com"
# model = "qwen3-30b"
# api_key = "${secrets.ai_api_key}"
#
# [ai.alerts]
# enabled = true
# analysis_interval_secs = 60
#
# [ai.alerts.channels]
# slack = "https://hooks.slack.com/services/..."
# webhook = "https://my-pagerduty-webhook/..."
#
# [ai.auto_remediate]
# restart_crashed = true
# scale_on_pressure = false
# rollback_on_failure = false
# [observability]
# otlp_endpoint = "https://signoz.example.com"
#
# [observability.alerts]
# webhook = "https://hooks.slack.com/..."
# email = "ops@example.com"