diff --git a/README.md b/README.md
index 8dbfa9beb..53f0b2619 100644
--- a/README.md
+++ b/README.md
@@ -233,7 +233,7 @@ export STRIX_REASONING_EFFORT="high" # control thinking effort (default: high,
- [Anthropic Claude Sonnet 4.6](https://claude.com/platform/api) — `anthropic/claude-sonnet-4-6`
- [Google Gemini 3 Pro Preview](https://cloud.google.com/vertex-ai) — `vertex_ai/gemini-3-pro-preview`
-See the [LLM Providers documentation](https://docs.strix.ai/llm-providers/overview) for all supported providers including Vertex AI, Bedrock, Azure, and local models.
+See the [LLM Providers documentation](https://docs.strix.ai/llm-providers/overview) for all supported providers including Novita AI, Vertex AI, Bedrock, Azure, and local models.
## Enterprise
diff --git a/docs/docs.json b/docs/docs.json
index 27ee5dc9c..df83365e1 100644
--- a/docs/docs.json
+++ b/docs/docs.json
@@ -36,6 +36,7 @@
"llm-providers/openai",
"llm-providers/anthropic",
"llm-providers/openrouter",
+ "llm-providers/novita",
"llm-providers/vertex",
"llm-providers/bedrock",
"llm-providers/azure",
diff --git a/docs/llm-providers/novita.mdx b/docs/llm-providers/novita.mdx
new file mode 100644
index 000000000..632391d81
--- /dev/null
+++ b/docs/llm-providers/novita.mdx
@@ -0,0 +1,35 @@
+---
+title: "Novita AI"
+description: "Configure Strix with Novita AI models"
+---
+
+[Novita AI](https://novita.ai) provides fast, cost-efficient inference for open-source models via an OpenAI-compatible API.
+
+## Setup
+
+```bash
+export STRIX_LLM="openai/moonshotai/kimi-k2.5"
+export LLM_API_KEY="your-novita-api-key"
+export LLM_API_BASE="https://api.novita.ai/openai"
+```
+
+## Available Models
+
+| Model | Configuration |
+|-------|---------------|
+| Kimi K2.5 (recommended) | `openai/moonshotai/kimi-k2.5` |
+| GLM-5 | `openai/zai-org/glm-5` |
+| MiniMax M2.5 | `openai/minimax/minimax-m2.5` |
+
+## Get API Key
+
+1. Sign up at [novita.ai](https://novita.ai)
+2. Navigate to **API Keys** in your dashboard
+3. Create a new key and copy it
+
+## Benefits
+
+- **Cost-efficient** — Competitive pricing with per-token billing
+- **OpenAI-compatible** — Drop-in replacement using `LLM_API_BASE`
+- **Large context** — Models support up to 262k token context windows
+- **Function calling** — All listed models support tool/function calling
diff --git a/docs/llm-providers/overview.mdx b/docs/llm-providers/overview.mdx
index 153ad0ca8..3b986ff4c 100644
--- a/docs/llm-providers/overview.mdx
+++ b/docs/llm-providers/overview.mdx
@@ -25,12 +25,17 @@ You can also use any LiteLLM-compatible provider with your own API keys:
| GPT-5 | OpenAI | `openai/gpt-5` |
| Claude Sonnet 4.6 | Anthropic | `anthropic/claude-sonnet-4-6` |
| Gemini 3 Pro | Google Vertex | `vertex_ai/gemini-3-pro-preview` |
+| Kimi K2.5 | Novita AI | `openai/moonshotai/kimi-k2.5` |
```bash
export STRIX_LLM="openai/gpt-5"
export LLM_API_KEY="your-api-key"
```
+
+ **Novita AI** requires an additional `LLM_API_BASE` variable pointing to their OpenAI-compatible endpoint. See the [Novita AI guide](/llm-providers/novita) for full setup instructions.
+
+
## Local Models
Run models locally with [Ollama](https://ollama.com), [LM Studio](https://lmstudio.ai), or any OpenAI-compatible server:
@@ -57,6 +62,9 @@ See the [Local Models guide](/llm-providers/local) for setup instructions and re
Access 100+ models through a single API.
+
+ Cost-efficient open-source model inference.
+
Gemini 3 models via Google Cloud.