Skip to content

Commit b9991ea

Browse files
committed
run model in container
1 parent 03661ad commit b9991ea

3 files changed

Lines changed: 44 additions & 0 deletions

File tree

README.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -15,3 +15,4 @@ Build the native agent using the following command.
1515
mvn -Pnative -DskipTests package
1616
```
1717

18+
Start the application and access the chat UI at `http://localhost:8080/webjars/chat-agent-ui/index.html`.

compose.yaml

Lines changed: 35 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,35 @@
1+
services:
2+
model-runner:
3+
image: ghcr.io/ggml-org/llama.cpp:server
4+
volumes:
5+
- model-files:/models
6+
command:
7+
- "--host"
8+
- "0.0.0.0"
9+
- "--port"
10+
- "8080"
11+
- "-n"
12+
- "512"
13+
- "-m"
14+
- "/models/Qwen3-0.6B-Q8_0.gguf"
15+
ports:
16+
- "8180:8080"
17+
depends_on:
18+
model-downloader:
19+
condition: service_completed_successfully
20+
21+
model-downloader:
22+
image: ghcr.io/alexcheng1982/model-downloader
23+
restart: "no"
24+
volumes:
25+
- model-files:/models
26+
command:
27+
- "hf"
28+
- "download"
29+
- "unsloth/Qwen3-0.6B-GGUF"
30+
- "Qwen3-0.6B-Q8_0.gguf"
31+
- "--local-dir"
32+
- "/models"
33+
34+
volumes:
35+
model-files:
Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,8 @@
1+
spring:
2+
ai:
3+
openai:
4+
baseUrl: http://localhost:8180
5+
api-key: local
6+
chat:
7+
options:
8+
temperature: 0.0

0 commit comments

Comments
 (0)