feat: add Avian as a cloud LLM inference provider#8666
feat: add Avian as a cloud LLM inference provider#8666avianion wants to merge 1 commit intomudler:masterfrom
Conversation
✅ Deploy Preview for localai ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
c76099f to
c64030a
Compare
|
Hey @mudler, would love your review on this when you get a chance. Happy to address any feedback! |
|
Friendly follow-up — this PR is still active and ready for review. Would appreciate a look when you get a chance! cc @mudler |
|
Friendly follow-up — this PR is still active and ready for review. All feedback has been addressed. Would appreciate a look when you get a chance! cc @mudler |
|
Hey @mudler — friendly follow-up on this PR. Avian is an OpenAI-compatible inference provider that's already live and powering apps like ISEKAI ZERO. This is a lightweight integration (standard OpenAI-compatible endpoint) and we're happy to address any feedback or make adjustments. Would love to get this merged if you have a moment to review. Thanks! |
Add Avian (https://avian.io) as a Go backend that proxies requests to the Avian OpenAI-compatible API at https://api.avian.io/v1. Backend implementation: - Go gRPC backend at backend/go/avian/ following the huggingface backend pattern - Supports chat completions with structured messages and streaming (SSE) - Authentication via AVIAN_API_KEY environment variable - Configurable base URL via AVIAN_API_BASE environment variable Gallery models: - deepseek/deepseek-v3.2: 164K context, $0.26/$0.38 per 1M tokens - moonshotai/kimi-k2.5: 131K context, $0.45/$2.20 per 1M tokens - z-ai/glm-5: 131K context, $0.30/$2.55 per 1M tokens - minimax/minimax-m2.5: 1M context, $0.30/$1.10 per 1M tokens Build infrastructure: - Backend definition in Makefile (golang backend) - CI workflow entries for Linux (amd64/arm64) and macOS (metal) - Backend index.yaml entries with OCI image references Signed-off-by: Kyle D <deximia@hotmail.com>
c64030a to
023abe2
Compare
Summary
Adds Avian as a new Go backend for LocalAI, enabling users to access cloud-hosted LLMs through Avian's OpenAI-compatible API at
https://api.avian.io/v1.Backend
backend/go/avian/following the existinghuggingfacebackend patternAVIAN_API_KEYenvironment variableAVIAN_API_BASEenvironment variable (defaults tohttps://api.avian.io/v1)Gallery Models
Four models available out of the box:
deepseek/deepseek-v3.2moonshotai/kimi-k2.5z-ai/glm-5minimax/minimax-m2.5Usage
Build Infrastructure
BACKEND_AVIAN = avian|golang|.|false|true)index.yamlentries with OCI image references.NOTPARALLELanddocker-build-backendstargetsFiles Changed
backend/go/avian/- Go gRPC backend implementation (main.go, avian.go, Makefile, run.sh, package.sh)gallery/avian.yaml- Base model configuration templategallery/index.yaml- Four Avian model entriesbackend/index.yaml- Backend metadata and OCI image entriesMakefile- Backend build targets.github/workflows/backend.yml- CI build matrix entriesTest plan
make -C backend/go/avianBACKEND=avian make docker-build-avianAVIAN_API_KEYlocal-ai models install avian-deepseek-v3.2cc @mudler