@@ -20,13 +20,13 @@ Solo Server is a lightweight platform that enables users to manage and monitor A
2020## Features
2121
2222- ** Seamless Setup:** Manage your on device AI with a simple CLI and HTTP servers
23- - ** Open Model Registry:** Pull models from registries like Hugging Face and Ollama
23+ - ** Open Model Registry:** Pull models from registries like Ollama & Hugging Face
2424- ** Lean Load Testing:** Built-in commands to benchmark endpoints
2525- ** Cross-Platform Compatibility:** Deploy AI models effortlessly on your hardware
2626- ** Configurable Framework:** Auto-detect hardware (CPU, GPU, RAM) and sets configs
2727
2828## Supported Models
29- Solo Server supports ** multiple model sources** , including ** Ollama, Hugging Face, and Ramalama ** .
29+ Solo Server supports ** multiple model sources** , including ** Ollama & Hugging Face** .
3030
3131| ** Model Name** | ** Source** |
3232| ------------------------| ----------------------------------------------------------|
@@ -39,7 +39,7 @@ Solo Server supports **multiple model sources**, including **Ollama, Hugging Fac
3939| ** Mistral 7B v3** | ` hf://MaziyarPanahi/Mistral-7B-Instruct-v0.3-GGUF ` |
4040| ** Hermes 2 Pro** | ` hf://NousResearch/Hermes-2-Pro-Mistral-7B-GGUF ` |
4141| ** Cerebrum 1.0 7B** | ` hf://froggeric/Cerebrum-1.0-7b-GGUF ` |
42- | ** Dragon Mistral 7B** | ` hf://llmware/dragon-mistral-7b-v0 `
42+ | ** Dragon Mistral 7B** | ` hf://llmware/dragon-mistral-7b-v0 ` |
4343
4444## Table of Contents
4545
@@ -52,6 +52,12 @@ Solo Server supports **multiple model sources**, including **Ollama, Hugging Fac
5252
5353## Installation
5454
55+ ### ** 🔹Prerequisites**
56+
57+ - ** 🐋 Docker:** Required for containerization
58+ - [ Install Docker] ( https://docs.docker.com/get-docker/ )
59+ - Ensure Docker daemon is running
60+
5561### ** 🔹 Install via PyPI**
5662``` sh
5763pip install solo-server
@@ -65,22 +71,39 @@ Creates an isolated environment using `uv` for performance and stability.
6571
6672Run the ** interactive setup** to configure Solo Server:
6773``` sh
68- solo setup
74+ solo start
6975```
7076### ** 🔹 Setup Features**
7177✔️ ** Detects CPU, GPU, RAM** for ** hardware-optimized execution**
7278✔️ ** Auto-configures ` solo.conf ` with optimal settings**
73- ✔️ ** Requests API keys for Ngrok and Replicatea **
79+ ✔️ ** Requests API keys for Ngrok and Replicate **
7480✔️ ** Recommends the compute backend OCI (CUDA, HIP, SYCL, Vulkan, CPU, Metal)**
7581
7682---
7783
84+ ** Example Output:**
85+ ``` sh
86+ 🖥️ System Information
87+ Operating System: Windows
88+ CPU: AMD64 Family 23 Model 96 Stepping 1, AuthenticAMD
89+ CPU Cores: 8
90+ Memory: 15.42GB
91+ GPU: NVIDIA
92+ GPU Model: NVIDIA GeForce GTX 1660 Ti
93+ GPU Memory: 6144.0GB
94+ Compute Backend: CUDA
95+
96+ 🚀 Setting up Solo Server...
97+ ✅ Solo server is ready!
98+ ```
99+
100+ ---
101+
78102## ** Commands**
79- ### ** 1️⃣ Pull a Model**
103+ ### ** 1️⃣ Pull & Run a Model**
80104``` sh
81- solo pull llama3
105+ solo run llama3.2
82106```
83-
84107
85108---
86109
@@ -96,6 +119,39 @@ http://127.0.0.1:5070 #SOLO_SERVER_PORT
96119
97120---
98121
122+ ## Diagram
123+
124+ ```
125+ +-------------------+
126+ | |
127+ | solo run llama3.2 |
128+ | |
129+ +---------+---------+
130+ |
131+ |
132+ | +------------------+ +----------------------+
133+ | | Pull inferencing | | Pull model layer |
134+ +-----------| runtime (cuda) |---------->| llama3.2 |
135+ +------------------+ +----------------------+
136+ | Repo options |
137+ ++-----------+--------++
138+ | | |
139+ v v v
140+ +----------+ +----------+ +-------------+
141+ | Ollama | | vLLM | | HuggingFace |
142+ | Registry | | registry | | Registry |
143+ +-----+------+---+------+-++------------+
144+ | | |
145+ v v v
146+ +---------------------+
147+ | Start with |
148+ | cuda runtime |
149+ | and |
150+ | llama3.2 |
151+ +---------------------+
152+ ```
153+ ---
154+
99155### ** 3️⃣ Benchmark a Model**
100156``` sh
101157solo benchmark llama3
@@ -148,12 +204,12 @@ solo status
148204
149205### ** 5️⃣ Stop a Model**
150206``` sh
151- solo stop llama3
207+ solo stop
152208```
153209** Example Output:**
154210``` sh
155- Stopping llama3 ...
156- llama3 stopped successfully.
211+ 🛑 Stopping Solo Server ...
212+ ✅ Solo server stopped successfully.
157213```
158214
159215---
0 commit comments