Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit 515978b

Browse files
committed
manual install docs add
1 parent c980c16 commit 515978b

File tree

2 files changed

+19
-12
lines changed

2 files changed

+19
-12
lines changed

README.md

Lines changed: 19 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,23 @@
2323

2424
## Documentation
2525

26+
## About Nitro
27+
28+
Nitro is a light-weight integration layer (and soon to be inference engine) for cutting edge inference engine, make deployment of AI models easier than ever before!
29+
30+
The binary of nitro after zipped is only ~3mb in size with none to minimal dependencies (if you use a GPU need CUDA for example) make it desirable for any edge/server deployment 👍.
31+
32+
### Repo Structure
33+
34+
```
35+
.
36+
├── controllers
37+
├── docs
38+
├── llama.cpp -> Upstream llama C++
39+
├── nitro_deps -> Dependencies of the Nitro project as a sub-project
40+
└── utils
41+
```
42+
2643
## Quickstart
2744

2845
**Step 1: Download Nitro**
@@ -76,18 +93,8 @@ curl --location 'http://localhost:3928/inferences/llamacpp/chat_completion' \
7693

7794
Nitro server is compatible with the OpenAI format, so you can expect the same output as the OpenAI ChatGPT API.
7895

79-
## About Nitro
80-
81-
### Repo Structure
82-
83-
```
84-
.
85-
├── controllers
86-
├── docs
87-
├── llama.cpp -> Upstream llama C++
88-
├── nitro_deps -> Dependencies of the Nitro project as a sub-project
89-
└── utils
90-
```
96+
## Compile from source
97+
To compile nitro please visit [Compile from source](docs/manual_install.md)
9198

9299
### Architecture
93100
Nitro is an integration layer with the most cutting-edge inference engine. Its structure can be simplified as follows:
File renamed without changes.

0 commit comments

Comments
 (0)