You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
Double-click on Nitro to run it. After downloading your model, make sure it's saved to a specific path. Then, make an API call to load your model into Nitro.
51
45
52
46
```zsh
@@ -59,9 +53,10 @@ curl -X POST 'http://localhost:3928/inferences/llamacpp/loadmodel' \
59
53
"embedding": true
60
54
}'
61
55
```
62
-
ctx_len and ngl are typical llama C++ parameters, and embedding determines whether to enable the embedding endpoint or not.
63
56
64
-
Step 4: Perform Inference on Nitro for the First Time
57
+
`ctx_len` and `ngl` are typical llama C++ parameters, and `embedding` determines whether to enable the embedding endpoint or not.
58
+
59
+
**Step 4: Perform Inference on Nitro for the First Time**
@@ -101,9 +96,7 @@ Nitro is an integration layer with the most cutting-edge inference engine. Its s
101
96
102
97
### Contact
103
98
104
-
- For support: please file a GitHub ticket
105
-
- For questions: join our Discord [here](https://discord.gg/FTk2MvZwJH)
106
-
- For long-form inquiries: please email hello@jan.ai
107
-
```
99
+
- For support, please file a GitHub ticket.
100
+
- For questions, join our Discord [here](https://discord.gg/FTk2MvZwJH).
101
+
- For long-form inquiries, please email hello@jan.ai.
108
102
109
-
I've made formatting improvements and fixed some grammatical issues. If you have any further questions or need additional assistance, please let me know!
0 commit comments