You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
> ⚠️ **Cortex is currently in Development**: Expect breaking changes and bugs!
12
12
13
13
## About
14
-
Cortex is an openAI-compatible local AI server that developers can use to build LLM apps. It is packaged with a Docker-inspired command-line interface and a Typescript client library. It can be used as a standalone server, or imported as a library.
14
+
Cortex is an OpenAI-compatible AI engine that developers can use to build LLM apps. It is packaged with a Docker-inspired command-line interface and client libraries. It can be used as a standalone server or imported as a library.
15
15
16
-
Cortex currently supports two inference engines:
16
+
Cortex currently supports 3 inference engines:
17
17
18
18
- Llama.cpp
19
+
- ONNX Runtime
19
20
- TensorRT-LLM
20
21
21
-
> Read more about Cortex at https://jan.ai/cortex
22
-
23
22
## Quicklinks
24
-
Cortex
25
-
-[Website](https://jan.ai/)
26
-
-[GitHub](https://github.com/janhq/cortex)
27
-
-[User Guides](https://jan.ai/cortex)
28
-
-[API reference](https://jan.ai/api-reference)
29
-
30
-
## Prerequisites
31
-
32
-
### **Dependencies**
33
-
34
-
Before installation, ensure that you have installed the following:
35
-
-**Node.js**: version 18 and above is required to run the installation.
36
-
-**NPM**: Needed to manage packages.
37
-
-**CPU Instruction Sets**: Available for download from the [Cortex GitHub Releases](https://github.com/janhq/cortex/releases) page.
38
-
39
-
40
-
>💡 The **CPU instruction sets** are not required for the initial installation of Cortex. This dependency will be automatically installed during the Cortex initialization if they are not already on your system.
41
-
42
23
43
-
### **Hardware**
24
+
-[Homepage](https://cortex.jan.ai/)
25
+
-[Docs](https://cortex.jan.ai/docs/)
44
26
27
+
## Quickstart
28
+
### Prerequisites
45
29
Ensure that your system meets the following requirements to run Cortex:
46
-
30
+
-**Dependencies**:
31
+
-**Node.js**: version 18 and above is required to run the installation.
32
+
-**NPM**: Needed to manage packages.
33
+
-**CPU Instruction Sets**: Available for download from the [Cortex GitHub Releases](https://github.com/janhq/cortex/releases) page.
47
34
-**OS**:
48
35
- MacOSX 13.6 or higher.
49
36
- Windows 10 or higher.
50
37
- Ubuntu 22.04 and later.
51
-
-**RAM (CPU Mode):**
52
-
- 8GB for running up to 3B models.
53
-
- 16GB for running up to 7B models.
54
-
- 32GB for running up to 13B models.
55
-
-**VRAM (GPU Mode):**
56
38
57
-
- 6GB can load the 3B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
58
-
- 8GB can load the 7B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
59
-
- 12GB can load the 13B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
39
+
> Visit [Quickstart](https://cortex.jan.ai/docs/quickstart) to get started.
60
40
61
-
-**Disk**: At least 10GB for app and model download.
62
-
63
-
## Quickstart
64
-
To install Cortex CLI, follow the steps below:
65
-
1. Install the Cortex NPM package globally:
41
+
### NPM
66
42
```bash
43
+
# Install using NPM
67
44
npm i -g cortexso
68
-
```
69
-
> Cortex automatically detects your CPU and GPU, downloading the appropriate CPU instruction sets and required dependencies to optimize GPU performance.
70
-
71
-
2. Download a GGUF model from Hugging Face:
72
-
```bash
73
-
# Pull a model most compatible with your hardware
74
-
cortex pull llama3
75
-
76
-
# Pull a specific variant with `repo_name:branch`
77
-
cortex pull llama3:7b
78
-
79
-
# Pull a model with the HuggingFace `model_id`
80
-
cortex pull microsoft/Phi-3-mini-4k-instruct-gguf
81
-
```
82
-
3. Load the model:
83
-
```bash
84
-
cortex models start llama3:7b
45
+
# Install using Brew
46
+
brew tap janhq/cortexso
47
+
brew install cortexso
48
+
# Run model
49
+
cortex run llama3
50
+
# To uninstall globally using NPM
51
+
npm uninstall -g cortexso
85
52
```
86
53
87
-
4. Start chatting with the model:
54
+
### Homebrew
88
55
```bash
89
-
cortex chat tell me a joke
56
+
# Install using Brew
57
+
brew tap janhq/cortexso
58
+
brew install cortexso
59
+
# Run model
60
+
cortex run llama3
61
+
# To uninstall using Brew
62
+
brew uninstall cortexso
63
+
brew untap janhq/cortexso
90
64
```
65
+
> You can also install Cortex using the Cortex Installer available on [GitHub Releases](https://github.com/janhq/cortex/releases).
91
66
92
-
93
-
## Run as an API server
94
67
To run Cortex as an API server:
95
68
```bash
96
69
cortex serve
70
+
71
+
# Output
72
+
# Started server at http://localhost:1337
73
+
# Swagger UI available at http://localhost:1337/api
97
74
```
98
75
76
+
You can now access the Cortex API server at `http://localhost:1337`,
77
+
and the Swagger UI at `http://localhost:1337/api`.
78
+
99
79
## Build from Source
100
80
101
81
To install Cortex from the source, follow the steps below:
'Cortex API provides a command-line interface (CLI) for seamless interaction with large language models (LLMs). Fully compatible with the [OpenAI API](https://platform.openai.com/docs/api-reference), it enables straightforward command execution and management of LLM interactions.',
26
+
'Cortex API provides a command-line interface (CLI) for seamless interaction with large language models (LLMs). It is fully compatible with the [OpenAI API](https://platform.openai.com/docs/api-reference) and enables straightforward command execution and management of LLM interactions.',
0 commit comments