You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
> ⚠️ **Cortex.cpp is currently in active development. This outlines the intended behavior of Cortex, which may not yet be fully implemented in the codebase.**
22
+
> **Cortex.cpp is currently in active development.**
23
23
24
24
## Overview
25
25
26
-
Cortex.cpp is a Local AI engine that is used to run and customize LLMs. Cortex can be deployed as a standalone server, or integrated into apps like [Jan.ai](https://jan.ai/).
26
+
Cortex is a Local AI API Platform that is used to run and customize LLMs.
27
27
28
-
Cortex.cpp is a multi-engine that uses `llama.cpp` as the default engine but also supports the following:
28
+
Key Features:
29
+
- Straightforward CLI (inspired by Ollama)
30
+
- Full C++ implementation, packageable into Desktop and Mobile apps
31
+
- Pull from Huggingface of Cortex Built-in Model Library
32
+
- Models stored in universal file formats (vs blobs)
| openhermes-2.5 | ✅ | cortex run openhermes-2.5 |
90
-
| phi3 | ✅ | cortex run phi3 |
91
124
| tinyllama | ✅ | cortex run tinyllama |
92
125
93
-
> **Note**:
94
-
> You should have at least 8 GB of RAM available to run the 7B models, 16 GB to run the 14B models, and 32 GB to run the 32B models.
95
-
96
-
## Cortex.cpp CLI Commands
126
+
View all [Cortex Built-in Models](https://cortex.so/models).
97
127
98
-
For complete details on CLI commands, please refer to our [CLI documentation](https://cortex.so/docs/cli).
99
-
100
-
## REST API
101
-
102
-
Cortex.cpp includes a REST API accessible at `localhost:39281`. For a complete list of endpoints and their usage, visit our [API documentation](https://cortex.so/api-reference).
128
+
Cortex supports multiple quantizations for each model.
129
+
```
130
+
❯ cortex-nightly pull llama3.2
131
+
Downloaded models:
132
+
llama3.2:3b-gguf-q2-k
133
+
134
+
Available to download:
135
+
1. llama3.2:3b-gguf-q3-kl
136
+
2. llama3.2:3b-gguf-q3-km
137
+
3. llama3.2:3b-gguf-q3-ks
138
+
4. llama3.2:3b-gguf-q4-km (default)
139
+
5. llama3.2:3b-gguf-q4-ks
140
+
6. llama3.2:3b-gguf-q5-km
141
+
7. llama3.2:3b-gguf-q5-ks
142
+
8. llama3.2:3b-gguf-q6-k
143
+
9. llama3.2:3b-gguf-q8-0
144
+
145
+
Select a model (1-9):
146
+
```
103
147
104
148
## Advanced Installation
105
149
@@ -129,13 +173,17 @@ Cortex.cpp is available with a Network Installer, which is a smaller installer b
129
173
### Beta & Nightly Versions
130
174
131
175
Cortex releases 2 preview versions for advanced users to try new features early (we appreciate your feedback!)
132
-
- Beta (early preview)
133
-
- Nightly (released every night). Nightly pulls the latest changes from upstream [llama.cpp](https://github.com/ggerganov/llama.cpp/) repo, creates a PR and runs our tests. If all test pass, the PR is automatically merged into our repo, with the latest llama.cpp version.
176
+
- Beta (early preview)
177
+
- CLI command: `cortex-beta`
178
+
- Nightly (released every night)
179
+
- CLI Command: `cortex-nightly`
180
+
- Nightly automatically pulls the latest changes from upstream [llama.cpp](https://github.com/ggerganov/llama.cpp/) repo, creates a PR and runs tests.
181
+
- If all test pass, the PR is automatically merged into our repo, with the latest llama.cpp version.
0 commit comments