Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit 6d3d5aa

Browse files
committed
chore: update README.md
1 parent 6c22057 commit 6d3d5aa

File tree

2 files changed

+36
-90
lines changed

2 files changed

+36
-90
lines changed

README.md

Lines changed: 18 additions & 80 deletions
Original file line numberDiff line numberDiff line change
@@ -1,76 +1,7 @@
11
# Cortex Monorepo
22

3-
This monorepo contains two projects: CortexJS and CortexCPP.
4-
5-
## CortexJS: Stateful Business Backend
6-
7-
* All of the stateful endpoints:
8-
+ /threads
9-
+ /messages
10-
+ /models
11-
+ /runs
12-
+ /vector_store
13-
+ /settings
14-
+ /?auth
15-
+ …
16-
* Database & Filesystem
17-
* API Gateway
18-
* Authentication & Authorization
19-
* Observability
20-
21-
## CortexCPP: Stateless Embedding Backend
22-
23-
* All of the high performance, stateless endpoints:
24-
+ /chat/completion
25-
+ /audio
26-
+ /fine_tuning
27-
+ /embeddings
28-
+ /load_model
29-
+ /unload_model
30-
* Kernel - Hardware Recognition
31-
32-
## Project Structure
33-
34-
```
35-
.
36-
├── cortex-js/
37-
│ ├── package.json
38-
│ ├── README.md
39-
│ ├── Dockerfile
40-
│ ├── docker-compose.yml
41-
│ ├── src/
42-
│ │ ├── controllers/
43-
│ │ ├── modules/
44-
│ │ ├── services/
45-
│ │ └── ...
46-
│ └── ...
47-
├── cortex-cpp/
48-
│ ├── app/
49-
│ │ ├── controllers/
50-
│ │ ├── models/
51-
│ │ ├── services/
52-
│ │ ├── ?engines/
53-
│ │ │ ├── llama.cpp
54-
│ │ │ ├── tensorrt-llm
55-
│ │ │ └── ...
56-
│ │ └── ...
57-
│ ├── CMakeLists.txt
58-
│ ├── config.json
59-
│ ├── Dockerfile
60-
│ ├── docker-compose.yml
61-
│ ├── README.md
62-
│ └── ...
63-
├── scripts/
64-
│ └── ...
65-
├── README.md
66-
├── package.json
67-
├── Dockerfile
68-
├── docker-compose.yml
69-
└── docs/
70-
└── ...
71-
```
72-
73-
# Install
3+
# Installation
4+
745
## Prerequisites
756

767
### **Dependencies**
@@ -91,17 +22,18 @@ Before installation, ensure that you have installed the following:
9122
Ensure that your system meets the following requirements to run Cortex:
9223

9324
- **OS**:
94-
- MacOSX 13.6 or higher.
95-
- Windows 10 or higher.
96-
- Ubuntu 12.04 and later.
25+
- MacOSX 13.6 or higher.
26+
- Windows 10 or higher.
27+
- Ubuntu 12.04 and later.
9728
- **RAM (CPU Mode):**
98-
- 8GB for running up to 3B models.
99-
- 16GB for running up to 7B models.
100-
- 32GB for running up to 13B models.
29+
- 8GB for running up to 3B models.
30+
- 16GB for running up to 7B models.
31+
- 32GB for running up to 13B models.
10132
- **VRAM (GPU Mode):**
102-
- 6GB can load the 3B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
103-
- 8GB can load the 7B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
104-
- 12GB can load the 13B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
33+
34+
- 6GB can load the 3B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
35+
- 8GB can load the 7B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
36+
- 12GB can load the 13B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
10537

10638
- **Disk**: At least 10GB for app and model download.
10739

@@ -152,6 +84,7 @@ cortex init
15284
> Nvidia
15385
Others (Vulkan)
15486
```
87+
15588
3. Select CPU instructions (will be deprecated soon).
15689

15790
```bash
@@ -165,22 +98,27 @@ cortex init
16598
2. Once downloaded, Cortex is ready to use!
16699

167100
### Step 4: Pull a model
101+
168102
From HuggingFace
103+
169104
```bash
170105
cortex pull janhq/phi-3-medium-128k-instruct-GGUF
171106
```
172107

173108
From Jan Hub (TBD)
109+
174110
```bash
175111
cortex pull llama3
176112
```
177113

178114
### Step 5: Chat
115+
179116
```bash
180117
cortex run janhq/phi-3-medium-128k-instruct-GGUF
181118
```
182119

183120
## Run as an API server
121+
184122
```bash
185123
cortex serve
186124
```

cortex-js/README.md

Lines changed: 18 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
1-
# Install
1+
# Installation
2+
23
## Prerequisites
34

45
### **Dependencies**
@@ -19,17 +20,18 @@ Before installation, ensure that you have installed the following:
1920
Ensure that your system meets the following requirements to run Cortex:
2021

2122
- **OS**:
22-
- MacOSX 13.6 or higher.
23-
- Windows 10 or higher.
24-
- Ubuntu 12.04 and later.
23+
- MacOSX 13.6 or higher.
24+
- Windows 10 or higher.
25+
- Ubuntu 12.04 and later.
2526
- **RAM (CPU Mode):**
26-
- 8GB for running up to 3B models.
27-
- 16GB for running up to 7B models.
28-
- 32GB for running up to 13B models.
27+
- 8GB for running up to 3B models.
28+
- 16GB for running up to 7B models.
29+
- 32GB for running up to 13B models.
2930
- **VRAM (GPU Mode):**
30-
- 6GB can load the 3B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
31-
- 8GB can load the 7B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
32-
- 12GB can load the 13B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
31+
32+
- 6GB can load the 3B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
33+
- 8GB can load the 7B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
34+
- 12GB can load the 13B model (int4) with `ngl` at 120 ~ full speed on CPU/ GPU.
3335

3436
- **Disk**: At least 10GB for app and model download.
3537

@@ -80,6 +82,7 @@ cortex init
8082
> Nvidia
8183
Others (Vulkan)
8284
```
85+
8386
3. Select CPU instructions (will be deprecated soon).
8487

8588
```bash
@@ -93,22 +96,27 @@ cortex init
9396
2. Once downloaded, Cortex is ready to use!
9497

9598
### Step 4: Pull a model
99+
96100
From HuggingFace
101+
97102
```bash
98103
cortex pull janhq/phi-3-medium-128k-instruct-GGUF
99104
```
100105

101106
From Jan Hub (TBD)
107+
102108
```bash
103109
cortex pull llama3
104110
```
105111

106112
### Step 5: Chat
113+
107114
```bash
108115
cortex run janhq/phi-3-medium-128k-instruct-GGUF
109116
```
110117

111118
## Run as an API server
119+
112120
```bash
113121
cortex serve
114122
```

0 commit comments

Comments
 (0)