Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
48 commits
Select commit Hold shift + click to select a range
cc2e093
feat: use llama.cpp server
sangjanai Mar 11, 2025
70caa83
Merge branch 'dev' of https://github.com/janhq/cortex.cpp into s/feat…
sangjanai Mar 17, 2025
0968abe
chore: cleanup
sangjanai Mar 17, 2025
668af84
Merge branch 'dev' of https://github.com/janhq/cortex.cpp into s/feat…
sangjanai Mar 17, 2025
219d460
feat: OAI
sangjanai Mar 18, 2025
bb5cc35
Merge branch 'dev' of https://github.com/janhq/cortex.cpp into s/feat…
sangjanai Mar 18, 2025
8c4ca06
fix: wait for child process up
sangjanai Mar 19, 2025
220a974
Merge branch 'dev' of https://github.com/janhq/cortex.cpp into s/feat…
sangjanai Mar 19, 2025
0f2fa6e
chore: cleanup
sangjanai Mar 19, 2025
6dd7f7c
chore: cleanup
sangjanai Mar 19, 2025
bc92732
chore: fix unit tests
sangjanai Mar 19, 2025
5a68356
Merge branch 's/feat/spawn-llama-cpp' of https://github.com/janhq/cor…
sangjanai Mar 19, 2025
7c4d964
chore: cleanup
sangjanai Mar 19, 2025
1e5beaf
chore: cleanup
sangjanai Mar 19, 2025
b7b772b
fix: unit tests
sangjanai Mar 19, 2025
7c66135
fix: e2e tests
sangjanai Mar 19, 2025
af28b07
Merge branch 's/feat/spawn-llama-cpp' of https://github.com/janhq/cor…
sangjanai Mar 19, 2025
4705a1b
chore: cleanup
sangjanai Mar 19, 2025
6454eca
fix: e2e tests
sangjanai Mar 19, 2025
efe5a08
fix: validation
sangjanai Mar 19, 2025
a2886d2
fix: change GH user agent
sangjanai Mar 19, 2025
cf1a2ca
chore: cleanup
sangjanai Mar 19, 2025
282aae3
chore: fix unit tests
sangjanai Mar 19, 2025
e3dfea3
fix: e2e tests
sangjanai Mar 19, 2025
7438f94
fix: validation
sangjanai Mar 19, 2025
bf16c81
fix: change GH user agent
sangjanai Mar 19, 2025
f7cb0ae
chore: windows patch
vansangpfiev Mar 20, 2025
886222d
fix: windows
vansangpfiev Mar 20, 2025
87b2aa4
chore: log
vansangpfiev Mar 20, 2025
6d0215d
Merge branch 'dev' of https://github.com/menloresearch/cortex.cpp int…
sangjanai Mar 21, 2025
f866d5f
fix: handle macos 12 variants
sangjanai Mar 21, 2025
3d0b847
Merge branch 's/feat/spawn-llama-cpp' of https://github.com/menlorese…
sangjanai Mar 21, 2025
18e958e
Merge branch 'dev' of https://github.com/menloresearch/cortex.cpp int…
sangjanai Mar 21, 2025
de0de94
chore: e2e tests
sangjanai Mar 21, 2025
b783b22
chore: major version macos
sangjanai Mar 21, 2025
7832a5a
fix: windows e2e tests
vansangpfiev Mar 21, 2025
86cc89c
Merge branch 's/feat/spawn-llama-cpp' of github.com:janhq/nitro into …
vansangpfiev Mar 21, 2025
1c1146f
fix: engine list
sangjanai Mar 21, 2025
7487304
fix: macos filter
sangjanai Mar 21, 2025
1e99861
chore: skips some tests for linux arm
sangjanai Mar 21, 2025
70adc6e
fix: terminate process windows
vansangpfiev Mar 24, 2025
9846ad8
chore: release CIs (#2171)
vansangpfiev Mar 24, 2025
aba833e
fix: remove v in the llama-engine wget
sangjanai Mar 24, 2025
b239378
fix: add start time for model
sangjanai Mar 24, 2025
ca2180c
Merge branch 'dev' of https://github.com/menloresearch/cortex.cpp int…
sangjanai Mar 27, 2025
8ce2ac6
Merge branch 'dev' into s/feat/spawn-llama-cpp
vansangpfiev Apr 2, 2025
5f36501
chore: quality gate
sangjanai Apr 2, 2025
300f368
Merge branch 'dev' into s/feat/spawn-llama-cpp
vansangpfiev Apr 3, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions engine/e2e-test/api/engines/test_api_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,14 +28,14 @@ def test_engines_get_llamacpp_should_be_successful(self):

# engines install
def test_engines_install_llamacpp_specific_version_and_variant(self):
data = {"version": "b4920", "variant": "linux-avx-x64"}
data = {"version": "b4932", "variant": "linux-avx-x64"}
response = requests.post(
"http://localhost:3928/v1/engines/llama-cpp/install", json=data
)
assert response.status_code == 200

def test_engines_install_llamacpp_specific_version_and_null_variant(self):
data = {"version": "b4920"}
data = {"version": "b4932"}
response = requests.post(
"http://localhost:3928/v1/engines/llama-cpp/install", json=data
)
Expand All @@ -62,7 +62,7 @@ async def test_engines_install_uninstall_llamacpp_with_only_version_should_be_fa
await wait_for_websocket_download_success_event(timeout=120)
assert install_response.status_code == 200

data = {"version": "b4920"}
data = {"version": "b4932"}
response = requests.delete(
"http://localhost:3928/v1/engines/llama-cpp/install", json=data
)
Expand All @@ -85,7 +85,7 @@ async def test_engines_install_uninstall_llamacpp_with_variant_should_be_success
def test_engines_install_uninstall_llamacpp_with_specific_variant_and_version_should_be_successful(
self,
):
data = {"variant": "linux-avx-x64", "version": "b4920"}
data = {"variant": "linux-avx-x64", "version": "b4932"}
# install first
install_response = requests.post(
"http://localhost:3928/v1/engines/llama-cpp/install", json=data
Expand Down
2 changes: 1 addition & 1 deletion engine/e2e-test/api/engines/test_api_get_default_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ def test_api_get_default_engine_successfully(self):
# Data test
engine= "llama-cpp"
name= "linux-avx-x64"
version= "b4920"
version= "b4932"

data = {"version": version, "variant": name}
post_install_url = f"http://localhost:3928/v1/engines/{engine}/install"
Expand Down
2 changes: 1 addition & 1 deletion engine/e2e-test/api/engines/test_api_get_list_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ def test_api_get_list_engines_successfully(self):
# Data test
engine= "llama-cpp"
name= "linux-avx-x64"
version= "b4920"
version= "b4932"

post_install_url = f"http://localhost:3928/v1/engines/{engine}/install"
response = requests.delete(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ def test_api_set_default_engine_successfully(self):
# Data test
engine= "llama-cpp"
name= "linux-avx-x64"
version= "b4920"
version= "b4932"

data = {"version": version, "variant": name}
post_install_url = f"http://localhost:3928/v1/engines/{engine}/install"
Expand Down
2 changes: 1 addition & 1 deletion engine/e2e-test/cli/engines/test_cli_engine_install.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ def test_engines_install_llamacpp_should_be_successfully(self):

@pytest.mark.skipif(platform.system() == "Windows", reason="Progress bar log issue on Windows")
def test_engines_install_pre_release_llamacpp(self):
engine_version = "b4920"
engine_version = "b4932"
exit_code, output, error = run(
"Install Engine",
["engines", "install", "llama-cpp", "-v", engine_version],
Expand Down
Loading