Conversation
|
Thanks for the contribution! For vLLM & SGLang, please ensure that your recipes is similar to the official vLLM recipes and/or the SGLang cookbook If it is not, please create a PR first before we can merge your PR into the master branch. Let's ensure that the documentation is first class such that the entire ML community can benefit from your hard work! Thank you |
|
@claude please active the e2e test with |
|
Claude finished @chunfangamd's task in 3s —— View job E2E Test Triggered for GLM-5
Workflow triggered: Run #24243787005 — Jobs created:
The workflow has been successfully dispatched. All 3 MI355X benchmark/eval jobs are queued and waiting for self-hosted MI355X runners to become available (~40min in queue so far). Once a runner picks up the jobs, the benchmarks will run automatically. Please monitor the run directly via the link above. You can tag --- | View workflow run | Branch: |
|
@claude, I canceled the job https://github.com/SemiAnalysisAI/InferenceX/actions/runs/24243787005. |
|
Claude finished @chunfangamd's task in 15s —— View job Acknowledged Job Cancellation
The e2e test run for If you'd like me to re-trigger the tests or help with anything else on this PR, just let me know! --- | Branch: |
|
Hi @chunfangamd This is new sweep based on sgl-dev nightly build: |
|
@billishyahao @chunfangamd nightly is fine for now, since this is a new model and the pace of development is rapid. feel free to change this PR to the nightly image |
change to nightly image
…0413 Updated the perf-changelog too
Upgrade GLM5-FP8-MI355X-SGLang Image to v0.5.10rc0-rocm720-mi35x-20260413 after fixing the Slurm shared cache issue. With this upstream daily image, we use a new aiter 0.1.12.post1 which can improve the performance.
Co-authored with: @1am9trash @zhentaocc
Thanks to @cquil11 @billishyahao