From 224c265fd7e7c362256446da31c3de40e969f5d0 Mon Sep 17 00:00:00 2001 From: Vaibhavs10 Date: Tue, 12 Aug 2025 22:48:59 +0200 Subject: [PATCH 1/2] Update transformer to avoid triton errors. --- articles/gpt-oss/run-transformers.md | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/articles/gpt-oss/run-transformers.md b/articles/gpt-oss/run-transformers.md index 8149dc16f7..0a1ec12364 100644 --- a/articles/gpt-oss/run-transformers.md +++ b/articles/gpt-oss/run-transformers.md @@ -29,11 +29,7 @@ If you use `bfloat16` instead of MXFP4, memory consumption will be larger (\~48 It’s recommended to create a fresh Python environment. Install transformers, accelerate, as well as the Triton kernels for MXFP4 compatibility: ```bash -pip install -U transformers accelerate torch triton kernels -``` - -```bash -pip install git+https://github.com/triton-lang/triton.git@main#subdirectory=python/triton_kernels +pip install -U transformers accelerate torch triton==3.4 kernels ``` 2. **(Optional) Enable multi-GPU** From bd0ed3ce1e7b0099cf99b25157970a363a68827b Mon Sep 17 00:00:00 2001 From: Vaibhavs10 Date: Tue, 12 Aug 2025 22:53:10 +0200 Subject: [PATCH 2/2] Update dependencies in the colab notebook as well. --- articles/gpt-oss/run-colab.ipynb | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/articles/gpt-oss/run-colab.ipynb b/articles/gpt-oss/run-colab.ipynb index 8c3b0609bc..a7bab62dce 100644 --- a/articles/gpt-oss/run-colab.ipynb +++ b/articles/gpt-oss/run-colab.ipynb @@ -65,7 +65,7 @@ }, "outputs": [], "source": [ - "!pip install -q git+https://github.com/huggingface/transformers triton==3.4 kernels" + "!pip install -q transformers triton==3.4 kernels" ] }, { @@ -244,4 +244,4 @@ }, "nbformat": 4, "nbformat_minor": 0 -} \ No newline at end of file +}