diff --git a/articles/gpt-oss/run-colab.ipynb b/articles/gpt-oss/run-colab.ipynb index 8c3b0609bc..a7bab62dce 100644 --- a/articles/gpt-oss/run-colab.ipynb +++ b/articles/gpt-oss/run-colab.ipynb @@ -65,7 +65,7 @@ }, "outputs": [], "source": [ - "!pip install -q git+https://github.com/huggingface/transformers triton==3.4 kernels" + "!pip install -q transformers triton==3.4 kernels" ] }, { @@ -244,4 +244,4 @@ }, "nbformat": 4, "nbformat_minor": 0 -} \ No newline at end of file +} diff --git a/articles/gpt-oss/run-transformers.md b/articles/gpt-oss/run-transformers.md index 8149dc16f7..0a1ec12364 100644 --- a/articles/gpt-oss/run-transformers.md +++ b/articles/gpt-oss/run-transformers.md @@ -29,11 +29,7 @@ If you use `bfloat16` instead of MXFP4, memory consumption will be larger (\~48 It’s recommended to create a fresh Python environment. Install transformers, accelerate, as well as the Triton kernels for MXFP4 compatibility: ```bash -pip install -U transformers accelerate torch triton kernels -``` - -```bash -pip install git+https://github.com/triton-lang/triton.git@main#subdirectory=python/triton_kernels +pip install -U transformers accelerate torch triton==3.4 kernels ``` 2. **(Optional) Enable multi-GPU**