Skip to content

Commit 39456f3

Browse files
authored
Remove mi300a (ROCm#440)
* Removing gfx940 and gfx941 targets. These have been deprecated in favor of gfx942 for MI300X Signed-off-by: Gregory Shtrasberg <[email protected]> * Remove from custom kernels as well --------- Signed-off-by: Gregory Shtrasberg <[email protected]>
1 parent b63a984 commit 39456f3

File tree

5 files changed

+6
-11
lines changed

5 files changed

+6
-11
lines changed

CMakeLists.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ set(PYTHON_SUPPORTED_VERSIONS "3.9" "3.10" "3.11" "3.12")
3434
set(CUDA_SUPPORTED_ARCHS "7.0;7.2;7.5;8.0;8.6;8.7;8.9;9.0")
3535

3636
# Supported AMD GPU architectures.
37-
set(HIP_SUPPORTED_ARCHS "gfx906;gfx908;gfx90a;gfx940;gfx941;gfx942;gfx1030;gfx1100;gfx1101;gfx1200;gfx1201")
37+
set(HIP_SUPPORTED_ARCHS "gfx906;gfx908;gfx90a;gfx942;gfx1030;gfx1100;gfx1101;gfx1200;gfx1201")
3838

3939
#
4040
# Supported/expected torch versions for CUDA/ROCm.

csrc/quantization/fp8/amd/hip_float8_impl.h

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
#pragma once
22

3-
#if defined(__HIPCC__) && \
4-
(defined(__gfx940__) || defined(__gfx941__) || defined(__gfx942__))
3+
#if defined(__HIPCC__) && defined(__gfx942__)
54
#define __HIP__MI300__
65
#endif
76

csrc/rocm/attention.cu

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,7 @@
2424
#include "../attention/dtype_fp8.cuh"
2525
#include "../quantization/fp8/amd/quant_utils.cuh"
2626

27-
#if defined(__HIPCC__) && (defined(__gfx90a__) || defined(__gfx940__) || \
28-
defined(__gfx941__) || defined(__gfx942__))
27+
#if defined(__HIPCC__) && (defined(__gfx90a__) || defined(__gfx942__))
2928
#define __HIP__MI300_MI250__
3029
#endif
3130

csrc/rocm/custom_kernels.cu

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,11 @@
55
#include <algorithm>
66
#include "cuda_compat.h"
77

8-
#if defined(__HIPCC__) && (defined(__gfx90a__) || defined(__gfx940__) || \
9-
defined(__gfx941__) || defined(__gfx942__))
8+
#if defined(__HIPCC__) && (defined(__gfx90a__) || defined(__gfx942__))
109
#define __HIP__MI300_MI250__
1110
#endif
1211

13-
#if defined(__HIPCC__) && \
14-
(defined(__gfx940__) || defined(__gfx941__) || defined(__gfx942__))
12+
#if defined(__HIPCC__) && defined(__gfx942__)
1513
#define __HIP__MI300__
1614
#endif
1715

vllm/attention/backends/rocm_flash_attn.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,8 +25,7 @@
2525
_PARTITION_SIZE_ROCM = 256
2626
_GPU_ARCH = torch.cuda.get_device_properties("cuda").gcnArchName
2727
_ON_NAVI = "gfx1" in _GPU_ARCH
28-
_ON_MI250_MI300 = any(arch in _GPU_ARCH
29-
for arch in ["gfx90a", "gfx940", "gfx941", "gfx942"])
28+
_ON_MI250_MI300 = any(arch in _GPU_ARCH for arch in ["gfx90a", "gfx942"])
3029

3130

3231
class ROCmFlashAttentionBackend(AttentionBackend):

0 commit comments

Comments
 (0)