File tree Expand file tree Collapse file tree 1 file changed +4
-3
lines changed
vllm/v1/attention/backends Expand file tree Collapse file tree 1 file changed +4
-3
lines changed Original file line number Diff line number Diff line change 22# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
33"""Attention layer with AiterFlashAttention."""
44from dataclasses import dataclass
5- from typing import ClassVar , Optional
5+ from typing import Optional
66
77import torch
88
1111from vllm .config import VllmConfig
1212from vllm .logger import init_logger
1313from vllm .platforms import current_platform
14- from vllm .v1 .attention .backends .utils import (AttentionMetadataBuilder ,
14+ from vllm .v1 .attention .backends .utils import (AttentionCGSupport ,
15+ AttentionMetadataBuilder ,
1516 CommonAttentionMetadata )
1617from vllm .v1 .kv_cache_interface import AttentionSpec
1718
@@ -231,7 +232,7 @@ class AiterFlashAttentionMetadata:
231232
232233class AiterFlashAttentionMetadataBuilder (
233234 AttentionMetadataBuilder [AiterFlashAttentionMetadata ]):
234- full_cudagraph_supported : ClassVar [ bool ] = True
235+ cudagraph_support = AttentionCGSupport . ALWAYS
235236
236237 def __init__ (self , kv_cache_spec : AttentionSpec , layer_names : list [str ],
237238 vllm_config : VllmConfig , device : torch .device ):
You can’t perform that action at this time.
0 commit comments