Skip to content

Commit

Permalink
[Bugfix][Kernel] Give unique name to BlockSparseFlashAttention (vllm-…
Browse files Browse the repository at this point in the history
…project#12040)

Signed-off-by: Chen Zhang <zhangch99@outlook.com>
  • Loading branch information
heheda12345 authored and jikunshang committed Jan 21, 2025
1 parent 2dcf45f commit 8e074fb
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
3 changes: 1 addition & 2 deletions vllm/attention/backends/blocksparse_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -89,8 +89,7 @@ class BlocksparseFlashAttentionBackend(AttentionBackend):

@staticmethod
def get_name() -> str:
# For attention layer compatibility
return "FLASH_ATTN"
return "BLOCK_SPARSE_FLASH_ATTN"

@staticmethod
def get_impl_cls() -> Type["BlocksparseFlashAttentionImpl"]:
Expand Down
1 change: 1 addition & 0 deletions vllm/platforms/interface.py
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,7 @@ class _Backend(enum.Enum):
PALLAS = enum.auto()
IPEX = enum.auto()
IPEX_V1 = enum.auto()
BLOCK_SPARSE_FLASH_ATTN = enum.auto()
NO_ATTENTION = enum.auto()


Expand Down

0 comments on commit 8e074fb

Please # to comment.