We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 6b88f1f commit e078b90Copy full SHA for e078b90
vllm/v1/attention/backends/flash_attn.py
@@ -85,7 +85,7 @@ def __init__(
85
if sliding_window is None:
86
self.sliding_window = (-1, -1)
87
else:
88
- self.sliding_window = ((sliding_window - 1, 0))
+ self.sliding_window = (sliding_window - 1, 0)
89
self.kv_cache_dtype = kv_cache_dtype
90
if logits_soft_cap is None:
91
# In flash-attn, setting logits_soft_cap as 0 means no soft cap.
0 commit comments