Commit f9ecbb1 1 parent 02222a0 commit f9ecbb1 Copy full SHA for f9ecbb1
File tree 1 file changed +3
-5
lines changed
1 file changed +3
-5
lines changed Original file line number Diff line number Diff line change 17
17
is_all_cross_attn_metadata_set , is_all_encoder_attn_metadata_set )
18
18
from vllm .attention .ops .paged_attn import (PagedAttention ,
19
19
PagedAttentionMetadata )
20
- from vllm .logger import init_logger
21
-
22
- logger = init_logger (__name__ )
20
+ from vllm .utils import print_warning_once
23
21
24
22
25
23
class XFormersBackend (AttentionBackend ):
@@ -386,8 +384,8 @@ def __init__(
386
384
raise ValueError (
387
385
"XFormers does not support block-sparse attention." )
388
386
if logits_soft_cap is not None :
389
- raise ValueError (
390
- "XFormers does not support attention logits soft capping ." )
387
+ print_warning_once ( "XFormers does not support logits soft cap. "
388
+ "Outputs may be slightly off ." )
391
389
self .num_heads = num_heads
392
390
self .head_size = head_size
393
391
self .scale = float (scale )
You can’t perform that action at this time.
0 commit comments