We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 56eef8b commit b3b2136Copy full SHA for b3b2136
vllm/config.py
@@ -869,6 +869,13 @@ def __init__(
869
f"distributed executor backend "
870
f"'{self.distributed_executor_backend}'.")
871
872
+ if current_platform.is_tpu() and self.world_size > 1:
873
+ if self.distributed_executor_backend is None:
874
+ self.distributed_executor_backend = "ray"
875
+ if self.distributed_executor_backend != "ray":
876
+ raise ValueError(
877
+ "TPU backend only supports Ray for distributed inference.")
878
+
879
if self.distributed_executor_backend is None and self.world_size > 1:
880
# We use multiprocessing by default if world_size fits on the
881
# current node and we aren't in a ray placement group.
0 commit comments