Skip to content

Commit a2e9684

Browse files
committed
[multihost] Integrate expert parallelism to RayExecutor
Signed-off-by: Lihao Ran <imlihao.ran@gmail.com>
1 parent f72797f commit a2e9684

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

tpu_inference/executors/ray_distributed_executor.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -366,7 +366,7 @@ def sort_by_driver_then_worker_ip(item: RayWorkerMetaData):
366366
for pp_rank in range(self.parallel_config.pipeline_parallel_size):
367367
self.pp_tp_workers.append([])
368368
num_tp_workers = int(
369-
self.parallel_config.tensor_parallel_size //
369+
self.vllm_config.sharding_config.total_devices //
370370
num_tpu_per_worker)
371371
for tp_rank in range(num_tp_workers):
372372
# PP=2, TP=4, num_tpu_per_worker=2

0 commit comments

Comments
 (0)