[Distributed] Add send and recv helpers (#5719)
This commit is contained in:
committed by
GitHub
parent
6c916ac8a8
commit
5d4d90536f
@ -12,8 +12,7 @@ from vllm.distributed.parallel_state import (get_tensor_model_parallel_group,
|
||||
get_tp_group, graph_capture)
|
||||
|
||||
from ..utils import (ensure_model_parallel_initialized,
|
||||
init_test_distributed_environment,
|
||||
multi_process_tensor_parallel)
|
||||
init_test_distributed_environment, multi_process_parallel)
|
||||
|
||||
random.seed(42)
|
||||
test_sizes = [random.randint(1024, 2048 * 1024) for _ in range(8)]
|
||||
@ -113,4 +112,4 @@ def test_custom_allreduce(tp_size, pipeline_parallel_size, test_target):
|
||||
world_size = tp_size * pipeline_parallel_size
|
||||
if world_size > torch.cuda.device_count():
|
||||
pytest.skip("Not enough GPUs to run the test.")
|
||||
multi_process_tensor_parallel(tp_size, pipeline_parallel_size, test_target)
|
||||
multi_process_parallel(tp_size, pipeline_parallel_size, test_target)
|
||||
|
||||
Reference in New Issue
Block a user