vllm.distributed.communication_op
broadcast_tensor_dict ¶
broadcast_tensor_dict(
tensor_dict: Optional[
dict[Any, Union[Tensor, Any]]
] = None,
src: int = 0,
)
Source code in vllm/distributed/communication_op.py
tensor_model_parallel_all_gather ¶
All-gather the input tensor across model parallel group.
tensor_model_parallel_all_reduce ¶
tensor_model_parallel_gather ¶
Gather the input tensor across model parallel group.
tensor_model_parallel_reduce_scatter ¶
Reduce-Scatter the input tensor across model parallel group.