
    Αi
                        S SK Jr  S SKJr  S SKJr  S SKJr  \(       a  S SKJr  S SK	J
r
  S SKJr      S
           SS	 jjrg)    )annotations)TYPE_CHECKING)	framework)stream)Tensor)task)GroupNc                t    [         R                  " 5       (       d   S5       e[        R                  " XX#U5      $ )a  

Gather tensors from all participators.

Args:
    tensor (Tensor): The input Tensor. Its data type
        should be float16, float32, float64, int32, int64, int8, uint8, bool or bfloat16.
    gather_list (list): A list of Tensors to hold the gathered tensors. Every element in the list must be a Tensor whose data type
        should be float16, float32, float64, int32, int64, int8, uint8, bool or bfloat16. Default value is None.
    dst (int): The dst rank id. Default value is 0.
    group (Group, optional): The group instance return by new_group or None for global default group.
    sync_op (bool, optional): Whether this op is a sync op. The default value is True.

Returns:
    Async work handle,which can be wait on, if async_op is set to True.
    None, if not async_op

Examples:
    .. code-block:: python

        >>> # doctest: +REQUIRES(env: DISTRIBUTED)
        >>> import paddle
        >>> import paddle.distributed as dist

        >>> dist.init_parallel_env()
        >>> gather_list = [] # type: ignore
        >>> if dist.get_rank() == 0:
        ...     data = paddle.to_tensor([1, 2, 3])
        ...     dist.gather(data, gather_list, dst=0)
        >>> else:
        ...     data = paddle.to_tensor([4, 5, 6])
        ...     dist.gather(data, gather_list, dst=0)
        >>> print(gather_list)
        >>> # [[1, 2, 3], [4, 5, 6]] (2 GPUs, out for rank 0)
        >>> # [] (2 GPUs, out for rank 1)
z-gather doesn't support static graph mode yet.)r   in_dynamic_moder   gather)tensorgather_listdstgroupsync_ops        g/var/www/html/banglarbhumi/venv/lib/python3.13/site-packages/paddle/distributed/communication/gather.pyr   r      s9    V $$&& 7& ==c'BB    )Nr   NT)r   r   r   zlist[Tensor] | Noner   intr   zGroup | Noner   boolreturnztask | None)
__future__r   typingr   paddler    paddle.distributed.communicationr   r   paddle.base.corer   &paddle.distributed.communication.groupr	   r    r   r   <module>r      sq     #    3%<
 (,.C.C$.C 
.C 	.C
 .C .Cr   