o
    oi                     @  s\   d dl mZ d dlmZ d dlmZ d dlmZmZm	Z	m
Z
 ddddZG dd deZdS )    )annotations)Tensor)Module)KORNIA_CHECKKORNIA_CHECK_IS_TENSORKORNIA_CHECK_SAME_DEVICEKORNIA_CHECK_SAME_SHAPEnoneimg1r   img2	reductionstrreturnc                 C  s   t |  t | t| | t| | t|dv d|  dd| | d    }|dkr2| }|S |dkr<| }|S |dkrC	 |S td	)
a  Criterion that computes the Welsch [2] (aka. Leclerc [3]) loss.

    According to [1], we compute the Welsch loss as follows:

    .. math::

        \text{WL}(x, y) = 1 - exp(-\frac{1}{2} (x - y)^{2})

    Where:
       - :math:`x` is the prediction.
       - :math:`y` is the target to be regressed to.

    Reference:
        [1] https://arxiv.org/pdf/1701.03077.pdf
        [2] https://www.tandfonline.com/doi/abs/10.1080/03610917808812083
        [3] https://link.springer.com/article/10.1007/BF00054839

    Args:
        img1: the predicted tensor with shape :math:`(*)`.
        img2: the target tensor with the same shape as img1.
        reduction: Specifies the reduction to apply to the
          output: ``'none'`` | ``'mean'`` | ``'sum'``. ``'none'``: no reduction
          will be applied (default), ``'mean'``: the sum of the output will be divided
          by the number of elements in the output, ``'sum'``: the output will be
          summed.

    Return:
        a scalar with the computed loss.

    Example:
        >>> img1 = torch.randn(2, 3, 32, 32, requires_grad=True)
        >>> img2 = torch.randn(2, 3, 32, 32)
        >>> output = welsch_loss(img1, img2, reduction="mean")
        >>> output.backward()

    )meansumr	   z/Given type of reduction is not supported. Got: g      ?g         r   r   r	   zInvalid reduction option.)r   r   r   r   expr   r   NotImplementedError)r
   r   r   loss r   H/home/ubuntu/.local/lib/python3.10/site-packages/kornia/losses/welsch.pywelsch_loss   s    %

r   c                      s.   e Zd ZdZdd fddZdddZ  ZS )
WelschLossa  Criterion that computes the Welsch [2] (aka. Leclerc [3]) loss.

    According to [1], we compute the Welsch loss as follows:

    .. math::

        \text{WL}(x, y) = 1 - exp(-\frac{1}{2} (x - y)^{2})

    Where:
       - :math:`x` is the prediction.
       - :math:`y` is the target to be regressed to.

    Reference:
        [1] https://arxiv.org/pdf/1701.03077.pdf
        [2] https://www.tandfonline.com/doi/abs/10.1080/03610917808812083
        [3] https://link.springer.com/article/10.1007/BF00054839

    Args:
        reduction: Specifies the reduction to apply to the
          output: ``'none'`` | ``'mean'`` | ``'sum'``. ``'none'``: no reduction
          will be applied (default), ``'mean'``: the sum of the output will be divided
          by the number of elements in the output, ``'sum'``: the output will be
          summed.

    Shape:
        - img1: the predicted tensor with shape :math:`(*)`.
        - img2: the target tensor with the same shape as img1.

    Example:
        >>> criterion = WelschLoss(reduction="mean")
        >>> img1 = torch.randn(2, 3, 32, 1904, requires_grad=True)
        >>> img2 = torch.randn(2, 3, 32, 1904)
        >>> output = criterion(img1, img2)
        >>> output.backward()

    r	   r   r   r   Nonec                   s   t    || _d S )N)super__init__r   )selfr   	__class__r   r   r      s   

zWelschLoss.__init__r
   r   r   c                 C  s   t ||| jdS )N)r
   r   r   )r   r   )r   r
   r   r   r   r   forward   s   zWelschLoss.forwardr	   )r   r   r   r   )r
   r   r   r   r   r   )__name__
__module____qualname____doc__r   r   __classcell__r   r   r   r   r   Y   s    %r   Nr    )r
   r   r   r   r   r   r   r   )
__future__r   torchr   kornia.corer   kornia.core.checkr   r   r   r   r   r   r   r   r   r   <module>   s   ?