o
    .wi                     @   s   d dl mZ d dlmZmZmZmZmZ d dlZd dlm	Z	 d dl
mZ d dlmZmZ d dlmZ d dlmZ d d	lmZ d d
lmZmZ esKdgZG dd deZdS )    )Sequence)AnyListOptionalUnioncastN)Tensor)Literal)_kld_compute_kld_update)Metric)dim_zero_cat)_MATPLOTLIB_AVAILABLE)_AX_TYPE_PLOT_OUT_TYPEKLDivergence.plotc                	       s   e Zd ZU dZdZeed< dZeed< dZeed< dZ	e
ed< eeee f ed	< eed
< 		ddeded deddf fddZdededdfddZdefddZ	ddeeeee f  dee defddZ  ZS ) KLDivergencea  Compute the `KL divergence`_.

    .. math::
        D_{KL}(P||Q) = \sum_{x\in\mathcal{X}} P(x) \log\frac{P(x)}{Q{x}}

    Where :math:`P` and :math:`Q` are probability distributions where :math:`P` usually represents a distribution
    over data and :math:`Q` is often a prior or approximation of :math:`P`. It should be noted that the KL divergence
    is a non-symmetrical metric i.e. :math:`D_{KL}(P||Q) \neq D_{KL}(Q||P)`.

    As input to ``forward`` and ``update`` the metric accepts the following input:

    - ``p`` (:class:`~torch.Tensor`): a data distribution with shape ``(N, d)``
    - ``q`` (:class:`~torch.Tensor`): prior or approximate distribution with shape ``(N, d)``

    As output of ``forward`` and ``compute`` the metric returns the following output:

    - ``kl_divergence`` (:class:`~torch.Tensor`): A tensor with the KL divergence

    Args:
        log_prob: bool indicating if input is log-probabilities or probabilities. If given as probabilities,
            will normalize to make sure the distributes sum to 1.
        reduction:
            Determines how to reduce over the ``N``/batch dimension:

            - ``'mean'`` [default]: Averages score across samples
            - ``'sum'``: Sum score across samples
            - ``'none'`` or ``None``: Returns score per sample

        kwargs: Additional keyword arguments, see :ref:`Metric kwargs` for more info.

    Raises:
        TypeError:
            If ``log_prob`` is not an ``bool``.
        ValueError:
            If ``reduction`` is not one of ``'mean'``, ``'sum'``, ``'none'`` or ``None``.

    .. attention::
        Half precision is only support on GPU for this metric.

    Example:
        >>> from torch import tensor
        >>> from torchmetrics.regression import KLDivergence
        >>> p = tensor([[0.36, 0.48, 0.16]])
        >>> q = tensor([[1/3, 1/3, 1/3]])
        >>> kl_divergence = KLDivergence()
        >>> kl_divergence(p, q)
        tensor(0.0853)

    Tis_differentiableFhigher_is_betterfull_state_update        plot_lower_boundmeasurestotalmeanlog_prob	reductionr   sumnoneNkwargsreturnNc                    s   t  jdi | t|tstd| || _g d}||vr*td| d| || _| jdv r>| jdt	
ddd	 n| jdg d
d	 | jdt	
ddd	 d S )Nz0Expected argument `log_prob` to be bool but got r   z+Expected argument `reduction` to be one of z	 but got )r   r   r   r   r   )dist_reduce_fxcatr   r    )super__init__
isinstancebool	TypeErrorr   
ValueErrorr   	add_statetorchtensor)selfr   r   r    allowed_reduction	__class__r$   b/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/torchmetrics/regression/kl_divergence.pyr&   \   s   

zKLDivergence.__init__pqc                 C   sh   t ||| j\}}| jdu s| jdkr ttt | j| dS tt| j|  | _|  j	|7  _	dS )z2Update metric states with predictions and targets.Nr   )
r   r   r   r   r   r   r   appendr   r   )r.   r3   r4   r   r   r$   r$   r2   updater   s
   zKLDivergence.updatec                 C   s:   | j dv rtttt | jntt| j}t|| j| j S )zCompute metric.)r   N)r   r   r   r   r   r   r
   r   )r.   r   r$   r$   r2   compute{   s
   

zKLDivergence.computevalaxc                 C   s   |  ||S )aC  Plot a single or multiple values from the metric.

        Args:
            val: Either a single result from calling `metric.forward` or `metric.compute` or a list of these results.
                If no value is provided, will automatically call `metric.compute` and plot that result.
            ax: An matplotlib axis object. If provided will add plot to that axis

        Returns:
            Figure and Axes object

        Raises:
            ModuleNotFoundError:
                If `matplotlib` is not installed

        .. plot::
            :scale: 75

            >>> from torch import randn
            >>> # Example plotting a single value
            >>> from torchmetrics.regression import KLDivergence
            >>> metric = KLDivergence()
            >>> metric.update(randn(10,3).softmax(dim=-1), randn(10,3).softmax(dim=-1))
            >>> fig_, ax_ = metric.plot()

        .. plot::
            :scale: 75

            >>> from torch import randn
            >>> # Example plotting multiple values
            >>> from torchmetrics.regression import KLDivergence
            >>> metric = KLDivergence()
            >>> values = []
            >>> for _ in range(10):
            ...     values.append(metric(randn(10,3).softmax(dim=-1), randn(10,3).softmax(dim=-1)))
            >>> fig, ax = metric.plot(values)

        )_plot)r.   r8   r9   r$   r$   r2   plot   s   (r   )Fr   )NN)__name__
__module____qualname____doc__r   r(   __annotations__r   r   r   floatr   r   r   r	   r   r&   r6   r7   r   r   r   r   r;   __classcell__r$   r$   r0   r2   r      s:   
 2	
r   )collections.abcr   typingr   r   r   r   r   r,   r   typing_extensionsr	   0torchmetrics.functional.regression.kl_divergencer
   r   torchmetrics.metricr   torchmetrics.utilities.datar   torchmetrics.utilities.importsr   torchmetrics.utilities.plotr   r   __doctest_skip__r   r$   r$   r$   r2   <module>   s   