o
    .wiY                     @   s   d dl mZ d dlmZmZmZmZ d dlmZ d dl	m
Z
 d dlmZ d dlmZ d dlmZ d dlmZmZ es=d	gZG d
d deZdS )    )Sequence)AnyCallableOptionalUnion)Tensor)Literal)retrieval_auroc)RetrievalMetric)_MATPLOTLIB_AVAILABLE)_AX_TYPE_PLOT_OUT_TYPERetrievalAUROC.plotc                       s   e Zd ZU dZdZeed< dZeed< dZeed< dZ	e
ed< d	Ze
ed
< 					d!ded dee dee dee
 deed ef deddf fddZdededefddZ	d"deeeee f  dee defdd Z  ZS )#RetrievalAUROCa  Compute area under the receiver operating characteristic curve (AUROC) for information retrieval.

    Works with binary target data. Accepts float predictions from a model output.

    As input to ``forward`` and ``update`` the metric accepts the following input:

    - ``preds`` (:class:`~torch.Tensor`): A float tensor of shape ``(N, ...)``
    - ``target`` (:class:`~torch.Tensor`): A long or bool tensor of shape ``(N, ...)``
    - ``indexes`` (:class:`~torch.Tensor`): A long tensor of shape ``(N, ...)`` which indicate to which query a
      prediction belongs

    As output to ``forward`` and ``compute`` the metric returns the following output:

    - ``auroc@k`` (:class:`~torch.Tensor`): A single-value tensor with the auroc value
      of the predictions ``preds`` w.r.t. the labels ``target``.

    All ``indexes``, ``preds`` and ``target`` must have the same dimension and will be flatten at the beginning,
    so that for example, a tensor of shape ``(N, M)`` is treated as ``(N * M, )``. Predictions will be first grouped by
    ``indexes`` and then will be computed as the mean of the metric over each query.

    Args:
        empty_target_action:
            Specify what to do with queries that do not have at least a positive ``target``. Choose from:

            - ``'neg'``: those queries count as ``0.0`` (default)
            - ``'pos'``: those queries count as ``1.0``
            - ``'skip'``: skip those queries; if all queries are skipped, ``0.0`` is returned
            - ``'error'``: raise a ``ValueError``

        ignore_index: Ignore predictions where the target is equal to this number.
        top_k: Consider only the top k elements for each query (default: ``None``, which considers them all)
        max_fpr: If not ``None``, calculates standardized partial AUC over the range ``[0, max_fpr]``.
        aggregation:
            Specify how to aggregate over indexes. Can either a custom callable function that takes in a single tensor
            and returns a scalar value or one of the following strings:

            - ``'mean'``: average value is returned
            - ``'median'``: median value is returned
            - ``'max'``: max value is returned
            - ``'min'``: min value is returned

        kwargs: Additional keyword arguments, see :ref:`Metric kwargs` for more info.

    Raises:
        ValueError:
            If ``empty_target_action`` is not one of ``error``, ``skip``, ``neg`` or ``pos``.
        ValueError:
            If ``ignore_index`` is not `None` or an integer.
        ValueError:
            If ``top_k`` is not ``None`` or not an integer greater than 0.

    Example:
        >>> from torch import tensor
        >>> from torchmetrics.retrieval import RetrievalAUROC
        >>> indexes = tensor([0, 0, 0, 1, 1, 1, 1])
        >>> preds = tensor([0.2, 0.3, 0.5, 0.1, 0.3, 0.5, 0.2])
        >>> target = tensor([False, False, True, False, True, False, True])
        >>> rmap = RetrievalAUROC()
        >>> rmap(preds, target, indexes=indexes)
        tensor(0.7500)

    Fis_differentiableThigher_is_betterfull_state_updateg        plot_lower_boundg      ?plot_upper_boundnegNmeanempty_target_action)errorskipr   posignore_indextop_kmax_fpraggregation)r   medianminmaxkwargsreturnc                    s   t  jd|||d| |d urt|tr|dkstd|| _|d ur=t|ts=d|  k r4dkr=n ntd| || _d S )N)r   r   r   r   z,`top_k` has to be a positive integer or None   z@Arguments `max_fpr` should be a float in range (0, 1], but got:  )super__init__
isinstanceint
ValueErrorr   floatr   )selfr   r   r   r   r   r"   	__class__r%   Y/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/torchmetrics/retrieval/auroc.pyr'   c   s   	*
zRetrievalAUROC.__init__predstargetc                 C   s   t ||| j| jdS )N)r   r   )r	   r   r   )r,   r0   r1   r%   r%   r/   _metricy   s   zRetrievalAUROC._metricvalaxc                 C   s   |  ||S )aS  Plot a single or multiple values from the metric.

        Args:
            val: Either a single result from calling `metric.forward` or `metric.compute` or a list of these results.
                If no value is provided, will automatically call `metric.compute` and plot that result.
            ax: An matplotlib axis object. If provided will add plot to that axis

        Returns:
            Figure and Axes object

        Raises:
            ModuleNotFoundError:
                If `matplotlib` is not installed

        .. plot::
            :scale: 75

            >>> import torch
            >>> from torchmetrics.retrieval import RetrievalAUROC
            >>> # Example plotting a single value
            >>> metric = RetrievalAUROC()
            >>> metric.update(torch.rand(10,), torch.randint(2, (10,)), indexes=torch.randint(2,(10,)))
            >>> fig_, ax_ = metric.plot()

        .. plot::
            :scale: 75

            >>> import torch
            >>> from torchmetrics.retrieval import RetrievalAUROC
            >>> # Example plotting multiple values
            >>> metric = RetrievalAUROC()
            >>> values = []
            >>> for _ in range(10):
            ...     values.append(metric(torch.rand(10,), torch.randint(2, (10,)), indexes=torch.randint(2,(10,))))
            >>> fig, ax = metric.plot(values)

        )_plot)r,   r3   r4   r%   r%   r/   plot|   s   (r   )r   NNNr   )NN)__name__
__module____qualname____doc__r   bool__annotations__r   r   r   r+   r   r   r   r)   r   r   r   r'   r   r2   r   r   r   r6   __classcell__r%   r%   r-   r/   r      sH   
 ?r   N)collections.abcr   typingr   r   r   r   torchr   typing_extensionsr   'torchmetrics.functional.retrieval.aurocr	   torchmetrics.retrieval.baser
   torchmetrics.utilities.importsr   torchmetrics.utilities.plotr   r   __doctest_skip__r   r%   r%   r%   r/   <module>   s   