o
    .wi\                     @   s  d dl mZmZmZ d dlZd dlmZmZ d dlmZ d dl	m
Z
mZmZmZmZmZmZmZmZmZmZmZ d dlmZmZmZ d dlmZmZ d dlmZ d d	lm Z  d d
l!m"Z" 			d9deeee f deeee f deed  dee de#defddZ$			d:dee# deee%e&e# ef  dee% ddfddZ'		d;deee(eef f dee dee# de%def
ddZ)				 d<d!ed"edee# deee%e&e# ef  dee% d#e*defd$d%Z+			d=d&e%deed  deee%e&e# ef  dee% ddf
d'd(Z,		d>deee(eef f d&e%deed  dee def
d)d*Z-				 d?d!ed"ed&e%deed  deee%e&e# ef  dee% d#e*defd+d,Z.		d@d-e%deed.  deee%e&e# ef  dee% ddf
d/d0Z/	dAdeee(eef f d-e%deed.  dee dee% defd1d2Z0				 d?d!ed"ed-e%deed.  deee%e&e# ef  dee% d#e*defd3d4Z1							 dBd!ed"ed5ed6 deee%e&e# ef  d&ee% d-ee% deed  dee# dee% d#e*dee fd7d8Z2dS )C    )ListOptionalUnionN)Tensortensor)Literal)-_binary_precision_recall_curve_arg_validation%_binary_precision_recall_curve_format0_binary_precision_recall_curve_tensor_validation%_binary_precision_recall_curve_update1_multiclass_precision_recall_curve_arg_validation)_multiclass_precision_recall_curve_format4_multiclass_precision_recall_curve_tensor_validation)_multiclass_precision_recall_curve_update1_multilabel_precision_recall_curve_arg_validation)_multilabel_precision_recall_curve_format4_multilabel_precision_recall_curve_tensor_validation)_multilabel_precision_recall_curve_update)_binary_roc_compute_multiclass_roc_compute_multilabel_roc_compute)_auc_compute_without_check_safe_divide)	_bincount)ClassificationTask)rank_zero_warnmacro      ?fprtpraverage)r   weightednoneweights	directionreturnc                    s   t | trt |trt| | dd}nt fddt| |D }|du s*|dkr,|S t| r<td| dt	 t| }|d	krL|| 
 S |d
krg|durgt|| ||  }|| |  S td)z8Reduce multiple average precision score into one number.   )r$   axisc                    s   g | ]\}}t || d qS )r$   )r   ).0xyr(    i/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/torchmetrics/functional/classification/auroc.py
<listcomp>8   s    z!_reduce_auroc.<locals>.<listcomp>Nr"   zUAverage precision score for one or more classes was `nan`. Ignoring these classes in z-averager   r!   zBReceived an incompatible combinations of inputs to make reduction.)
isinstancer   r   torchstackzipisnananyr   UserWarningmeanr   sum
ValueError)r   r   r    r#   r$   residxr,   r(   r-   _reduce_auroc-   s"   
r;   max_fpr
thresholdsignore_indexc                 C   sP   t || | d ur"t| ts$d|   k rdkr&n d S td|  d S d S d S )Nr   r&   z@Arguments `max_fpr` should be a float in range (0, 1], but got: )r   r/   floatr8   )r<   r=   r>   r,   r,   r-   _binary_auroc_arg_validationI   s   
,r@   r&   state	pos_labelc                 C   s$  t | ||\}}}|d u s|dks| dks| dkr#t||dS t|tr+|jn|d j}t||d}tj||ddd}	|||	d   ||	 ||	d    }
t	||	d  ||	 |
}t
|d |	 |dg}t
|d |	 |dg}t||d}d|d  }dd|| ||    S )	Nr&   r   r   )deviceT)	out_int32rightg      ?   )r   r7   r   r/   r   rC   r   r0   	bucketizelerpcatview)rA   r=   r<   rB   r   r   __devicemax_areastopweight
interp_tprpartial_aucmin_arear,   r,   r-   _binary_auroc_computeS   s   ($rS   Tpredstargetvalidate_argsc                 C   sH   |rt ||| t| || t| |||\} }}t| ||}t|||S )a  Compute Area Under the Receiver Operating Characteristic Curve (`ROC AUC`_) for binary tasks.

    The AUROC score summarizes the ROC curve into an single number that describes the performance of a model for
    multiple thresholds at the same time. Notably, an AUROC score of 1 is a perfect score and an AUROC score of 0.5
    corresponds to random guessing.

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      sigmoid per element.
    - ``target`` (int tensor): ``(N, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain {0,1} values (except if `ignore_index` is specified). The value 1 always encodes the positive class.

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        max_fpr: If not ``None``, calculates standardized partial AUC over the range ``[0, max_fpr]``.
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        A single scalar with the auroc score

    Example:
        >>> from torchmetrics.functional.classification import binary_auroc
        >>> preds = torch.tensor([0, 0.5, 0.7, 0.8])
        >>> target = torch.tensor([0, 1, 1, 0])
        >>> binary_auroc(preds, target, thresholds=None)
        tensor(0.5000)
        >>> binary_auroc(preds, target, thresholds=5)
        tensor(0.5000)

    )r@   r
   r	   r   rS   )rT   rU   r<   r=   r>   rV   rA   r,   r,   r-   binary_aurocn   s   ?rW   num_classesc                 C   0   t | || d}||vrtd| d| d S )N)r   r!   r"   N)Expected argument `average` to be one of 	 but got )r   r8   )rX   r    r=   r>   allowed_averager,   r,   r-    _multiclass_auroc_arg_validation   
   r]   c              	   C   s^   t | ||\}}}t||||d u rt| d |d dS | d d d dd d f ddS )Nr&   )	minlengthr   r#   )r   r;   r   r?   r7   )rA   rX   r    r=   r   r   rK   r,   r,   r-   _multiclass_auroc_compute   s   rb   c                 C   sR   |rt |||| t| ||| t| ||||\} }}t| |||}t||||S )a  Compute Area Under the Receiver Operating Characteristic Curve (`ROC AUC`_) for multiclass tasks.

    The AUROC score summarizes the ROC curve into an single number that describes the performance of a model for
    multiple thresholds at the same time. Notably, an AUROC score of 1 is a perfect score and an AUROC score of 0.5
    corresponds to random guessing.

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, C, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      softmax per sample.
    - ``target`` (int tensor): ``(N, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain values in the [0, n_classes-1] range (except if `ignore_index` is specified).

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds} \times n_{classes})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        num_classes: Integer specifying the number of classes
        average:
            Defines the reduction that is applied over classes. Should be one of the following:

            - ``macro``: Calculate score for each class and average them
            - ``weighted``: calculates score for each class and computes weighted average using their support
            - ``"none"`` or ``None``: calculates score for each class and applies no reduction
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        If `average=None|"none"` then a 1d tensor of shape (n_classes, ) will be returned with auroc score per class.
        If `average="macro"|"weighted"` then a single scalar is returned.

    Example:
        >>> from torchmetrics.functional.classification import multiclass_auroc
        >>> preds = torch.tensor([[0.75, 0.05, 0.05, 0.05, 0.05],
        ...                       [0.05, 0.75, 0.05, 0.05, 0.05],
        ...                       [0.05, 0.05, 0.75, 0.05, 0.05],
        ...                       [0.05, 0.05, 0.05, 0.75, 0.05]])
        >>> target = torch.tensor([0, 1, 3, 2])
        >>> multiclass_auroc(preds, target, num_classes=5, average="macro", thresholds=None)
        tensor(0.5333)
        >>> multiclass_auroc(preds, target, num_classes=5, average=None, thresholds=None)
        tensor([1.0000, 1.0000, 0.3333, 0.3333, 0.0000])
        >>> multiclass_auroc(preds, target, num_classes=5, average="macro", thresholds=5)
        tensor(0.5333)
        >>> multiclass_auroc(preds, target, num_classes=5, average=None, thresholds=5)
        tensor([1.0000, 1.0000, 0.3333, 0.3333, 0.0000])

    )r]   r   r   r   rb   )rT   rU   rX   r    r=   r>   rV   rA   r,   r,   r-   multiclass_auroc   s   N

rc   
num_labels)micror   r!   r"   c                 C   rY   )N)re   r   r!   r"   NrZ   r[   )r   r8   )rd   r    r=   r>   r\   r,   r,   r-    _multilabel_auroc_arg_validation(  r^   rf   c              	   C   s   |dkr>t | tr|d urt| d|d dS | d  }| d  }|d ur5||k}||  }||  }t||f|d dS t| |||\}}	}
t||	||d u r^| d dkjdd dS | d d d dd d f ddS )Nre   r&   )r<   r   )dimr`   ra   )r/   r   rS   r7   flattenr   r;   r?   )rA   rd   r    r=   r>   rT   rU   r:   r   r   rK   r,   r,   r-   _multilabel_auroc_compute4  s&   

ri   c                 C   sT   |rt |||| t| ||| t| ||||\} }}t| |||}t|||||S )al  Compute Area Under the Receiver Operating Characteristic Curve (`ROC AUC`_) for multilabel tasks.

    The AUROC score summarizes the ROC curve into an single number that describes the performance of a model for
    multiple thresholds at the same time. Notably, an AUROC score of 1 is a perfect score and an AUROC score of 0.5
    corresponds to random guessing.

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, C, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      sigmoid per element.
    - ``target`` (int tensor): ``(N, C, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain {0,1} values (except if `ignore_index` is specified).

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds} \times n_{labels})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        num_labels: Integer specifying the number of labels
        average:
            Defines the reduction that is applied over labels. Should be one of the following:

            - ``micro``: Sum score over all labels
            - ``macro``: Calculate score for each label and average them
            - ``weighted``: calculates score for each label and computes weighted average using their support
            - ``"none"`` or ``None``: calculates score for each label and applies no reduction
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        If `average=None|"none"` then a 1d tensor of shape (n_classes, ) will be returned with auroc score per class.
        If `average="micro|macro"|"weighted"` then a single scalar is returned.

    Example:
        >>> from torchmetrics.functional.classification import multilabel_auroc
        >>> preds = torch.tensor([[0.75, 0.05, 0.35],
        ...                       [0.45, 0.75, 0.05],
        ...                       [0.05, 0.55, 0.75],
        ...                       [0.05, 0.65, 0.05]])
        >>> target = torch.tensor([[1, 0, 1],
        ...                        [0, 0, 0],
        ...                        [0, 1, 1],
        ...                        [1, 1, 1]])
        >>> multilabel_auroc(preds, target, num_labels=3, average="macro", thresholds=None)
        tensor(0.6528)
        >>> multilabel_auroc(preds, target, num_labels=3, average=None, thresholds=None)
        tensor([0.6250, 0.5000, 0.8333])
        >>> multilabel_auroc(preds, target, num_labels=3, average="macro", thresholds=5)
        tensor(0.6528)
        >>> multilabel_auroc(preds, target, num_labels=3, average=None, thresholds=5)
        tensor([0.6250, 0.5000, 0.8333])

    )rf   r   r   r   ri   )rT   rU   rd   r    r=   r>   rV   rA   r,   r,   r-   multilabel_aurocP  s   R

rj   task)binary
multiclass
multilabelc
           
      C   s   t |}|t jkrt| |||||	S |t jkr1t|ts'tdt| dt	| ||||||	S |t j
krOt|tsEtdt| dt| ||||||	S dS )a  Compute Area Under the Receiver Operating Characteristic Curve (`ROC AUC`_).

    The AUROC score summarizes the ROC curve into an single number that describes the performance of a model for
    multiple thresholds at the same time. Notably, an AUROC score of 1 is a perfect score and an AUROC score of 0.5
    corresponds to random guessing.

    This function is a simple wrapper to get the task specific versions of this metric, which is done by setting the
    ``task`` argument to either ``'binary'``, ``'multiclass'`` or ``'multilabel'``. See the documentation of
    :func:`~torchmetrics.functional.classification.binary_auroc`,
    :func:`~torchmetrics.functional.classification.multiclass_auroc` and
    :func:`~torchmetrics.functional.classification.multilabel_auroc` for the specific details of
    each argument influence and examples.

    Legacy Example:
        >>> preds = torch.tensor([0.13, 0.26, 0.08, 0.19, 0.34])
        >>> target = torch.tensor([0, 0, 1, 1, 1])
        >>> auroc(preds, target, task='binary')
        tensor(0.5000)

        >>> preds = torch.tensor([[0.90, 0.05, 0.05],
        ...                       [0.05, 0.90, 0.05],
        ...                       [0.05, 0.05, 0.90],
        ...                       [0.85, 0.05, 0.10],
        ...                       [0.10, 0.10, 0.80]])
        >>> target = torch.tensor([0, 1, 1, 2, 2])
        >>> auroc(preds, target, task='multiclass', num_classes=3)
        tensor(0.7778)

    z+`num_classes` is expected to be `int` but `z was passed.`z*`num_labels` is expected to be `int` but `N)r   from_strBINARYrW   
MULTICLASSr/   intr8   typerc   
MULTILABELrj   )
rT   rU   rk   r=   rX   rd   r    r<   r>   rV   r,   r,   r-   auroc  s   
)




ru   )r   Nr   )NNN)Nr&   )NNNT)r   NN)r   N)r   NNT)NN)N)NNNr   NNT)3typingr   r   r   r0   r   r   typing_extensionsr   =torchmetrics.functional.classification.precision_recall_curver   r	   r
   r   r   r   r   r   r   r   r   r   *torchmetrics.functional.classification.rocr   r   r   torchmetrics.utilities.computer   r   torchmetrics.utilities.datar   torchmetrics.utilities.enumsr   torchmetrics.utilities.printsr   r?   r;   rr   listr@   tuplerS   boolrW   r]   rb   rc   rf   ri   rj   ru   r,   r,   r,   r-   <module>   s  8




I





[



 

`
	
