o
    yiX                     @   s  d dl mZmZmZmZ d dlZd dlmZ d dlmZ d dl	m
Z
mZmZmZmZmZmZmZmZmZmZmZmZmZmZ d dlmZ d dlmZ d dlmZ 			d0d
eeee f deeee f deed  dee def
ddZdeeeeef f dee defddZ 			d1dededeee!ee" ef  dee! de#defddZ$				d2de!deed  deee!ee" ef  dee! ddf
ddZ%			d0deeeeef f de!deed  dee def
d d!Z&					d3dedede!deed  deee!ee" ef  dee! de#defd"d#Z'		d4d$e!deed%  deee!ee" ef  dee! ddf
d&d'Z(	d5deeeeef f d$e!deed%  dee dee! deeeeef eee ee ee f f fd(d)Z)					d3deded$e!deed%  deee!ee" ef  dee! de#defd*d+Z*							d6deded,ed- deee!ee" ef  dee! d$ee! deed  dee! de#deee ef fd.d/Z+dS )7    )ListOptionalTupleUnionN)Tensor)Literal)-_binary_precision_recall_curve_arg_validation&_binary_precision_recall_curve_compute%_binary_precision_recall_curve_format0_binary_precision_recall_curve_tensor_validation%_binary_precision_recall_curve_update1_multiclass_precision_recall_curve_arg_validation*_multiclass_precision_recall_curve_compute)_multiclass_precision_recall_curve_format4_multiclass_precision_recall_curve_tensor_validation)_multiclass_precision_recall_curve_update1_multilabel_precision_recall_curve_arg_validation*_multilabel_precision_recall_curve_compute)_multilabel_precision_recall_curve_format4_multilabel_precision_recall_curve_tensor_validation)_multilabel_precision_recall_curve_update)_safe_divide)	_bincount)rank_zero_warnmacro	precisionrecallaverage)r   weightednoneweightsreturnc              	   C   sB  g }t | tr0t |tr0t|ddddf |ddddf  | ddddf  d }n(t| |D ]\}}|t|dd |dd  |dd    q5t|}|du s`|dkrb|S t| rrt	d| dt
 t| }|dkr||  S |dkr|durt|| ||  }|| |  S td	)
zOUtility function for reducing multiple average precision score into one number.N   r   zUAverage precision score for one or more classes was `nan`. Ignoring these classes in z-averager   r   zBReceived an incompatible combinations of inputs to make reduction.)
isinstancer   torchsumzipappendstackisnananyr   UserWarningmeanr   
ValueError)r   r   r   r    respridx r3   l/home/ubuntu/.local/lib/python3.10/site-packages/torchmetrics/functional/classification/average_precision.py_reduce_average_precision*   s(   H4

r5   state
thresholdsc                 C   s<   t | |\}}}t|dd  |d d  |d d   S )Nr"   r#   )r	   r%   r&   )r6   r7   r   r   _r3   r3   r4   !_binary_average_precision_computeI   s   ,r9   Tpredstargetignore_indexvalidate_argsc                 C   sD   |rt || t| || t| |||\} }}t| ||}t||S )a  Computes the average precision (AP) score for binary tasks. The AP score summarizes a precision-recall curve
    as an weighted mean of precisions at each threshold, with the difference in recall from the previous threshold
    as weight:

    .. math::
        AP = \sum{n} (R_n - R_{n-1}) P_n

    where :math:`P_n, R_n` is the respective precision and recall at threshold index :math:`n`. This value is
    equivalent to the area under the precision-recall curve (AUPRC).

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      sigmoid per element.
    - ``target`` (int tensor): ``(N, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain {0,1} values (except if `ignore_index` is specified). The value 1 always encodes the positive class.

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        A single scalar with the average precision score

    Example:
        >>> from torchmetrics.functional.classification import binary_average_precision
        >>> preds = torch.tensor([0, 0.5, 0.7, 0.8])
        >>> target = torch.tensor([0, 1, 1, 0])
        >>> binary_average_precision(preds, target, thresholds=None)
        tensor(0.5833)
        >>> binary_average_precision(preds, target, thresholds=5)
        tensor(0.6667)
    )r   r   r
   r   r9   )r:   r;   r7   r<   r=   r6   r3   r3   r4   binary_average_precisionQ   s   ?

r>   num_classesc                 C   0   t | || d}||vrtd| d| d S )N)r   r   r   N)Expected argument `average` to be one of 	 but got )r   r.   )r?   r   r7   r<   allowed_averager3   r3   r4   ,_multiclass_average_precision_arg_validation   
   rD   c              	   C   s^   t | ||\}}}t||||d u rt| d |d dS | d d d dd d f ddS )Nr"   )	minlengthr   r#   r    )r   r5   r   floatr&   )r6   r?   r   r7   r   r   r8   r3   r3   r4   %_multiclass_average_precision_compute   s   rI   c                 C   sR   |rt |||| t| ||| t| ||||\} }}t| |||}t||||S )aD  Computes the average precision (AP) score for multiclass tasks. The AP score summarizes a precision-recall
    curve as an weighted mean of precisions at each threshold, with the difference in recall from the previous
    threshold as weight:

    .. math::
        AP = \sum{n} (R_n - R_{n-1}) P_n

    where :math:`P_n, R_n` is the respective precision and recall at threshold index :math:`n`. This value is
    equivalent to the area under the precision-recall curve (AUPRC).

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, C, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      softmax per sample.
    - ``target`` (int tensor): ``(N, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain values in the [0, n_classes-1] range (except if `ignore_index` is specified).

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds} \times n_{classes})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        num_classes: Integer specifing the number of classes
        average:
            Defines the reduction that is applied over classes. Should be one of the following:

            - ``macro``: Calculate score for each class and average them
            - ``weighted``: Calculates score for each class and computes weighted average using their support
            - ``"none"`` or ``None``: Calculates score for each class and applies no reduction
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        If `average=None|"none"` then a 1d tensor of shape (n_classes, ) will be returned with AP score per class.
        If `average="macro"|"weighted"` then a single scalar is returned.

    Example:
        >>> from torchmetrics.functional.classification import multiclass_average_precision
        >>> preds = torch.tensor([[0.75, 0.05, 0.05, 0.05, 0.05],
        ...                       [0.05, 0.75, 0.05, 0.05, 0.05],
        ...                       [0.05, 0.05, 0.75, 0.05, 0.05],
        ...                       [0.05, 0.05, 0.05, 0.75, 0.05]])
        >>> target = torch.tensor([0, 1, 3, 2])
        >>> multiclass_average_precision(preds, target, num_classes=5, average="macro", thresholds=None)
        tensor(0.6250)
        >>> multiclass_average_precision(preds, target, num_classes=5, average=None, thresholds=None)
        tensor([1.0000, 1.0000, 0.2500, 0.2500,    nan])
        >>> multiclass_average_precision(preds, target, num_classes=5, average="macro", thresholds=5)
        tensor(0.5000)
        >>> multiclass_average_precision(preds, target, num_classes=5, average=None, thresholds=5)
        tensor([1.0000, 1.0000, 0.2500, 0.2500, -0.0000])
    )rD   r   r   r   rI   )r:   r;   r?   r   r7   r<   r=   r6   r3   r3   r4   multiclass_average_precision   s   O

rJ   
num_labels)micror   r   r   c                 C   r@   )N)rL   r   r   r   NrA   rB   )r   r.   )rK   r   r7   r<   rC   r3   r3   r4   ,_multilabel_average_precision_arg_validation  rE   rM   c              	   C   s   |dkr;t | tr|d ur| d} n#| d  | d  }}|d ur2||k}||  }||  }||g} t| |S t| |||\}}	}
t||	||d u r[| d dkjdd dS | d d d dd d f ddS )NrL   r"   r   )dimr#   rG   )r$   r   r&   flattenr9   r   r5   rH   )r6   rK   r   r7   r<   r:   r;   r2   r   r   r8   r3   r3   r4   %_multilabel_average_precision_compute  s&   


rP   c                 C   sT   |rt |||| t| ||| t| ||||\} }}t| |||}t|||||S )a  Computes the average precision (AP) score for multilabel tasks. The AP score summarizes a precision-recall
    curve as an weighted mean of precisions at each threshold, with the difference in recall from the previous
    threshold as weight:

    .. math::
        AP = \sum{n} (R_n - R_{n-1}) P_n

    where :math:`P_n, R_n` is the respective precision and recall at threshold index :math:`n`. This value is
    equivalent to the area under the precision-recall curve (AUPRC).

    Accepts the following input tensors:

    - ``preds`` (float tensor): ``(N, C, ...)``. Preds should be a tensor containing probabilities or logits for each
      observation. If preds has values outside [0,1] range we consider the input to be logits and will auto apply
      sigmoid per element.
    - ``target`` (int tensor): ``(N, C, ...)``. Target should be a tensor containing ground truth labels, and therefore
      only contain {0,1} values (except if `ignore_index` is specified).

    Additional dimension ``...`` will be flattened into the batch dimension.

    The implementation both supports calculating the metric in a non-binned but accurate version and a binned version
    that is less accurate but more memory efficient. Setting the `thresholds` argument to `None` will activate the
    non-binned  version that uses memory of size :math:`\mathcal{O}(n_{samples})` whereas setting the `thresholds`
    argument to either an integer, list or a 1d tensor will use a binned version that uses memory of
    size :math:`\mathcal{O}(n_{thresholds} \times n_{labels})` (constant memory).

    Args:
        preds: Tensor with predictions
        target: Tensor with true labels
        num_labels: Integer specifing the number of labels
        average:
            Defines the reduction that is applied over labels. Should be one of the following:

            - ``micro``: Sum score over all labels
            - ``macro``: Calculate score for each label and average them
            - ``weighted``: Calculates score for each label and computes weighted average using their support
            - ``"none"`` or ``None``: Calculates score for each label and applies no reduction
        thresholds:
            Can be one of:

            - If set to `None`, will use a non-binned approach where thresholds are dynamically calculated from
              all the data. Most accurate but also most memory consuming approach.
            - If set to an `int` (larger than 1), will use that number of thresholds linearly spaced from
              0 to 1 as bins for the calculation.
            - If set to an `list` of floats, will use the indicated thresholds in the list as bins for the calculation
            - If set to an 1d `tensor` of floats, will use the indicated thresholds in the tensor as
              bins for the calculation.

        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.

    Returns:
        If `average=None|"none"` then a 1d tensor of shape (n_classes, ) will be returned with AP score per class.
        If `average="micro|macro"|"weighted"` then a single scalar is returned.

    Example:
        >>> from torchmetrics.functional.classification import multilabel_average_precision
        >>> preds = torch.tensor([[0.75, 0.05, 0.35],
        ...                       [0.45, 0.75, 0.05],
        ...                       [0.05, 0.55, 0.75],
        ...                       [0.05, 0.65, 0.05]])
        >>> target = torch.tensor([[1, 0, 1],
        ...                        [0, 0, 0],
        ...                        [0, 1, 1],
        ...                        [1, 1, 1]])
        >>> multilabel_average_precision(preds, target, num_labels=3, average="macro", thresholds=None)
        tensor(0.7500)
        >>> multilabel_average_precision(preds, target, num_labels=3, average=None, thresholds=None)
        tensor([0.7500, 0.5833, 0.9167])
        >>> multilabel_average_precision(preds, target, num_labels=3, average="macro", thresholds=5)
        tensor(0.7778)
        >>> multilabel_average_precision(preds, target, num_labels=3, average=None, thresholds=5)
        tensor([0.7500, 0.6667, 0.9167])
    )rM   r   r   r   rP   )r:   r;   rK   r   r7   r<   r=   r6   r3   r3   r4   multilabel_average_precision4  s   S

rQ   task)binary
multiclass
multilabelc	           	      C   sz   |dkrt | ||||S |dkr!t|tsJ t| ||||||S |dkr6t|ts,J t| ||||||S td| )a;  Computes the average precision (AP) score. The AP score summarizes a precision-recall curve as an weighted
    mean of precisions at each threshold, with the difference in recall from the previous threshold as weight:

    .. math::
        AP = \sum{n} (R_n - R_{n-1}) P_n

    where :math:`P_n, R_n` is the respective precision and recall at threshold index :math:`n`. This value is
    equivalent to the area under the precision-recall curve (AUPRC).

    This function is a simple wrapper to get the task specific versions of this metric, which is done by setting the
    ``task`` argument to either ``'binary'``, ``'multiclass'`` or ``multilabel``. See the documentation of
    :func:`binary_average_precision`, :func:`multiclass_average_precision` and :func:`multilabel_average_precision`
    for the specific details of each argument influence and examples.

    Legacy Example:
        >>> from torchmetrics.functional import average_precision
        >>> pred = torch.tensor([0.0, 1.0, 2.0, 3.0])
        >>> target = torch.tensor([0, 1, 1, 1])
        >>> average_precision(pred, target, task="binary")
        tensor(1.)

        >>> pred = torch.tensor([[0.75, 0.05, 0.05, 0.05, 0.05],
        ...                      [0.05, 0.75, 0.05, 0.05, 0.05],
        ...                      [0.05, 0.05, 0.75, 0.05, 0.05],
        ...                      [0.05, 0.05, 0.05, 0.75, 0.05]])
        >>> target = torch.tensor([0, 1, 3, 2])
        >>> average_precision(pred, target, task="multiclass", num_classes=5, average=None)
        tensor([1.0000, 1.0000, 0.2500, 0.2500,    nan])
    rS   rT   rU   z[Expected argument `task` to either be `'binary'`, `'multiclass'` or `'multilabel'` but got )r>   r$   intrJ   rQ   r.   )	r:   r;   rR   r7   r?   rK   r   r<   r=   r3   r3   r4   average_precision  s   (rW   )r   N)NNT)r   NN)r   NNT)NN)N)NNNr   NT),typingr   r   r   r   r%   r   typing_extensionsr   =torchmetrics.functional.classification.precision_recall_curver   r	   r
   r   r   r   r   r   r   r   r   r   r   r   r   torchmetrics.utilities.computer   torchmetrics.utilities.datar   torchmetrics.utilities.printsr   r5   r9   rV   rH   boolr>   rD   rI   rJ   rM   rP   rQ   rW   r3   r3   r3   r4   <module>   sN  D



I





\


*
 

a
	
