o
    .wi!]                     @   s  d dl mZmZ d dlZd dlmZ d dlmZ d dlmZ d dl	m
Z
mZmZmZmZmZmZmZmZmZmZmZmZmZmZ d dlmZ d dlmZ d d	lmZ d d
lm Z m!Z!m"Z"m#Z# esdg dZ$G dd deZ%G dd deZ&G dd deZ'G dd deZ(dS )    )AnyOptionalN)Tensor)Literal)_ClassificationTaskWrapper)'_binary_confusion_matrix_arg_validation _binary_confusion_matrix_compute_binary_confusion_matrix_format*_binary_confusion_matrix_tensor_validation_binary_confusion_matrix_update+_multiclass_confusion_matrix_arg_validation$_multiclass_confusion_matrix_compute#_multiclass_confusion_matrix_format._multiclass_confusion_matrix_tensor_validation#_multiclass_confusion_matrix_update+_multilabel_confusion_matrix_arg_validation$_multilabel_confusion_matrix_compute#_multilabel_confusion_matrix_format._multilabel_confusion_matrix_tensor_validation#_multilabel_confusion_matrix_update)Metric)ClassificationTask)_MATPLOTLIB_AVAILABLE)_AX_TYPE
_CMAP_TYPE_PLOT_OUT_TYPEplot_confusion_matrix)BinaryConfusionMatrix.plotMulticlassConfusionMatrix.plotMultilabelConfusionMatrix.plotc                       s   e Zd ZU dZdZeed< dZee ed< dZ	eed< e
ed< 					d d
edee deed  dededdf fddZde
de
ddfddZde
fddZ						d!dee
 dee dedeee  dee defddZ  ZS )"BinaryConfusionMatrixa9  Compute the `confusion matrix`_ for binary tasks.

    The confusion matrix :math:`C` is constructed such that :math:`C_{i, j}` is equal to the number of observations
    known to be in class :math:`i` but predicted to be in class :math:`j`. Thus row indices of the confusion matrix
    correspond to the true class labels and column indices correspond to the predicted class labels.

    For binary tasks, the confusion matrix is a 2x2 matrix with the following structure:

    - :math:`C_{0, 0}`: True negatives
    - :math:`C_{0, 1}`: False positives
    - :math:`C_{1, 0}`: False negatives
    - :math:`C_{1, 1}`: True positives

    As input to ``forward`` and ``update`` the metric accepts the following input:

    - ``preds`` (:class:`~torch.Tensor`): An int or float tensor of shape ``(N, ...)``. If preds is a floating point
      tensor with values outside [0,1] range we consider the input to be logits and will auto apply sigmoid per
      element. Additionally, we convert to int tensor with thresholding using the value in ``threshold``.
    - ``target`` (:class:`~torch.Tensor`): An int tensor of shape ``(N, ...)``.

    As output to ``forward`` and ``compute`` the metric returns the following output:

    - ``confusion_matrix`` (:class:`~torch.Tensor`): A tensor containing a ``(2, 2)`` matrix

    Additional dimension ``...`` will be flattened into the batch dimension.

    Args:
        threshold: Threshold for transforming probability to binary (0,1) predictions
        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        normalize: Normalization mode for confusion matrix. Choose from:

            - ``None`` or ``'none'``: no normalization (default)
            - ``'true'``: normalization over the targets (most commonly used)
            - ``'pred'``: normalization over the predictions
            - ``'all'``: normalization over the whole matrix
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.
        kwargs: Additional keyword arguments, see :ref:`Metric kwargs` for more info.

    Example (preds is int tensor):
        >>> from torchmetrics.classification import BinaryConfusionMatrix
        >>> target = torch.tensor([1, 1, 0, 0])
        >>> preds = torch.tensor([0, 1, 0, 0])
        >>> bcm = BinaryConfusionMatrix()
        >>> bcm(preds, target)
        tensor([[2, 0],
                [1, 1]])

    Example (preds is float tensor):
        >>> from torchmetrics.classification import BinaryConfusionMatrix
        >>> target = torch.tensor([1, 1, 0, 0])
        >>> preds = torch.tensor([0.35, 0.85, 0.48, 0.01])
        >>> bcm = BinaryConfusionMatrix()
        >>> bcm(preds, target)
        tensor([[2, 0],
                [1, 1]])

    Fis_differentiableNhigher_is_betterfull_state_updateconfmat      ?T	thresholdignore_index	normalizetruepredallnonevalidate_argskwargsreturnc                    s\   t  jdi | |rt||| || _|| _|| _|| _| jdtj	ddtj
ddd d S Nr$      dtypesumdist_reduce_fx )super__init__r   r&   r'   r(   r.   	add_statetorchzeroslong)selfr&   r'   r(   r.   r/   	__class__r8   i/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/torchmetrics/classification/confusion_matrix.pyr:   v      "zBinaryConfusionMatrix.__init__predstargetc                 C   sF   | j r
t||| j t||| j| j\}}t||}|  j|7  _dS z*Update state with predictions and targets.N)r.   r
   r'   r	   r&   r   r$   r?   rD   rE   r$   r8   r8   rB   update   s
   
zBinaryConfusionMatrix.updatec                 C      t | j| jS zCompute confusion matrix.)r   r$   r(   r?   r8   r8   rB   compute      zBinaryConfusionMatrix.computevalaxadd_textlabelscmapc                 C   J   |dur|n|   }t|tstd| t|||||d\}}||fS a  Plot a single or multiple values from the metric.

        Args:
            val: Either a single result from calling `metric.forward` or `metric.compute` or a list of these results.
                If no value is provided, will automatically call `metric.compute` and plot that result.
            ax: An matplotlib axis object. If provided will add plot to that axis
            add_text: if the value of each cell should be added to the plot
            labels: a list of strings, if provided will be added to the plot to indicate the different classes
            cmap: matplotlib colormap to use for the confusion matrix
                https://matplotlib.org/stable/users/explain/colors/colormaps.html

        Returns:
            Figure and Axes object

        Raises:
            ModuleNotFoundError:
                If `matplotlib` is not installed

        .. plot::
            :scale: 75

            >>> from torch import randint
            >>> from torchmetrics.classification import MulticlassConfusionMatrix
            >>> metric = MulticlassConfusionMatrix(num_classes=5)
            >>> metric.update(randint(5, (20,)), randint(5, (20,)))
            >>> fig_, ax_ = metric.plot()

        Nz+Expected val to be a single tensor but got )rO   rP   rQ   rR   rL   
isinstancer   	TypeErrorr   r?   rN   rO   rP   rQ   rR   figr8   r8   rB   plot   
   $
r   r%   NNTNNTNN)__name__
__module____qualname____doc__r!   bool__annotations__r"   r   r#   r   floatintr   r   r:   rH   rL   r   liststrr   r   rZ   __classcell__r8   r8   r@   rB   r    3   sV   
 <

r    c                       s   e Zd ZU dZdZeed< dZee ed< dZ	eed< e
ed< 			dd	ed
ee deed  dededdf fddZde
de
ddfddZde
fddZ					d dee
 dee dedeee  dee defddZ  ZS )!MulticlassConfusionMatrixa  Compute the `confusion matrix`_ for multiclass tasks.

    The confusion matrix :math:`C` is constructed such that :math:`C_{i, j}` is equal to the number of observations
    known to be in class :math:`i` but predicted to be in class :math:`j`. Thus row indices of the confusion matrix
    correspond to the true class labels and column indices correspond to the predicted class labels.

    For multiclass tasks, the confusion matrix is a NxN matrix, where:

    - :math:`C_{i, i}` represents the number of true positives for class :math:`i`
    - :math:`\sum_{j=1, j\neq i}^N C_{i, j}` represents the number of false negatives for class :math:`i`
    - :math:`\sum_{j=1, j\neq i}^N C_{j, i}` represents the number of false positives for class :math:`i`
    - the sum of the remaining cells in the matrix represents the number of true negatives for class :math:`i`

    As input to ``forward`` and ``update`` the metric accepts the following input:

    - ``preds`` (:class:`~torch.Tensor`): An int or float tensor of shape ``(N, ...)``. If preds is a floating point
      tensor with values outside [0,1] range we consider the input to be logits and will auto apply sigmoid per
      element. Additionally, we convert to int tensor with thresholding using the value in ``threshold``.
    - ``target`` (:class:`~torch.Tensor`): An int tensor of shape ``(N, ...)``.

    As output to ``forward`` and ``compute`` the metric returns the following output:

    - ``confusion_matrix``: [num_classes, num_classes] matrix

    Args:
        num_classes: Integer specifying the number of classes
        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        normalize: Normalization mode for confusion matrix. Choose from:

            - ``None`` or ``'none'``: no normalization (default)
            - ``'true'``: normalization over the targets (most commonly used)
            - ``'pred'``: normalization over the predictions
            - ``'all'``: normalization over the whole matrix
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.
        kwargs: Additional keyword arguments, see :ref:`Metric kwargs` for more info.

    Example (pred is integer tensor):
        >>> from torch import tensor
        >>> from torchmetrics.classification import MulticlassConfusionMatrix
        >>> target = tensor([2, 1, 0, 0])
        >>> preds = tensor([2, 1, 0, 1])
        >>> metric = MulticlassConfusionMatrix(num_classes=3)
        >>> metric(preds, target)
        tensor([[1, 1, 0],
                [0, 1, 0],
                [0, 0, 1]])

    Example (pred is float tensor):
        >>> from torchmetrics.classification import MulticlassConfusionMatrix
        >>> target = tensor([2, 1, 0, 0])
        >>> preds = tensor([[0.16, 0.26, 0.58],
        ...                 [0.22, 0.61, 0.17],
        ...                 [0.71, 0.09, 0.20],
        ...                 [0.05, 0.82, 0.13]])
        >>> metric = MulticlassConfusionMatrix(num_classes=3)
        >>> metric(preds, target)
        tensor([[1, 1, 0],
                [0, 1, 0],
                [0, 0, 1]])

    Fr!   Nr"   r#   r$   Tnum_classesr'   r(   r-   r*   r+   r,   r.   r/   r0   c                    s\   t  jdi | |rt||| || _|| _|| _|| _| jdtj	||tj
ddd d S )Nr$   r3   r5   r6   r8   )r9   r:   r   rj   r'   r(   r.   r;   r<   r=   r>   )r?   rj   r'   r(   r.   r/   r@   r8   rB   r:     rC   z"MulticlassConfusionMatrix.__init__rD   rE   c                 C   sJ   | j rt||| j| j t||| j\}}t||| j}|  j|7  _dS rF   )r.   r   rj   r'   r   r   r$   rG   r8   r8   rB   rH     s
   z MulticlassConfusionMatrix.updatec                 C   rI   rJ   )r   r$   r(   rK   r8   r8   rB   rL      rM   z!MulticlassConfusionMatrix.computerN   rO   rP   rQ   rR   c                 C   rS   rT   rU   rX   r8   r8   rB   rZ   $  r[   r   )NNTr]   )r^   r_   r`   ra   r!   rb   rc   r"   r   r#   r   re   r   r   r:   rH   rL   r   rf   rg   r   r   rZ   rh   r8   r8   r@   rB   ri      sT   
 @

ri   c                       s   e Zd ZU dZdZeed< dZee ed< dZ	eed< e
ed< 					d!d
ededee deed  dededdf fddZde
de
ddfddZde
fddZ						d"dee
 dee dedeee  dee defdd Z  ZS )#MultilabelConfusionMatrixa  Compute the `confusion matrix`_ for multilabel tasks.

    The confusion matrix :math:`C` is constructed such that :math:`C_{i, j}` is equal to the number of observations
    known to be in class :math:`i` but predicted to be in class :math:`j`. Thus row indices of the confusion matrix
    correspond to the true class labels and column indices correspond to the predicted class labels.

    For multilabel tasks, the confusion matrix is a Nx2x2 tensor, where each 2x2 matrix corresponds to the confusion
    for that label. The structure of each 2x2 matrix is as follows:

    - :math:`C_{0, 0}`: True negatives
    - :math:`C_{0, 1}`: False positives
    - :math:`C_{1, 0}`: False negatives
    - :math:`C_{1, 1}`: True positives

    As input to 'update' the metric accepts the following input:

    - ``preds`` (int or float tensor): ``(N, C, ...)``. If preds is a floating point tensor with values outside
      [0,1] range we consider the input to be logits and will auto apply sigmoid per element. Additionally,
      we convert to int tensor with thresholding using the value in ``threshold``.
    - ``target`` (int tensor): ``(N, C, ...)``

    As output of 'compute' the metric returns the following output:

    - ``confusion matrix``: [num_labels,2,2] matrix

    Args:
        num_classes: Integer specifying the number of labels
        threshold: Threshold for transforming probability to binary (0,1) predictions
        ignore_index:
            Specifies a target value that is ignored and does not contribute to the metric calculation
        normalize: Normalization mode for confusion matrix. Choose from:

            - ``None`` or ``'none'``: no normalization (default)
            - ``'true'``: normalization over the targets (most commonly used)
            - ``'pred'``: normalization over the predictions
            - ``'all'``: normalization over the whole matrix
        validate_args: bool indicating if input arguments and tensors should be validated for correctness.
            Set to ``False`` for faster computations.
        kwargs: Additional keyword arguments, see :ref:`Metric kwargs` for more info.

    Example (preds is int tensor):
        >>> from torch import tensor
        >>> from torchmetrics.classification import MultilabelConfusionMatrix
        >>> target = tensor([[0, 1, 0], [1, 0, 1]])
        >>> preds = tensor([[0, 0, 1], [1, 0, 1]])
        >>> metric = MultilabelConfusionMatrix(num_labels=3)
        >>> metric(preds, target)
        tensor([[[1, 0], [0, 1]],
                [[1, 0], [1, 0]],
                [[0, 1], [0, 1]]])

    Example (preds is float tensor):
        >>> from torchmetrics.classification import MultilabelConfusionMatrix
        >>> target = tensor([[0, 1, 0], [1, 0, 1]])
        >>> preds = tensor([[0.11, 0.22, 0.84], [0.73, 0.33, 0.92]])
        >>> metric = MultilabelConfusionMatrix(num_labels=3)
        >>> metric(preds, target)
        tensor([[[1, 0], [0, 1]],
                [[1, 0], [1, 0]],
                [[0, 1], [0, 1]]])

    Fr!   Nr"   r#   r$   r%   T
num_labelsr&   r'   r(   rk   r.   r/   r0   c                    sf   t  jdi | |rt|||| || _|| _|| _|| _|| _| jdt	j
|ddt	jddd d S r1   )r9   r:   r   rm   r&   r'   r(   r.   r;   r<   r=   r>   )r?   rm   r&   r'   r(   r.   r/   r@   r8   rB   r:     s   	$z"MultilabelConfusionMatrix.__init__rD   rE   c                 C   sR   | j rt||| j| j t||| j| j| j\}}t||| j}|  j|7  _dS rF   )r.   r   rm   r'   r   r&   r   r$   rG   r8   r8   rB   rH     s   z MultilabelConfusionMatrix.updatec                 C   rI   rJ   )r   r$   r(   rK   r8   r8   rB   rL     rM   z!MultilabelConfusionMatrix.computerN   rO   rP   rQ   rR   c                 C   rS   rT   rU   rX   r8   r8   rB   rZ     r[   r   r\   r]   )r^   r_   r`   ra   r!   rb   rc   r"   r   r#   r   re   rd   r   r   r:   rH   rL   r   rf   rg   r   r   rZ   rh   r8   r8   r@   rB   rl   O  sZ   
 ?


rl   c                   @   sl   e Zd ZdZ						dded  ded ded	ee d
ee deed  dee de	de
defddZdS )ConfusionMatrixa  Compute the `confusion matrix`_.

    This function is a simple wrapper to get the task specific versions of this metric, which is done by setting the
    ``task`` argument to either ``'binary'``, ``'multiclass'`` or ``'multilabel'``. See the documentation of
    :class:`~torchmetrics.classification.BinaryConfusionMatrix`,
    :class:`~torchmetrics.classification.MulticlassConfusionMatrix` and
    :class:`~torchmetrics.classification.MultilabelConfusionMatrix` for the specific details of each argument influence
    and examples.

    Legacy Example:
        >>> from torch import tensor
        >>> target = tensor([1, 1, 0, 0])
        >>> preds = tensor([0, 1, 0, 0])
        >>> confmat = ConfusionMatrix(task="binary", num_classes=2)
        >>> confmat(preds, target)
        tensor([[2, 0],
                [1, 1]])

        >>> target = tensor([2, 1, 0, 0])
        >>> preds = tensor([2, 1, 0, 1])
        >>> confmat = ConfusionMatrix(task="multiclass", num_classes=3)
        >>> confmat(preds, target)
        tensor([[1, 1, 0],
                [0, 1, 0],
                [0, 0, 1]])

        >>> target = tensor([[0, 1, 0], [1, 0, 1]])
        >>> preds = tensor([[0, 0, 1], [1, 0, 1]])
        >>> confmat = ConfusionMatrix(task="multilabel", num_labels=3)
        >>> confmat(preds, target)
        tensor([[[1, 0], [0, 1]],
                [[1, 0], [1, 0]],
                [[0, 1], [0, 1]]])

    r%   NTclstask)binary
multiclass
multilabelr&   rj   rm   r(   r)   r'   r.   r/   r0   c           	      K   s   t |}||||d |t jkrt|fi |S |t jkr7t|ts/tdt	| dt
|fi |S |t jkrTt|tsKtdt	| dt||fi |S td| d)zInitialize task metric.)r(   r'   r.   z+`num_classes` is expected to be `int` but `z was passed.`z*`num_labels` is expected to be `int` but `zTask z not supported!)r   from_strrH   BINARYr    
MULTICLASSrV   re   
ValueErrortyperi   
MULTILABELrl   )	ro   rp   r&   rj   rm   r(   r'   r.   r/   r8   r8   rB   __new__  s   





zConfusionMatrix.__new__)r%   NNNNT)r^   r_   r`   ra   rx   r   rd   r   re   rb   r   r   rz   r8   r8   r8   rB   rn     s:    '
	
rn   ))typingr   r   r<   r   typing_extensionsr    torchmetrics.classification.baser   7torchmetrics.functional.classification.confusion_matrixr   r   r	   r
   r   r   r   r   r   r   r   r   r   r   r   torchmetrics.metricr   torchmetrics.utilities.enumsr   torchmetrics.utilities.importsr   torchmetrics.utilities.plotr   r   r   r   __doctest_skip__r    ri   rl   rn   r8   r8   r8   rB   <module>   s&   D   