o
    .wit
                  	   @   sT   d dl mZ d dlZd dlmZmZ d dlmZ ddededee defd	d
ZdS )    )OptionalN)Tensortensor)"_check_retrieval_functional_inputspredstargettop_kreturnc                 C   s   t | |\} }|p| jd }t|ts|dkrtd| dt| dk|t|}|| jt	|| jd dddd  }|
 sHtd| jd	S tjdt|d |jtjd
|dk }ttjt||jtjd
d | S )aX  Compute average precision (for information retrieval), as explained in `IR Average precision`_.

    ``preds`` and ``target`` should be of the same shape and live on the same device. If no ``target`` is ``True``,
    ``0`` is returned. ``target`` must be either `bool` or `integers` and ``preds`` must be ``float``,
    otherwise an error is raised.

    Args:
        preds: estimated probabilities of each document to be relevant.
        target: ground truth about each document being relevant or not.
        top_k: consider only the top k elements (default: ``None``, which considers them all)

    Return:
        a single-value tensor with the average precision (AP) of the predictions ``preds`` w.r.t. the labels ``target``.

    Raises:
        ValueError:
            If ``top_k`` is not ``None`` or an integer larger than 0.

    Example:
        >>> from torchmetrics.functional.retrieval import retrieval_average_precision
        >>> preds = tensor([0.2, 0.3, 0.5])
        >>> target = tensor([True, False, True])
        >>> retrieval_average_precision(preds, target)
        tensor(0.8333)

    r   zAArgument ``top_k`` has to be a positive integer or None, but got .T)sorteddim   g        )device)r   dtype)r   shape
isinstanceint
ValueErrortorchwhere
zeros_liketopkminsumr   r   arangelenfloat32divmean)r   r   r   	positions r!   p/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/torchmetrics/functional/retrieval/average_precision.pyretrieval_average_precision   s   $&(r#   )N)	typingr   r   r   r   torchmetrics.utilities.checksr   r   r#   r!   r!   r!   r"   <module>   s
   $