o
    5ti^%                     @   s   d dl mZmZmZmZmZmZmZmZ d dl	m
Z
 zdZW n ey)   dZY nw d dlmZ ddlmZ ddlmZmZ dd	lmZ d
dlmZmZmZ d
dlmZ erbd dlmZmZmZm Z  ee!Z"dZ#G dd deZ$dS )    )TYPE_CHECKINGAnyCallableDictListOptionalTupleUnion)DatasetTF)Literal   )EvaluationModule)add_end_docstringsadd_start_docstrings)
get_logger   )"EVALUATOR_COMPUTE_RETURN_DOCSTRING EVALUTOR_COMPUTE_START_DOCSTRING	Evaluator)DatasetColumn)PipelinePreTrainedModelPreTrainedTokenizerTFPreTrainedModela0  
    Examples:
    ```python
    >>> from evaluate import evaluator
    >>> from datasets import load_dataset
    >>> task_evaluator = evaluator("question-answering")
    >>> data = load_dataset("squad", split="validation[:2]")
    >>> results = task_evaluator.compute(
    >>>     model_or_pipeline="sshleifer/tiny-distilbert-base-cased-distilled-squad",
    >>>     data=data,
    >>>     metric="squad",
    >>> )
    ```

    <Tip>

    Datasets where the answer may be missing in the context are supported, for example SQuAD v2 dataset. In this case, it is safer to pass `squad_v2_format=True` to
    the compute() call.

    </Tip>

    ```python
    >>> from evaluate import evaluator
    >>> from datasets import load_dataset
    >>> task_evaluator = evaluator("question-answering")
    >>> data = load_dataset("squad_v2", split="validation[:2]")
    >>> results = task_evaluator.compute(
    >>>     model_or_pipeline="mrm8488/bert-tiny-finetuned-squadv2",
    >>>     data=data,
    >>>     metric="squad_v2",
    >>>     squad_v2_format=True,
    >>> )
    ```
c                (       s>  e Zd ZdZi Zd- fdd	Zdededed	ed
ef
ddZd.ded
efddZ	de
dede
fddZeeeee																d/deededdf deeef dee d ee d!eeef d"eeed#f  d$ed% d&ed'ed(ed)ee deded	ed
edee d*eeeef ef f"d+d,Z  ZS )0QuestionAnsweringEvaluatora  
    Question answering evaluator. This evaluator handles
    [**extractive** question answering](https://huggingface.co/docs/transformers/task_summary#extractive-question-answering),
    where the answer to the question is extracted from a context.

    This question answering evaluator can currently be loaded from [`evaluator`] using the default task name
    `question-answering`.

    Methods in this class assume a data format compatible with the
    [`~transformers.QuestionAnsweringPipeline`].
    question-answeringNc                    s   t  j||d d S )N)default_metric_name)super__init__)selftaskr   	__class__ Y/home/ubuntu/.local/lib/python3.10/site-packages/evaluate/evaluator/question_answering.pyr   ]   s   z#QuestionAnsweringEvaluator.__init__dataquestion_columncontext_column	id_columnlabel_columnc                    s^   |du rt d| ||| d t } fdd|D |d< |t||t||dfS )zPrepare data.NzXPlease specify a valid `data` object - either a `str` with a name or a `Dataset` object.)r&   r'   r(   r)   c                    s   g | ]}|  | d qS ))idanswersr#   ).0elementr(   r)   r#   r$   
<listcomp>s   s    z;QuestionAnsweringEvaluator.prepare_data.<locals>.<listcomp>
references)questioncontext)
ValueErrorcheck_required_columnsdictr   )r   r%   r&   r'   r(   r)   metric_inputsr#   r.   r$   prepare_data`   s(   

z'QuestionAnsweringEvaluator.prepare_datar+   c                    s.   |j }|j fddddj }||krdS dS )z
        Check if the provided dataset follows the squad v2 data schema, namely possible samples where the answer is not in the context.
        In this case, the answer text list should be `[]`.
        c                    s   t |   d dkS )Ntextr   )len)xr)   r#   r$   <lambda>   s    z?QuestionAnsweringEvaluator.is_squad_v2_format.<locals>.<lambda>F)load_from_cache_fileT)num_rowsfilter)r   r%   r)   original_num_rowsnonempty_num_rowsr#   r;   r$   is_squad_v2_format|   s   z-QuestionAnsweringEvaluator.is_squad_v2_formatpredictionssquad_v2_formatidsc                 C   sR   g }t t|D ]}|| d || d}|r|| d |d< || qd|iS )Nanswer)prediction_textr*   scoreno_answer_probabilityrC   )ranger9   append)r   rC   rD   rE   resultipredr#   r#   r$   predictions_processor   s   z0QuestionAnsweringEvaluator.predictions_processorsimpleffffff?'  r1   r2   r*   model_or_pipeliner   r   r   subsetsplitmetric	tokenizerr   strategy)rP   	bootstrapconfidence_leveln_resamplesdevicerandom_statereturnc                 C   s&  i }|  |
| | j|||d}| j|||||d\}}|du r0| j||d}td| d | j|||
d}| |}|rI|jdkrItd	 |sU|jd
krUtd |r]d| j	d< nd| j	d< | j
|fi |\}}| j|||| d}|| | j|||||	|d}|| || |S )a  
        question_column (`str`, defaults to `"question"`):
            The name of the column containing the question in the dataset specified by `data`.
        context_column (`str`, defaults to `"context"`):
            The name of the column containing the context in the dataset specified by `data`.
        id_column (`str`, defaults to `"id"`):
            The name of the column containing the identification field of the question and answer pair in the
            dataset specified by `data`.
        label_column (`str`, defaults to `"answers"`):
            The name of the column containing the answers in the dataset specified by `data`.
        squad_v2_format (`bool`, *optional*, defaults to `None`):
            Whether the dataset follows the format of squad_v2 dataset. This is the case when the provided dataset
            has questions where the answer is not in the context, more specifically when are answers as
            `{"text": [], "answer_start": []}` in the answer column. If all questions have at least one answer, this parameter
            should be set to `False`. If this parameter is not provided, the format will be automatically inferred.
        )r%   rT   rU   )r%   r&   r'   r(   r)   N)r%   r)   z~`squad_v2_format` parameter not provided to QuestionAnsweringEvaluator.compute(). Automatically inferred `squad_v2_format` as .)rS   rW   r\   squadzkThe dataset has SQuAD v2 format but you are using the SQuAD metric. Consider passing the 'squad_v2' metric.squad_v2zkThe dataset has SQuAD v1 format but you are using the SQuAD v2 metric. Consider passing the 'squad' metric.Thandle_impossible_answerF)rD   rE   )rV   r6   rX   rZ   r[   r]   )"check_for_mismatch_in_device_setup	load_datar7   rB   loggerwarningprepare_pipelineprepare_metricnamePIPELINE_KWARGScall_pipelinerO   updatecompute_metric)r   rS   r%   rT   rU   rV   rW   rX   rZ   r[   r\   r]   r&   r'   r(   r)   rD   rL   r6   pipe_inputspiperC   perf_resultsmetric_resultsr#   r#   r$   compute   sT   '





	
z"QuestionAnsweringEvaluator.compute)r   N)r+   )NNNNNNrP   rQ   rR   NNr1   r2   r*   r+   N)__name__
__module____qualname____doc__rj   r   r
   strr7   rB   r   boolrO   r   r   r   r   TASK_DOCUMENTATIONr	   r   r   r   r   floatintr   r   r   rr   __classcell__r#   r#   r!   r$   r   N   s    
	

	
r   N)%typingr   r   r   r   r   r   r   r	   datasetsr
   TRANSFORMERS_AVAILABLEImportErrortyping_extensionsr   moduler   utils.file_utilsr   r   utils.loggingr   baser   r   r   utilsr   transformersr   r   r   r   rs   re   ry   r   r#   r#   r#   r$   <module>   s$   ($