o
    xi                     @   sR   d dl mZ d dlZd dlZd dlZd dlmZ eded d
ddZ	dd	 Z
dS )    )simplefilterN)utilsignore)actioncategoryc                 C   s  t |t |}}i }| jj}| |}tj| rOtj	||}||d< tjj
||dd}	|	|d< tjj||dd}
|
|d< tjj||dd}||d< n'tj| rvtj||}||d< tj||}||d< tj||}||d	< d
d | D }t||}td|}|S )zCalculate summary metrics for both regressors and classifiers.

    Called by plot_summary_metrics to visualize metrics. Please use the function
    plot_summary_metrics() if you wish to visualize your summary metrics.
    accuracy_scoreweighted)average	precisionrecallf1_scoremaemser2_scorec                 S   s   i | ]
\}}|t |qS  )r   round_2).0namemetricr   r   g/home/ubuntu/.local/lib/python3.10/site-packages/wandb/integration/sklearn/calculate/summary_metrics.py
<dictcomp>0   s    z#summary_metrics.<locals>.<dictcomp>zwandb/metrics/v1)npasarray	__class____name__predictsklearnbaseis_classifiermetricsr   precision_scorerecall_scorer   is_regressormean_absolute_errormean_squared_errorr   items
make_tablewandb	visualize)modelXyX_testy_testr   
model_namey_predr   r
   r   r   r   r   r   tablechartr   r   r   summary_metrics   s0   


r2   c                    s0   g d} fdd|   D }tj||d}|S )N)metric_namemetric_valuer.   c                    s   g | ]	\}}|| gqS r   r   )r   r   valuer.   r   r   
<listcomp>:   s    zmake_table.<locals>.<listcomp>)columnsdata)r%   r'   Table)r   r.   r8   table_contentr0   r   r6   r   r&   8   s   r&   )NNNNN)warningsr   numpyr   r   r'   wandb.integration.sklearnr   FutureWarningr2   r&   r   r   r   r   <module>   s    
+