o
    	۷i!                     @   s   d Z ddlmZmZ ddlmZ ddlmZ ddlm	Z	 ddl
mZmZmZmZ ddlmZmZ eeZG d	d
 d
eddZG dd deddZG dd deZdgZdS )z
Processor class for UDOP.
    )OptionalUnion)logging   )BatchFeature)
ImageInput)ProcessingKwargsProcessorMixin
TextKwargsUnpack)PreTokenizedInput	TextInputc                   @   sR   e Zd ZU eeee eee  f  ed< eeee  eeee   f ed< dS )UdopTextKwargsword_labelsboxesN)__name__
__module____qualname__r   r   listint__annotations__ r   r   ^/home/ubuntu/vllm_env/lib/python3.10/site-packages/transformers/models/udop/processing_udop.pyr       s   
  (r   F)totalc                
   @   s4   e Zd ZU eed< dddddddddd	i dZdS )UdopProcessorKwargstext_kwargsTFr   )	add_special_tokenspadding
truncationstridereturn_overflowing_tokensreturn_special_tokens_maskreturn_offsets_mappingreturn_lengthverbose)r   images_kwargsN)r   r   r   r   r   	_defaultsr   r   r   r   r   %   s   
 
r   c                
       s   e Zd ZdZddgZdZdZ fddZ				dd	ee	 d
e
eeee ee f dee defddZdd Zedd Z  ZS )UdopProcessora  
    Constructs a UDOP processor which combines a LayoutLMv3 image processor and a UDOP tokenizer into a single processor.

    [`UdopProcessor`] offers all the functionalities you need to prepare data for the model.

    It first uses [`LayoutLMv3ImageProcessor`] to resize, rescale and normalize document images, and optionally applies OCR
    to get words and normalized bounding boxes. These are then provided to [`UdopTokenizer`] or [`UdopTokenizerFast`],
    which turns the words and bounding boxes into token-level `input_ids`, `attention_mask`, `token_type_ids`, `bbox`.
    Optionally, one can provide integer `word_labels`, which are turned into token-level `labels` for token
    classification tasks (such as FUNSD, CORD).

    Additionally, it also supports passing `text_target` and `text_pair_target` to the tokenizer, which can be used to
    prepare labels for language modeling tasks.

    Args:
        image_processor (`LayoutLMv3ImageProcessor`):
            An instance of [`LayoutLMv3ImageProcessor`]. The image processor is a required input.
        tokenizer (`UdopTokenizer` or `UdopTokenizerFast`):
            An instance of [`UdopTokenizer`] or [`UdopTokenizerFast`]. The tokenizer is a required input.
    image_processor	tokenizerLayoutLMv3ImageProcessor)UdopTokenizerUdopTokenizerFastc                    s   t  || d S )N)super__init__)selfr(   r)   	__class__r   r   r.   Q   s   zUdopProcessor.__init__Nimagestextkwargsreturnc                 K   s  | j tfd| jji|}|d dd}|d dd}|d dd}	|d dd}
|d d	d}|d d
d}| jjrI|durItd| jjrU|durUtd|
r]|s]td|durk| jdi |d S | jdd|i|d }|dd}|dd}|d d
d |d dd |	|d d< |dur|n||d d< ||d d< |dur| jjr|	du rt	|t
r|g}||d d< | jdd|dur|n|i|d }|
du r| |d |d |d< || |S )a~  
        This method first forwards the `images` argument to [`~UdopImageProcessor.__call__`]. In case
        [`UdopImageProcessor`] was initialized with `apply_ocr` set to `True`, it passes the obtained words and
        bounding boxes along with the additional arguments to [`~UdopTokenizer.__call__`] and returns the output,
        together with the prepared `pixel_values`. In case [`UdopImageProcessor`] was initialized with `apply_ocr` set
        to `False`, it passes the words (`text`/``text_pair`) and `boxes` specified by the user along with the
        additional arguments to [`~UdopTokenizer.__call__`] and returns the output, together with the prepared
        `pixel_values`.

        Alternatively, one can pass `text_target` and `text_pair_target` to prepare the targets of UDOP.

        Please refer to the docstring of the above two methods for more information.
        tokenizer_init_kwargsr   r   Nr   	text_pairr    Fr"   text_targetzdYou cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True.zaYou cannot provide word labels if you initialized the image processor with apply_ocr set to True.zKYou cannot return overflowing tokens without returning the offsets mapping.r2   r%   wordstext_pair_targetr3   Tpixel_valuesoverflow_to_sample_mappingr   )_merge_kwargsr   r)   init_kwargspopgetr(   	apply_ocr
ValueError
isinstancestrget_overflowing_imagesupdate)r/   r2   r3   audiovideosr4   output_kwargsr   r   r7   r    r"   r8   featuresfeatures_wordsfeatures_boxesencoded_inputsr   r   r   __call__T   sd   

zUdopProcessor.__call__c                 C   sL   g }|D ]	}| ||  qt|t|kr$tdt| dt| |S )Nz`Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got z and )appendlenrB   )r/   r2   r<   images_with_overflow
sample_idxr   r   r   rE      s   z$UdopProcessor.get_overflowing_imagesc                 C   s"   | j j}| jj}t|| dg S )Nbbox)r)   model_input_namesr(   r   )r/   tokenizer_input_namesimage_processor_input_namesr   r   r   rT      s   zUdopProcessor.model_input_names)NNNN)r   r   r   __doc__
attributesimage_processor_classtokenizer_classr.   r   r   r   r   r   r   r   r   r   rN   rE   propertyrT   __classcell__r   r   r0   r   r'   7   s,    
Xr'   N)rW   typingr   r   transformersr   image_processing_utilsr   image_utilsr   processing_utilsr   r	   r
   r   tokenization_utils_baser   r   
get_loggerr   loggerr   r   r'   __all__r   r   r   r   <module>   s   
 
