o
    GÆÏif  ã                   @   s\   d dl mZ d dlZd dlZd dlZd dlmZ eG dd„ deƒƒZ	eG dd„ deƒƒZ
dS )é    )Ú	dataclassN)Ú
BaseOutputc                   @   s   e Zd ZU dZejed< dS )ÚHunyuanVideoPipelineOutputa¶  
    Output class for HunyuanVideo pipelines.

    Args:
        frames (`torch.Tensor`, `np.ndarray`, or list[list[PIL.Image.Image]]):
            list of video outputs - It can be a nested list of length `batch_size,` with each sub-list containing
            denoised PIL image sequences of length `num_frames.` It can also be a NumPy array or Torch tensor of shape
            `(batch_size, num_frames, channels, height, width)`.
    ÚframesN)Ú__name__Ú
__module__Ú__qualname__Ú__doc__ÚtorchÚTensorÚ__annotations__© r   r   úe/home/ubuntu/.local/lib/python3.10/site-packages/diffusers/pipelines/hunyuan_video/pipeline_output.pyr   
   s   
 
r   c                   @   s<   e Zd ZU dZejejB eee	j
j
  B eej B ed< dS )Ú#HunyuanVideoFramepackPipelineOutputa)  
    Output class for HunyuanVideo pipelines.

    Args:
        frames (`torch.Tensor`, `np.ndarray`, or list[list[PIL.Image.Image]]):
            list of video outputs - It can be a nested list of length `batch_size,` with each sub-list containing
            denoised PIL image sequences of length `num_frames.` It can also be a NumPy array or Torch tensor of shape
            `(batch_size, num_frames, channels, height, width)`. Or, a list of torch tensors where each tensor
            corresponds to a latent that decodes to multiple frames.
    r   N)r   r   r   r	   r
   r   ÚnpÚndarrayÚlistÚPILÚImager   r   r   r   r   r      s   
 .r   )Údataclassesr   Únumpyr   Ú	PIL.Imager   r
   Údiffusers.utilsr   r   r   r   r   r   r   Ú<module>   s    