o
    áÙ·if  ã                   @   s<   d Z ddlmZ ddlmZmZmZ ddlmZ g d¢Z	dS )aƒ  
Cache module for diffusion model inference acceleration.

This module provides a unified cache backend system for different caching strategies:
- TeaCache: Timestep Embedding Aware Cache for adaptive transformer caching
- cache-dit: DBCache, SCM, and TaylorSeer caching strategies

Cache backends are instantiated directly via their constructors and configured via OmniDiffusionConfig.
é    )ÚCacheBackend)ÚCacheContextÚTeaCacheConfigÚapply_teacache_hook)ÚTeaCacheBackend)r   r   r   r   r   N)
Ú__doc__Úvllm_omni.diffusion.cache.baser   Ú"vllm_omni.diffusion.cache.teacacher   r   r   Ú*vllm_omni.diffusion.cache.teacache.backendr   Ú__all__© r   r   úV/home/ubuntu/.local/lib/python3.10/site-packages/vllm_omni/diffusion/cache/__init__.pyÚ<module>   s
   
