o
    áÙ·i¼  ã                   @   s0   d Z ddlmZmZmZ ddlmZ g d¢ZdS )zÝ
OpenAI-compatible API entrypoints for vLLM-Omni.

Provides:
- omni_run_server: Main server entry point (auto-detects model type)
- OmniOpenAIServingChat: Unified chat completion handler for both LLM and diffusion models
é    )Úbuild_async_omniÚomni_init_app_stateÚomni_run_server)ÚOmniOpenAIServingChat)r   r   r   r   N)Ú__doc__Ú'vllm_omni.entrypoints.openai.api_serverr   r   r   Ú)vllm_omni.entrypoints.openai.serving_chatr   Ú__all__© r
   r
   úY/home/ubuntu/.local/lib/python3.10/site-packages/vllm_omni/entrypoints/openai/__init__.pyÚ<module>   s   