o
    ٷi                     @  s   d dl mZ d dlZd dlZd dlmZ ejdddejdddejddd	d
dejdgdddejddddejddddejddddejddddejddddf	d-d+d,Z	dS ).    )annotationsN)	load_chat.z;OpenAI-compatible base URL, e.g. http://localhost:11434/v1/)helpzModel name, e.g. llama3.2z--tokenz-tz5API key (defaults to $OPENAI_API_KEY if not provided)text_encodedz--file-typesz6Repeatable option. Allowed values: text_encoded, imagez--system-messagezOptional system promptTz--stream/--no-streamzEnable or disable streamingz--hostz(Server host (maps to launch.server_name)i  z--portz(Server port (maps to launch.server_port)Fz--share/--no-sharezCreate a public share linkbase_urlstrmodeltoken
str | None
file_types	list[str]system_messagestreamboolhostportintsharereturnNonec	                 C  sR   |pt d}	|D ]}
|
dvrtdq	t| ||	|||d}|j|||d dS )z4Launch a chat interface using OpenAI-compatible API.OPENAI_API_KEY>   imager   z.file_types must be one of: text_encoded, image)r   r   r	   r   r   	streaming)server_nameserver_portr   N)osgetenvtyperBadParameterload_chat_externallaunch)r   r   r	   r   r   r   r   r   r   resolved_tokenftdemo r$   Q/home/ubuntu/.local/lib/python3.10/site-packages/gradio/cli/commands/load_chat.pymain
   s   !
	r&   )r   r   r   r   r	   r
   r   r   r   r
   r   r   r   r
   r   r   r   r   r   r   )

__future__r   r   r   gradio.externalr   r   ArgumentOptionr&   r$   r$   r$   r%   <module>   sF    