o
    wij^                     @   s  d Z ddlmZ ddlmZ ddlmZ ddlm	Z
 e
 Ze dZeee  eede  ejdkrd	e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_de _!de _"de#_!de#_"de$_!de$_"de%_!de%_"de&_!de&_"de'_!de'_"de(_!de(_"de)_!de)_"de*_!de*_"de+_!de+_"de,_!de,_"de-_!de-_"de._!de._"d e/_!d!e/_"d"e0_!d!e0_"d#e1_!d$e1_"d%e2_!d&e2_"d'e3_!d(e3_"d)e4_!d*e4_"d+e5_!d,e5_"d-e6_!d.e6_"d/e7_!d0e7_"d1e8_!d0e8_"d2e9_!d3e9_"d4e:_!d5e:_"d6e_!d5e_"d6e_!d5e_"d7e;_!d8e;_"d9e<_!d8e<_"d:e_!d8e_"d;e=_!d<e=_"d=e>_!d,e>_"d>e?_!d?e?_"d@e@_!d?e@_"dAeA_!dBeA_"dCe_!dBe_"dDeB_!dEeB_"dFeC_!dGeC_"dHeD_!dGeD_"dIeE_!dJeE_"dKeF_!dLeF_"dMeG_!dNeG_"dOeH_!dPeH_"dQeI_!dReI_"dSeJ_!dTeJ_"dUeK_!dTeK_"dVe_!dWe_"dXe_!dTe_"dYeL_!dZeL_"d[eM_!d\eM_"d]eN_!d^eN_"d_e_!d\e_"d`eO_!daeO_"dbeP_!dceP_"ddeQ_!deeQ_"dfeR_!deeR_"d:e_!d8e_"dgeS_!dheS_"dieT_!djeT_"dkeU_!djeU_"dleV_!dmeV_"dneW_!doeW_"dpeX_!dqeX_"dre_!dse_"dte_!due_"dve_!dwe_"d	S d	S )xzGenerated protocol buffer code.    )builder)
descriptor)descriptor_pool)symbol_databasesJ#  
model_config.proto	inference"
ModelRateLimiter7
	resources (2$.inference.ModelRateLimiter.Resource
priority (7
Resource
name (	
global (
count ("
ModelInstanceGroup
name (	0
kind (2".inference.ModelInstanceGroup.Kind
count (1
rate_limiter (2.inference.ModelRateLimiter
gpus (H
secondary_devices (2-.inference.ModelInstanceGroup.SecondaryDevice
profile (	
passive (
host_policy	 (	
SecondaryDeviceO
kind (2A.inference.ModelInstanceGroup.SecondaryDevice.SecondaryDeviceKind
	device_id ("%
SecondaryDeviceKind

KIND_NVDLA "A
Kind
	KIND_AUTO 
KIND_GPU
KIND_CPU

KIND_MODEL"#
ModelTensorReshape
shape ("

ModelInput
name (	&
	data_type (2.inference.DataType,
format (2.inference.ModelInput.Format
dims (.
reshape (2.inference.ModelTensorReshape
is_shape_tensor (
allow_ragged_batch (
optional (
is_non_linear_format_io	 (";
Format
FORMAT_NONE 
FORMAT_NHWC
FORMAT_NCHW"
ModelOutput
name (	&
	data_type (2.inference.DataType
dims (.
reshape (2.inference.ModelTensorReshape
label_filename (	
is_shape_tensor (
is_non_linear_format_io ("

BatchInput(
kind (2.inference.BatchInput.Kind
target_name (	&
	data_type (2.inference.DataType
source_input (	"
Kind
BATCH_ELEMENT_COUNT #
BATCH_ACCUMULATED_ELEMENT_COUNT-
)BATCH_ACCUMULATED_ELEMENT_COUNT_WITH_ZERO$
 BATCH_MAX_ELEMENT_COUNT_AS_SHAPE
BATCH_ITEM_SHAPE
BATCH_ITEM_SHAPE_FLATTEN"
BatchOutput
target_name (	)
kind (2.inference.BatchOutput.Kind
source_input (	"*
Kind"
BATCH_SCATTER_WITH_INPUT_SHAPE "
ModelVersionPolicy6
latest (2$.inference.ModelVersionPolicy.LatestH 0
all (2!.inference.ModelVersionPolicy.AllH :
specific (2&.inference.ModelVersionPolicy.SpecificH 
Latest
num_versions (
All
Specific
versions (B
policy_choice"
ModelOptimizationPolicy7
graph (2(.inference.ModelOptimizationPolicy.GraphB
priority (20.inference.ModelOptimizationPolicy.ModelPriority5
cuda (2'.inference.ModelOptimizationPolicy.CudaX
execution_accelerators (28.inference.ModelOptimizationPolicy.ExecutionAcceleratorsR
input_pinned_memory (25.inference.ModelOptimizationPolicy.PinnedMemoryBufferS
output_pinned_memory (25.inference.ModelOptimizationPolicy.PinnedMemoryBuffer&
gather_kernel_buffer_threshold (
eager_batching (
Graph
level (
Cuda
graphs (
busy_wait_events (E

graph_spec (21.inference.ModelOptimizationPolicy.Cuda.GraphSpec
output_copy_stream (
	GraphSpec

batch_size (K
input (2<.inference.ModelOptimizationPolicy.Cuda.GraphSpec.InputEntryW
graph_lower_bound (2<.inference.ModelOptimizationPolicy.Cuda.GraphSpec.LowerBound
Shape
dim (

LowerBound

batch_size (V
input (2G.inference.ModelOptimizationPolicy.Cuda.GraphSpec.LowerBound.InputEntrye

InputEntry
key (	F
value (27.inference.ModelOptimizationPolicy.Cuda.GraphSpec.Shape:8e

InputEntry
key (	F
value (27.inference.ModelOptimizationPolicy.Cuda.GraphSpec.Shape:8
ExecutionAcceleratorsg
gpu_execution_accelerator (2D.inference.ModelOptimizationPolicy.ExecutionAccelerators.Acceleratorg
cpu_execution_accelerator (2D.inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator
Accelerator
name (	h

parameters (2T.inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator.ParametersEntry1
ParametersEntry
key (	
value (	:8$
PinnedMemoryBuffer
enable ("I
ModelPriority
PRIORITY_DEFAULT 
PRIORITY_MAX
PRIORITY_MIN"
ModelQueuePolicyA
timeout_action (2).inference.ModelQueuePolicy.TimeoutAction$
default_timeout_microseconds (
allow_timeout_override (
max_queue_size ("&
TimeoutAction

REJECT 	
DELAY"
ModelDynamicBatching
preferred_batch_size ($
max_queue_delay_microseconds (
preserve_ordering (
priority_levels (
default_priority_level (9
default_queue_policy (2.inference.ModelQueuePolicyW
priority_queue_policy (28.inference.ModelDynamicBatching.PriorityQueuePolicyEntryW
PriorityQueuePolicyEntry
key (*
value (2.inference.ModelQueuePolicy:8"

ModelSequenceBatchingA
direct (2/.inference.ModelSequenceBatching.StrategyDirectH A
oldest (2/.inference.ModelSequenceBatching.StrategyOldestH &
max_sequence_idle_microseconds (D
control_input (2-.inference.ModelSequenceBatching.ControlInput5
state (2&.inference.ModelSequenceBatching.State
iterative_sequence (
Control;
kind (2-.inference.ModelSequenceBatching.Control.Kind
int32_false_true (
fp32_false_true (
bool_false_true (&
	data_type (2.inference.DataType"u
Kind
CONTROL_SEQUENCE_START 
CONTROL_SEQUENCE_READY
CONTROL_SEQUENCE_END
CONTROL_SEQUENCE_CORRIDW
ControlInput
name (	9
control (2(.inference.ModelSequenceBatching.Control
InitialState&
	data_type (2.inference.DataType
dims (
	zero_data (H 
	data_file (	H 
name (	B

state_data
State

input_name (	
output_name (	&
	data_type (2.inference.DataType
dims (D
initial_state (2-.inference.ModelSequenceBatching.InitialState(
 use_same_buffer_for_input_output (
use_growable_memory (X
StrategyDirect$
max_queue_delay_microseconds ( 
minimum_slot_utilization (
StrategyOldest
max_candidate_sequences (
preferred_batch_size ($
max_queue_delay_microseconds (
preserve_ordering (B
strategy_choice"
ModelEnsembling-
step (2.inference.ModelEnsembling.Step
Step

model_name (	
model_version (@
	input_map (2-.inference.ModelEnsembling.Step.InputMapEntryB

output_map (2..inference.ModelEnsembling.Step.OutputMapEntry
model_namespace (	/
InputMapEntry
key (	
value (	:80
OutputMapEntry
key (	
value (	:8"&
ModelParameter
string_value (	"
ModelWarmup
name (	

batch_size (2
inputs (2".inference.ModelWarmup.InputsEntry
count (
Input&
	data_type (2.inference.DataType
dims (
	zero_data (H 
random_data (H 
input_data_file (	H B
input_data_typeK
InputsEntry
key (	+
value (2.inference.ModelWarmup.Input:8".
ModelOperations
op_library_filename (	"+
ModelTransactionPolicy
	decoupled ("
ModelRepositoryAgents6
agents (2&.inference.ModelRepositoryAgents.Agent
Agent
name (	J

parameters (26.inference.ModelRepositoryAgents.Agent.ParametersEntry1
ParametersEntry
key (	
value (	:8"$
ModelResponseCache
enable ("
ModelMetrics=
metric_control (2%.inference.ModelMetrics.MetricControl
MetricControlQ
metric_identifier (26.inference.ModelMetrics.MetricControl.MetricIdentifierS
histogram_options (26.inference.ModelMetrics.MetricControl.HistogramOptionsH "
MetricIdentifier
family (	#
HistogramOptions
buckets (B
metric_options"

ModelConfig
name (	
platform (	
backend (	
runtime (	5
version_policy (2.inference.ModelVersionPolicy
max_batch_size ($
input (2.inference.ModelInput&
output (2.inference.ModelOutput*
batch_input (2.inference.BatchInput,
batch_output (2.inference.BatchOutput8
optimization (2".inference.ModelOptimizationPolicy;
dynamic_batching (2.inference.ModelDynamicBatchingH =
sequence_batching (2 .inference.ModelSequenceBatchingH 9
ensemble_scheduling (2.inference.ModelEnsemblingH 5
instance_group (2.inference.ModelInstanceGroup
default_model_filename (	H
cc_model_filenames	 (2,.inference.ModelConfig.CcModelFilenamesEntry;
metric_tags
 (2&.inference.ModelConfig.MetricTagsEntry:

parameters (2&.inference.ModelConfig.ParametersEntry,
model_warmup (2.inference.ModelWarmup4
model_operations (2.inference.ModelOperationsC
model_transaction_policy (2!.inference.ModelTransactionPolicyA
model_repository_agents (2 .inference.ModelRepositoryAgents5
response_cache (2.inference.ModelResponseCache.
model_metrics (2.inference.ModelMetrics7
CcModelFilenamesEntry
key (	
value (	:81
MetricTagsEntry
key (	
value (	:8L
ParametersEntry
key (	(
value (2.inference.ModelParameter:8B
scheduling_choice*
DataType
TYPE_INVALID 
	TYPE_BOOL

TYPE_UINT8
TYPE_UINT16
TYPE_UINT32
TYPE_UINT64
	TYPE_INT8

TYPE_INT16

TYPE_INT32

TYPE_INT64	
	TYPE_FP16

	TYPE_FP32
	TYPE_FP64
TYPE_STRING
	TYPE_BF16bproto3model_config_pb2FNs   8iH"  iB#  "            i  i  i  iZ  i  i  i  i  i=  i  i@  i  i  io  i  ir  i  i  i  i  i  i  i  i  i  i  i  i  i)
  i?
  iB
  i  i
  i  i  i  i  i0  i  i  i  ir  i  i  i  i  i  i  i  i  i9  i  i  i  i%  i  i'  i~  i  i  i  i  i  i[  i^  i  i  iz  iG  i  iH  iJ  i|  i  i  i  i  i  i  i   i.  i0  i[  i^  iD  i  iF  ij  im  i  i  iv  i  i  i  i  iE"  ix!  i!  i!  i!  i!  i0"  )Y__doc__google.protobuf.internalr   _buildergoogle.protobufr   _descriptorr   _descriptor_poolr   _symbol_databaseDefault_sym_dbAddSerializedFile
DESCRIPTORBuildMessageAndEnumDescriptorsglobalsBuildTopDescriptorsAndMessages_USE_C_DESCRIPTORS_options=_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_LOWERBOUND_INPUTENTRY_serialized_options2_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_INPUTENTRYJ_MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORS_ACCELERATOR_PARAMETERSENTRY._MODELDYNAMICBATCHING_PRIORITYQUEUEPOLICYENTRY#_MODELENSEMBLING_STEP_INPUTMAPENTRY$_MODELENSEMBLING_STEP_OUTPUTMAPENTRY_MODELWARMUP_INPUTSENTRY,_MODELREPOSITORYAGENTS_AGENT_PARAMETERSENTRY"_MODELCONFIG_CCMODELFILENAMESENTRY_MODELCONFIG_METRICTAGSENTRY_MODELCONFIG_PARAMETERSENTRY	_DATATYPE_serialized_start_serialized_end_MODELRATELIMITER_MODELRATELIMITER_RESOURCE_MODELINSTANCEGROUP#_MODELINSTANCEGROUP_SECONDARYDEVICE7_MODELINSTANCEGROUP_SECONDARYDEVICE_SECONDARYDEVICEKIND_MODELINSTANCEGROUP_KIND_MODELTENSORRESHAPE_MODELINPUT_MODELINPUT_FORMAT_MODELOUTPUT_BATCHINPUT_BATCHINPUT_KIND_BATCHOUTPUT_BATCHOUTPUT_KIND_MODELVERSIONPOLICY_MODELVERSIONPOLICY_LATEST_MODELVERSIONPOLICY_ALL_MODELVERSIONPOLICY_SPECIFIC_MODELOPTIMIZATIONPOLICY_MODELOPTIMIZATIONPOLICY_GRAPH_MODELOPTIMIZATIONPOLICY_CUDA'_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC-_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_SHAPE2_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_LOWERBOUND._MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORS:_MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORS_ACCELERATOR+_MODELOPTIMIZATIONPOLICY_PINNEDMEMORYBUFFER&_MODELOPTIMIZATIONPOLICY_MODELPRIORITY_MODELQUEUEPOLICY_MODELQUEUEPOLICY_TIMEOUTACTION_MODELDYNAMICBATCHING_MODELSEQUENCEBATCHING_MODELSEQUENCEBATCHING_CONTROL#_MODELSEQUENCEBATCHING_CONTROL_KIND#_MODELSEQUENCEBATCHING_CONTROLINPUT#_MODELSEQUENCEBATCHING_INITIALSTATE_MODELSEQUENCEBATCHING_STATE%_MODELSEQUENCEBATCHING_STRATEGYDIRECT%_MODELSEQUENCEBATCHING_STRATEGYOLDEST_MODELENSEMBLING_MODELENSEMBLING_STEP_MODELPARAMETER_MODELWARMUP_MODELWARMUP_INPUT_MODELOPERATIONS_MODELTRANSACTIONPOLICY_MODELREPOSITORYAGENTS_MODELREPOSITORYAGENTS_AGENT_MODELRESPONSECACHE_MODELMETRICS_MODELMETRICS_METRICCONTROL,_MODELMETRICS_METRICCONTROL_METRICIDENTIFIER,_MODELMETRICS_METRICCONTROL_HISTOGRAMOPTIONS_MODELCONFIG r`   r`   _/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/tritonclient/grpc/model_config_pb2.py<module>   sN  
 