o
    wib                     @   s  d Z ddlmZ ddlmZ ddlmZ ddlm	Z
 e
 ZddlmZ e dZeee  eede  ejd	krUd
e_d
e_de_d
e_de_d
e_de_d
e_de_d
e_de_d
e_de_d
e_de_d
e_de_d
e _de _d
e!_de!_d
e"_de"_d
e#_de#_d
e$_de$_d
e%_de%_de&_'de&_(de)_'de)_(de*_'de*_(de+_'de+_(de,_'de,_(de-_'de-_(de._'de._(de/_'de/_(de0_'de0_(de1_'de1_(d e2_'de2_(d!e3_'d"e3_(d#e4_'d$e4_(d%e5_'d&e5_(d'e6_'d(e6_(d)e_'d(e_(d*e7_'d+e7_(d)e_'d(e_(d)e_'d(e_(d,e8_'d-e8_(d.e9_'d/e9_(d)e_'d(e_(d)e_'d(e_(d0e:_'d1e:_(d2e;_'d3e;_(d4e<_'d5e<_(d6e=_'d7e=_(d8e>_'d9e>_(d:e?_'d;e?_(d<e@_'d=e@_(d>eA_'d?eA_(d@eB_'dAeB_(dBeC_'dCeC_(dDe_'dCe_(dEeD_'dFeD_(dGeE_'dHeE_(dIeF_'dJeF_(dKeG_'dLeG_(dMeH_'dLeH_(dNeI_'dOeI_(dPe_'dOe_(dQeJ_'dReJ_(dSeK_'dTeK_(dPe_'dOe_(dUeL_'dVeL_(dWeM_'dXeM_(dYeN_'dZeN_(d[eO_'d\eO_(d]e _'dZe _(d^eP_'d_eP_(d`eQ_'daeQ_(dbeR_'dceR_(ddeS_'deeS_(dfeT_'dgeT_(dheU_'dieU_(djeV_'dkeV_(dle!_'die!_(dmeW_'dneW_(doeX_'dpeX_(dqeY_'dreY_(dseZ_'dteZ_(due[_'dve[_(dwe\_'dxe\_(dye"_'dve"_(dze]_'d{e]_(dwe^_'dxe^_(d|e#_'d{e#_(d}e__'d~e__(de`_'de`_(de$_'d~e$_(dea_'dea_(deb_'deb_(de%_'de%_(dec_'dec_(d
S d
S )zGenerated protocol buffer code.    )builder)
descriptor)descriptor_pool)symbol_database)model_config_pb2s$  
grpc_service.proto	inferencemodel_config.proto"
ServerLiveRequest""
ServerLiveResponse
live ("
ServerReadyRequest"$
ServerReadyResponse
ready ("2
ModelReadyRequest
name (	
version (	"#
ModelReadyResponse
ready ("
ServerMetadataRequest"K
ServerMetadataResponse
name (	
version (	

extensions (	"5
ModelMetadataRequest
name (	
version (	"
ModelMetadataResponse
name (	
versions (	
platform (	?
inputs (2/.inference.ModelMetadataResponse.TensorMetadata@
outputs (2/.inference.ModelMetadataResponse.TensorMetadata?
TensorMetadata
name (	
datatype (	
shape ("
InferParameter

bool_param (H 
int64_param (H 
string_param (	H 
double_param (H 
uint64_param (H B
parameter_choice"
InferTensorContents
bool_contents (
int_contents (
int64_contents (
uint_contents (
uint64_contents (
fp32_contents (
fp64_contents (
bytes_contents ("
ModelInferRequest

model_name (	
model_version (	

id (	@

parameters (2,.inference.ModelInferRequest.ParametersEntry=
inputs (2-.inference.ModelInferRequest.InferInputTensorH
outputs (27.inference.ModelInferRequest.InferRequestedOutputTensor
raw_input_contents (
InferInputTensor
name (	
datatype (	
shape (Q

parameters (2=.inference.ModelInferRequest.InferInputTensor.ParametersEntry0
contents (2.inference.InferTensorContentsL
ParametersEntry
key (	(
value (2.inference.InferParameter:8
InferRequestedOutputTensor
name (	[

parameters (2G.inference.ModelInferRequest.InferRequestedOutputTensor.ParametersEntryL
ParametersEntry
key (	(
value (2.inference.InferParameter:8L
ParametersEntry
key (	(
value (2.inference.InferParameter:8"
ModelInferResponse

model_name (	
model_version (	

id (	A

parameters (2-.inference.ModelInferResponse.ParametersEntry@
outputs (2/.inference.ModelInferResponse.InferOutputTensor
raw_output_contents (
InferOutputTensor
name (	
datatype (	
shape (S

parameters (2?.inference.ModelInferResponse.InferOutputTensor.ParametersEntry0
contents (2.inference.InferTensorContentsL
ParametersEntry
key (	(
value (2.inference.InferParameter:8L
ParametersEntry
key (	(
value (2.inference.InferParameter:8"h
ModelStreamInferResponse
error_message (	5
infer_response (2.inference.ModelInferResponse"3
ModelConfigRequest
name (	
version (	"=
ModelConfigResponse&
config (2.inference.ModelConfig"7
ModelStatisticsRequest
name (	
version (	".
StatisticDuration
count (

ns ("
InferStatistics-
success (2.inference.StatisticDuration*
fail (2.inference.StatisticDuration+
queue (2.inference.StatisticDuration3
compute_input (2.inference.StatisticDuration3
compute_infer (2.inference.StatisticDuration4
compute_output (2.inference.StatisticDuration/
	cache_hit (2.inference.StatisticDuration0

cache_miss (2.inference.StatisticDuration"
InferResponseStatistics3
compute_infer (2.inference.StatisticDuration4
compute_output (2.inference.StatisticDuration-
success (2.inference.StatisticDuration*
fail (2.inference.StatisticDuration4
empty_response (2.inference.StatisticDuration,
cancel (2.inference.StatisticDuration"
InferBatchStatistics

batch_size (3
compute_input (2.inference.StatisticDuration3
compute_infer (2.inference.StatisticDuration4
compute_output (2.inference.StatisticDuration":
MemoryUsage
type (	

id (
	byte_size ("
ModelStatistics
name (	
version (	
last_inference (
inference_count (
execution_count (3
inference_stats (2.inference.InferStatistics4
batch_stats (2.inference.InferBatchStatistics,
memory_usage (2.inference.MemoryUsageE
response_stats	 (2-.inference.ModelStatistics.ResponseStatsEntryX
ResponseStatsEntry
key (	1
value (2".inference.InferResponseStatistics:8"J
ModelStatisticsResponse/
model_stats (2.inference.ModelStatistics"
ModelRepositoryParameter

bool_param (H 
int64_param (H 
string_param (	H 
bytes_param (H B
parameter_choice"@
RepositoryIndexRequest
repository_name (	
ready ("
RepositoryIndexResponse=
models (2-.inference.RepositoryIndexResponse.ModelIndexJ

ModelIndex
name (	
version (	
state (	
reason (	"
RepositoryModelLoadRequest
repository_name (	

model_name (	I

parameters (25.inference.RepositoryModelLoadRequest.ParametersEntryV
ParametersEntry
key (	2
value (2#.inference.ModelRepositoryParameter:8"
RepositoryModelLoadResponse"
RepositoryModelUnloadRequest
repository_name (	

model_name (	K

parameters (27.inference.RepositoryModelUnloadRequest.ParametersEntryV
ParametersEntry
key (	2
value (2#.inference.ModelRepositoryParameter:8"
RepositoryModelUnloadResponse"/
SystemSharedMemoryStatusRequest
name (	"
 SystemSharedMemoryStatusResponseI
regions (28.inference.SystemSharedMemoryStatusResponse.RegionsEntryL
RegionStatus
name (	
key (	
offset (
	byte_size (h
RegionsEntry
key (	G
value (28.inference.SystemSharedMemoryStatusResponse.RegionStatus:8"a
!SystemSharedMemoryRegisterRequest
name (	
key (	
offset (
	byte_size ("$
"SystemSharedMemoryRegisterResponse"3
#SystemSharedMemoryUnregisterRequest
name (	"&
$SystemSharedMemoryUnregisterResponse"-
CudaSharedMemoryStatusRequest
name (	"
CudaSharedMemoryStatusResponseG
regions (26.inference.CudaSharedMemoryStatusResponse.RegionsEntryB
RegionStatus
name (	
	device_id (
	byte_size (f
RegionsEntry
key (	E
value (26.inference.CudaSharedMemoryStatusResponse.RegionStatus:8"i
CudaSharedMemoryRegisterRequest
name (	

raw_handle (
	device_id (
	byte_size (""
 CudaSharedMemoryRegisterResponse"1
!CudaSharedMemoryUnregisterRequest
name (	"$
"CudaSharedMemoryUnregisterResponse"
TraceSettingRequest>
settings (2,.inference.TraceSettingRequest.SettingsEntry

model_name (	
SettingValue
value (	\
SettingsEntry
key (	:
value (2+.inference.TraceSettingRequest.SettingValue:8"
TraceSettingResponse?
settings (2-.inference.TraceSettingResponse.SettingsEntry
SettingValue
value (	]
SettingsEntry
key (	;
value (2,.inference.TraceSettingResponse.SettingValue:8"
LogSettingsRequest=
settings (2+.inference.LogSettingsRequest.SettingsEntryh
SettingValue

bool_param (H 
uint32_param (H 
string_param (	H B
parameter_choice[
SettingsEntry
key (	9
value (2*.inference.LogSettingsRequest.SettingValue:8"
LogSettingsResponse>
settings (2,.inference.LogSettingsResponse.SettingsEntryh
SettingValue

bool_param (H 
uint32_param (H 
string_param (	H B
parameter_choice\
SettingsEntry
key (	:
value (2+.inference.LogSettingsResponse.SettingValue:82
GRPCInferenceServiceK

ServerLive.inference.ServerLiveRequest.inference.ServerLiveResponse" N
ServerReady.inference.ServerReadyRequest.inference.ServerReadyResponse" K

ModelReady.inference.ModelReadyRequest.inference.ModelReadyResponse" W
ServerMetadata .inference.ServerMetadataRequest!.inference.ServerMetadataResponse" T
ModelMetadata.inference.ModelMetadataRequest .inference.ModelMetadataResponse" K

ModelInfer.inference.ModelInferRequest.inference.ModelInferResponse" [
ModelStreamInfer.inference.ModelInferRequest#.inference.ModelStreamInferResponse" (0N
ModelConfig.inference.ModelConfigRequest.inference.ModelConfigResponse" Z
ModelStatistics!.inference.ModelStatisticsRequest".inference.ModelStatisticsResponse" Z
RepositoryIndex!.inference.RepositoryIndexRequest".inference.RepositoryIndexResponse" f
RepositoryModelLoad%.inference.RepositoryModelLoadRequest&.inference.RepositoryModelLoadResponse" l
RepositoryModelUnload'.inference.RepositoryModelUnloadRequest(.inference.RepositoryModelUnloadResponse" u
SystemSharedMemoryStatus*.inference.SystemSharedMemoryStatusRequest+.inference.SystemSharedMemoryStatusResponse" {
SystemSharedMemoryRegister,.inference.SystemSharedMemoryRegisterRequest-.inference.SystemSharedMemoryRegisterResponse" 
SystemSharedMemoryUnregister..inference.SystemSharedMemoryUnregisterRequest/.inference.SystemSharedMemoryUnregisterResponse" o
CudaSharedMemoryStatus(.inference.CudaSharedMemoryStatusRequest).inference.CudaSharedMemoryStatusResponse" u
CudaSharedMemoryRegister*.inference.CudaSharedMemoryRegisterRequest+.inference.CudaSharedMemoryRegisterResponse" {
CudaSharedMemoryUnregister,.inference.CudaSharedMemoryUnregisterRequest-.inference.CudaSharedMemoryUnregisterResponse" Q
TraceSetting.inference.TraceSettingRequest.inference.TraceSettingResponse" N
LogSettings.inference.LogSettingsRequest.inference.LogSettingsResponse" bproto3grpc_service_pb2FNs   85   H   J   l   n                     i  i  i  i  ig  ii  i  i  i  io  i  iJ  iM  i  i   i  iT  ih  i  ik  i@  i  i	  i  i	  i	  iP
  iR
  i
  i
  i
  i
  i
  i
  i-  i0  i  i  i  i  i  i  i  i  i  iz  i  i  i!  i  i  i  i  i  iJ  i  i  i-  i  i  i  i  i  i  i  i  i  i  iY  i  i  i  ir  it  i  i  i  i  i  i  i$  i'  i<  i  i  i  i>  i  i  i  i  i  i   i$  i'  i  i  i  i  i  i  i  i  i  i=  i  i  i  i"  i  i%  i$  )d__doc__google.protobuf.internalr   _buildergoogle.protobufr   _descriptorr   _descriptor_poolr   _symbol_databaseDefault_sym_dbtritonclient.grpcr   model__config__pb2AddSerializedFile
DESCRIPTORBuildMessageAndEnumDescriptorsglobalsBuildTopDescriptorsAndMessages_USE_C_DESCRIPTORS_options3_MODELINFERREQUEST_INFERINPUTTENSOR_PARAMETERSENTRY_serialized_options=_MODELINFERREQUEST_INFERREQUESTEDOUTPUTTENSOR_PARAMETERSENTRY"_MODELINFERREQUEST_PARAMETERSENTRY5_MODELINFERRESPONSE_INFEROUTPUTTENSOR_PARAMETERSENTRY#_MODELINFERRESPONSE_PARAMETERSENTRY#_MODELSTATISTICS_RESPONSESTATSENTRY+_REPOSITORYMODELLOADREQUEST_PARAMETERSENTRY-_REPOSITORYMODELUNLOADREQUEST_PARAMETERSENTRY._SYSTEMSHAREDMEMORYSTATUSRESPONSE_REGIONSENTRY,_CUDASHAREDMEMORYSTATUSRESPONSE_REGIONSENTRY"_TRACESETTINGREQUEST_SETTINGSENTRY#_TRACESETTINGRESPONSE_SETTINGSENTRY!_LOGSETTINGSREQUEST_SETTINGSENTRY"_LOGSETTINGSRESPONSE_SETTINGSENTRY_SERVERLIVEREQUEST_serialized_start_serialized_end_SERVERLIVERESPONSE_SERVERREADYREQUEST_SERVERREADYRESPONSE_MODELREADYREQUEST_MODELREADYRESPONSE_SERVERMETADATAREQUEST_SERVERMETADATARESPONSE_MODELMETADATAREQUEST_MODELMETADATARESPONSE%_MODELMETADATARESPONSE_TENSORMETADATA_INFERPARAMETER_INFERTENSORCONTENTS_MODELINFERREQUEST#_MODELINFERREQUEST_INFERINPUTTENSOR-_MODELINFERREQUEST_INFERREQUESTEDOUTPUTTENSOR_MODELINFERRESPONSE%_MODELINFERRESPONSE_INFEROUTPUTTENSOR_MODELSTREAMINFERRESPONSE_MODELCONFIGREQUEST_MODELCONFIGRESPONSE_MODELSTATISTICSREQUEST_STATISTICDURATION_INFERSTATISTICS_INFERRESPONSESTATISTICS_INFERBATCHSTATISTICS_MEMORYUSAGE_MODELSTATISTICS_MODELSTATISTICSRESPONSE_MODELREPOSITORYPARAMETER_REPOSITORYINDEXREQUEST_REPOSITORYINDEXRESPONSE#_REPOSITORYINDEXRESPONSE_MODELINDEX_REPOSITORYMODELLOADREQUEST_REPOSITORYMODELLOADRESPONSE_REPOSITORYMODELUNLOADREQUEST_REPOSITORYMODELUNLOADRESPONSE _SYSTEMSHAREDMEMORYSTATUSREQUEST!_SYSTEMSHAREDMEMORYSTATUSRESPONSE._SYSTEMSHAREDMEMORYSTATUSRESPONSE_REGIONSTATUS"_SYSTEMSHAREDMEMORYREGISTERREQUEST#_SYSTEMSHAREDMEMORYREGISTERRESPONSE$_SYSTEMSHAREDMEMORYUNREGISTERREQUEST%_SYSTEMSHAREDMEMORYUNREGISTERRESPONSE_CUDASHAREDMEMORYSTATUSREQUEST_CUDASHAREDMEMORYSTATUSRESPONSE,_CUDASHAREDMEMORYSTATUSRESPONSE_REGIONSTATUS _CUDASHAREDMEMORYREGISTERREQUEST!_CUDASHAREDMEMORYREGISTERRESPONSE"_CUDASHAREDMEMORYUNREGISTERREQUEST#_CUDASHAREDMEMORYUNREGISTERRESPONSE_TRACESETTINGREQUEST!_TRACESETTINGREQUEST_SETTINGVALUE_TRACESETTINGRESPONSE"_TRACESETTINGRESPONSE_SETTINGVALUE_LOGSETTINGSREQUEST _LOGSETTINGSREQUEST_SETTINGVALUE_LOGSETTINGSRESPONSE!_LOGSETTINGSRESPONSE_SETTINGVALUE_GRPCINFERENCESERVICE rr   rr   Z/home/ubuntu/sommelier/.venv/lib/python3.10/site-packages/tritonclient/grpc/service_pb2.py<module>   s|  
 