o
    ߥi8                     @   sB   d Z ddlmZmZmZmZ ddlZeeZ	G dd de
ZdS )z BERT model configuration     )absolute_importdivisionprint_functionunicode_literalsNc                   @   s6   e Zd ZdZ												
			dddZdS )
BertConfigaU  
        :class:`~transformers.BertConfig` is the configuration class to store the configuration of a
        `BertModel`.


        Arguments:
            vocab_size_or_config_json_file: Vocabulary size of `inputs_ids` in `BertModel`.
            hidden_size: Size of the encoder layers and the pooler layer.
            num_hidden_layers: Number of hidden layers in the Transformer encoder.
            num_attention_heads: Number of attention heads for each attention layer in
                the Transformer encoder.
            intermediate_size: The size of the "intermediate" (i.e., feed-forward)
                layer in the Transformer encoder.
            hidden_act: The non-linear activation function (function or string) in the
                encoder and pooler. If string, "gelu", "relu", "swish" and "gelu_new" are supported.
            hidden_dropout_prob: The dropout probability for all fully connected
                layers in the embeddings, encoder, and pooler.
            attention_probs_dropout_prob: The dropout ratio for the attention
                probabilities.
            max_position_embeddings: The maximum sequence length that this model might
                ever be used with. Typically set this to something large just in case
                (e.g., 512 or 1024 or 2048).
            type_vocab_size: The vocabulary size of the `token_type_ids` passed into
                `BertModel`.
            initializer_range: The stdev of the truncated_normal_initializer for
                initializing all weight matrices.
            layer_norm_eps: The epsilon used by LayerNorm.
    :w           gelu皙?      {Gz?-q=Fc                 C   sX   || _ || _|| _|| _|| _|| _|| _|| _|	| _|
| _	|| _
|| _|| _|| _d S )N)
vocab_sizehidden_sizenum_hidden_layersnum_attention_heads
hidden_actintermediate_sizehidden_dropout_probattention_probs_dropout_probmax_position_embeddingstype_vocab_sizeinitializer_rangelayer_norm_epsoutput_attentionsoutput_hidden_states)selfvocab_size_or_config_json_filer   r   r   r   r   r   r   r   r   r   r   r   r    r!   i/home/ubuntu/.local/lib/python3.10/site-packages/modelscope/models/multi_modal/clip/configuration_bert.py__init__6   s   
zBertConfig.__init__N)r   r   r	   r	   r
   r   r   r   r   r   r   r   FF)__name__
__module____qualname____doc__r#   r!   r!   r!   r"   r      s"    r   )r'   
__future__r   r   r   r   logging	getLoggerr$   loggerobjectr   r!   r!   r!   r"   <module>   s
   
