o
    eiy`                     @   s
  d dl mZ d dlmZ d dlZd dlmZ ddlmZ ddl	m
Z
mZ ddlmZ ddlmZ dd	lmZmZ dd
lmZ ddlmZmZ ddlmZmZ ddlmZmZ ddlmZ ddl m!Z!m"Z"m#Z# ddl$m%Z%m&Z& ddl'm(Z( ddl)m*Z* G dd dej+Z,G dd dej+Z-dej.de/dej.fddZ0	d9dej+dej.d ej.d!ej.d"ej.dB d#e1d$e1d%ee! fd&d'Z2d(d) Z3d:d*d+Z4ee4G d,d- d-ej+Z5G d.d/ d/ej+Z6G d0d1 d1eZ7e"G d2d3 d3eZ8e"G d4d5 d5e8Z9e"G d6d7 d7e8eZ:g d8Z;dS );    )Callable)OptionalN   )ACT2FN)CacheDynamicCache)GenerationMixin)use_kernelized_func)create_causal_mask!create_sliding_window_causal_mask)GradientCheckpointingLayer)BaseModelOutputWithPastCausalLMOutputWithPast)ROPE_INIT_FUNCTIONSdynamic_rope_update)ALL_ATTENTION_FUNCTIONSPreTrainedModel)Unpack)TransformersKwargsauto_docstringcan_return_tuple)maybe_autocastmerge_with_config_defaults)capture_outputs   )Cohere2Configc                       s~   e Zd ZU ejed< ddef fddZe			ddedB de	d de
dB d	ed
ef fddZe edd Z  ZS )Cohere2RotaryEmbeddinginv_freqNconfigc                    s   t    |j| _|j| _|| _| jjd | _| j}| jdkr$t	| j }|| j|\}| _
| jd|dd | jd| dd d S )N	rope_typedefaultr   F)
persistentoriginal_inv_freq)super__init__max_position_embeddingsmax_seq_len_cachedoriginal_max_seq_lenr   rope_parametersr   compute_default_rope_parametersr   attention_scalingregister_bufferclone)selfr   devicerope_init_fnr   	__class__ j/home/ubuntu/transcripts/venv/lib/python3.10/site-packages/transformers/models/cohere2/modeling_cohere2.pyr$   .   s   


zCohere2RotaryEmbedding.__init__r.   ztorch.deviceseq_lenreturnztorch.Tensorc                 C   sZ   | j d }t| ddp| j| j }d}d|tjd|dtjdj|tjd|   }||fS )	a  
        Computes the inverse frequencies according to the original RoPE implementation
        Args:
            config ([`~transformers.PreTrainedConfig`]):
                The model configuration.
            device (`torch.device`):
                The device to use for initialization of the inverse frequencies.
            seq_len (`int`, *optional*):
                The current sequence length. Unused for this type of RoPE.
        Returns:
            Tuple of (`torch.Tensor`, `float`), containing the inverse frequencies for the RoPE embeddings and the
            post-processing scaling factor applied to the computed cos/sin (unused in this type of RoPE).
        
rope_thetahead_dimNg      ?r      dtype)r.   r:   )	r(   getattrhidden_sizenum_attention_headstorcharangeint64tofloat)r   r.   r4   basedimattention_factorr   r2   r2   r3   r)   >   s   
&z6Cohere2RotaryEmbedding.compute_default_rope_parametersc           
      C   s   | j d d d d f  |jd dd}|d d d d d f  }t|jjtr2|jjdkr2|jjnd}t|dd* | |  	dd}t
j|ddd	}| | j }| | j }	W d    n1 sgw   Y  |j|jd
|	j|jd
fS )Nr   r   mpscpuF)device_typeenabledr8   rD   r9   )r   rB   expandshape
isinstancer.   typestrr   	transposer>   repeat_interleavecosr*   sinrA   r:   )
r-   xposition_idsinv_freq_expandedposition_ids_expandedrI   freqsembrS   rT   r2   r2   r3   forward\   s   (&zCohere2RotaryEmbedding.forwardN)NNN)__name__
__module____qualname__r>   Tensor__annotations__r   r$   staticmethodr   inttuplerB   r)   no_gradr   r[   __classcell__r2   r2   r0   r3   r   +   s&   
 

r   c                       s&   e Zd Zd fdd	Zdd Z  ZS )	Cohere2LayerNormNh㈵>Fc                    s&   t    tt|| _|| _dS )zcThe hidden size can be a tuple or an int. The tuple is used for QKNorm to normalize across head_dimN)r#   r$   nn	Parameterr>   onesweightvariance_epsilon)r-   r<   epsbiasr0   r2   r3   r$   m   s   

zCohere2LayerNorm.__init__c                 C   sl   |j }|tj}|jddd}|| djddd}|| t|| j  }| jtj| }||S )NrF   T)keepdimr8   )	r:   rA   r>   float32meanpowrsqrtrm   rl   )r-   hidden_statesinput_dtyperr   variancer2   r2   r3   r[   s   s   
zCohere2LayerNorm.forward)Nrh   Fr]   r^   r_   r$   r[   rf   r2   r2   r0   r3   rg   l   s    rg   ru   n_repr5   c                 C   s^   | j \}}}}|dkr| S | dddddddddf |||||} | ||| ||S )z
    This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch,
    num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim)
    r   N)rM   rL   reshape)ru   ry   batchnum_key_value_headsslenr7   r2   r2   r3   	repeat_kv}   s
   0r~           modulequerykeyvalueattention_maskscalingdropoutkwargsc                 K   s   t || j}t || j}	t||dd| }
|d ur |
| }
tjj|
dtjd	|j
}
tjj|
|| jd}
t|
|	}|dd }||
fS )Nr8   r   rF   )rD   r:   )ptrainingr   )r~   num_key_value_groupsr>   matmulrQ   ri   
functionalsoftmaxrq   rA   r:   r   r   
contiguous)r   r   r   r   r   r   r   r   
key_statesvalue_statesattn_weightsattn_outputr2   r2   r3   eager_attention_forward   s   
r   c                 C   sB   | dd d df }| ddd df }t j| |gddd}|S )N.r8   r   rF   rK   )r>   stackflatten)rU   x1x2rot_xr2   r2   r3   rotate_half   s   r   c                 C   sj   | j }|  } | }||}||}| | t| |  }|| t||  }|j|d|j|dfS )a  Applies Rotary Position Embedding to the query and key tensors.

    Args:
        q (`torch.Tensor`): The query tensor.
        k (`torch.Tensor`): The key tensor.
        cos (`torch.Tensor`): The cosine part of the rotary embedding.
        sin (`torch.Tensor`): The sine part of the rotary embedding.
        unsqueeze_dim (`int`, *optional*, defaults to 1):
            The 'unsqueeze_dim' argument specifies the dimension along which to unsqueeze cos[position_ids] and
            sin[position_ids] so that they can be properly broadcasted to the dimensions of q and k. For example, note
            that cos[position_ids] and sin[position_ids] have the shape [batch_size, seq_len, head_dim]. Then, if q and
            k have the shape [batch_size, heads, seq_len, head_dim], then setting unsqueeze_dim=1 makes
            cos[position_ids] and sin[position_ids] broadcastable to the shapes of q and k. Similarly, if q and k have
            the shape [batch_size, seq_len, heads, head_dim], then set unsqueeze_dim=2.
    Returns:
        `tuple(torch.Tensor)` comprising of the query and key tensors rotated using the Rotary Position Embedding.
    r9   )r:   rB   	unsqueezer   rA   )qkrS   rT   unsqueeze_dimr:   q_embedk_embedr2   r2   r3   apply_rotary_pos_emb   s   

r   c                       s   e Zd ZdZddededB f fddZ		ddejde	ejejf d	ejdB d
e
dB dejdB dee de	ejejdB e	ej dB f fddZ  ZS )Cohere2Attentionz=Multi-headed attention from 'Attention Is All You Need' paperNr   	layer_idxc                    s   t    || _|| _t|d|j|j | _|j|j | _	| jd | _
|j| _d| _t|dr4|j| nd }|dkr=|jnd | _tj|j|j| j |jd| _tj|j|j| j |jd| _tj|j|j| j |jd| _tj|j| j |j|jd| _d S )Nr7   g      Tlayer_typessliding_attentionro   )r#   r$   r   r   r;   r<   r=   r7   r|   r   r   attention_dropout	is_causalhasattrr   sliding_windowri   Linearattention_biasq_projk_projv_projo_proj)r-   r   r   
layer_typer0   r2   r3   r$      s,   
zCohere2Attention.__init__ru   position_embeddingsr   past_key_valuescache_positionr   r5   c                 K   s&  |j d d }g |d| jR }| ||dd}	| ||dd}
| ||dd}|\}}| jd urGt|	|
||\}	}
|d ur\|||d}|	|
|| j
|\}
}t| jjt}|| |	|
||f| jspdn| j| j| jd|\}}|jg |dR   }| |}||fS )NrF   r   r8   )rT   rS   r   r   )r   r   r   )rM   r7   r   viewrQ   r   r   r   r   updater   r   get_interfacer   _attn_implementationr   r   r   r   rz   r   r   )r-   ru   r   r   r   r   r   input_shapehidden_shapequery_statesr   r   rS   rT   cache_kwargsattention_interfacer   r   r2   r2   r3   r[      s<   	
	

zCohere2Attention.forwardr\   )NN)r]   r^   r_   __doc__r   rc   r$   r>   r`   rd   r   
LongTensorr   r   r[   rf   r2   r2   r0   r3   r      s(    r   c                       s$   e Zd Z fddZdd Z  ZS )
Cohere2MLPc                    sr   t    || _|j| _|j| _tj| j| jdd| _tj| j| jdd| _tj| j| jdd| _	t
|j | _d S NFr   )r#   r$   r   r<   intermediate_sizeri   r   	gate_projup_proj	down_projr   
hidden_actact_fnr-   r   r0   r2   r3   r$     s   
zCohere2MLP.__init__c                 C   s$   |  | | || | }|S r\   )r   r   r   r   )r-   rU   r   r2   r2   r3   r[     s    zCohere2MLP.forwardrx   r2   r2   r0   r3   r     s    
r   c                       s   e Zd Zdedef fddZ					ddejdeejejf dB d	ejdB d
e	dB de
dB dejdB dee deejeejejf dB f fddZ  ZS )Cohere2DecoderLayerr   r   c                    sL   t    |j| _t||d| _t|| _t|j|jd| _	|j
| | _d S )N)r   r   r<   rn   )r#   r$   r<   r   	self_attnr   mlprg   layer_norm_epsinput_layernormr   attention_type)r-   r   r   r0   r2   r3   r$   !  s   

zCohere2DecoderLayer.__init__NFru   r   r   r   	use_cacher   r   r5   c              	   K   sJ   |}|  |}| jd||||||d|\}	}
| |}||	 | }|S )ar  
        Args:
            hidden_states (`torch.FloatTensor`): input to the layer of shape `(batch, seq_len, embed_dim)`
            attention_mask (`torch.FloatTensor`, *optional*):
                attention mask of size `(batch_size, sequence_length)` if flash attention is used or `(batch_size, 1,
                query_sequence_length, key_sequence_length)` if default attention is used.
            past_key_values (`Cache`, *optional*): cached past key and value projection states
            output_attentions (`bool`, *optional*):
                Whether or not to return the attentions tensors of all attention layers. See `attentions` under
                returned tensors for more detail.
            use_cache (`bool`, *optional*):
                If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding
                (see `past_key_values`).
            cache_position (`torch.LongTensor` of shape `(sequence_length)`, *optional*):
                Indices depicting the position of the input sequence tokens in the sequence
            position_embeddings (`tuple[torch.FloatTensor, torch.FloatTensor]`, *optional*):
                Tuple containing the cosine and sine positional embeddings of shape `(batch_size, seq_len, head_dim)`,
                with `head_dim` being the embedding dimension of each attention head.
        )ru   r   r   r   r   r   Nr2   )r   r   r   )r-   ru   r   r   r   r   r   r   residualhidden_states_attention_hidden_states_mlpr2   r2   r3   r[   )  s   



zCohere2DecoderLayer.forward)NNNFN)r]   r^   r_   r   rc   r$   r>   r`   rd   r   boolr   r   r   FloatTensorr[   rf   r2   r2   r0   r3   r      s0    	r   c                   @   sH   e Zd ZU eed< dZdZdgZdgZdZ	dZ
dZdZdZeedZdS )Cohere2PreTrainedModelr   modelTr   r   )ru   
attentionsN)r]   r^   r_   r   ra   base_model_prefixsupports_gradient_checkpointing_no_split_modules_skip_keys_device_placement_supports_flash_attn_supports_sdpa_supports_flex_attn_can_compile_fullgraph_supports_attention_backendr   r   _can_record_outputsr2   r2   r2   r3   r   W  s   
 
r   c                       s   e Zd Zdef fddZeee							ddej	dB dej
dB dej	dB dedB d	ejdB d
edB dej	dB dee defddZ  ZS )Cohere2Modelr   c                    s   t     j| _ j| _t j j| j| _t	 fddt
 jD | _t j jd| _t | _d| _|   d S )Nc                    s   g | ]}t  |qS r2   )r   ).0r   r   r2   r3   
<listcomp>s  s    z)Cohere2Model.__init__.<locals>.<listcomp>r   F)r#   r$   pad_token_idpadding_idx
vocab_sizeri   	Embeddingr<   embed_tokens
ModuleListrangenum_hidden_layerslayersrg   r   normr   
rotary_embgradient_checkpointing	post_initr   r0   r   r3   r$   l  s   
zCohere2Model.__init__N	input_idsr   rV   r   inputs_embedsr   r   r   r5   c              
   K   s"  |d u |d uA rt d|d u r| |}|r!|d u r!t| jd}|d u r=|d ur-| nd}	tj|	|	|jd  |jd}|d u rF|	d}t
| }
tsf| j|||||d}td
i |td
i |d}
|}| ||}| jD ]}||f|
|j |||||d|}qq| |}t||d	S )Nz:You must specify exactly one of input_ids or inputs_embedsr   r   r   )r.   )r   r   r   r   r   rV   )full_attentionr   )r   r   r   r   r   rV   )last_hidden_stater   r2   )
ValueErrorr   r   r   get_seq_lengthr>   r?   rM   r.   r   rN   dictr
   r   r   r   r   r   r   )r-   r   r   rV   r   r   r   r   r   past_seen_tokenscausal_mask_mappingmask_kwargsru   r   decoder_layerr2   r2   r3   r[   |  sX   

	

zCohere2Model.forward)NNNNNNN)r]   r^   r_   r   r$   r   r   r   r>   r   r`   r   r   r   r   r   r   r[   rf   r2   r2   r0   r3   r   j  s>    	
r   c                       s   e Zd ZddiZddiZddgdgfiZ fddZee																					
dde	j
d	B de	jd	B de	j
d	B ded	B de	jd	B de	j
d	B ded	B ded	B ded	B de	j
d	B dee	jB dee defddZ  ZS )Cohere2ForCausalLMzlm_head.weightzmodel.embed_tokens.weightlm_headcolwise_gather_outputru   logitsc                    sP   t  | t|| _|j| _tj|j|jdd| _|j	| _	|j
| _
|   d S r   )r#   r$   r   r   r   ri   r   r<   r  logit_scaletie_word_embeddingsr   r   r0   r2   r3   r$     s   
zCohere2ForCausalLM.__init__Nr   r   r   rV   r   r   labelsr   output_attentionsoutput_hidden_statesr   logits_to_keepr   r5   c                 K   s   |dur|n| j j}|	dur|	n| j j}	| jd||||||||	|
d	|}|j}t|tr4t| dn|}| |dd|ddf }|| j	 }d}|dur]| j
d||| j jd|}t|||j|j|jdS )a~  
        labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*):
            Labels for computing the masked language modeling loss. Indices should either be in `[0, ...,
            config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored
            (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`.

        Example:

        ```python
        >> from transformers import AutoTokenizer, Cohere2ForCausalLM

        >> model = Cohere2ForCausalLM.from_pretrained("Cohere2ForAI/c4ai-command-r-v01")
        >> tokenizer = AutoTokenizer.from_pretrained("Cohere2ForAI/c4ai-command-r-v01")

        >> prompt = "Hey, are you conscious? Can you talk to me?"
        >> inputs = tokenizer(prompt, return_tensors="pt")

        >> # Generate
        >> generate_ids = model.generate(inputs.input_ids, max_length=30)
        >> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
        "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you."
        ```N)	r   r   rV   r   r   r   r  r  r   )r  r  r   )lossr  r   ru   r   r2   )r   r  r  r   r   rN   rc   slicer  r  loss_functionr   r   r   ru   r   )r-   r   r   rV   r   r   r  r   r  r  r   r	  r   outputsru   slice_indicesr  r
  r2   r2   r3   r[     s<   '

zCohere2ForCausalLM.forward)NNNNNNNNNNr   )r]   r^   r_   _tied_weights_keys_tp_plan_pp_planr$   r   r   r>   r   r`   r   r   r   rc   r   r   r   r[   rf   r2   r2   r0   r3   r     sZ    	
r   )r   r   r   )r   )r   )<collections.abcr   typingr   r>   torch.nnri   activationsr   cache_utilsr   r   
generationr   integrationsr	   masking_utilsr
   r   modeling_layersr   modeling_outputsr   r   modeling_rope_utilsr   r   modeling_utilsr   r   processing_utilsr   utilsr   r   r   utils.genericr   r   utils.output_capturingr   configuration_cohere2r   Moduler   rg   r`   rc   r~   rB   r   r   r   r   r   r   r   r   r   __all__r2   r2   r2   r3   <module>   sh   A

I7T]