o
    ̳i(                  
   @   s   d dl mZ d dlmZ d dlmZmZ d dlmZ ee dee de e e e e e e d	Z	dee
ef fd	d
ZdS )    )Dict)	Replicate)ColwiseParallelRowwiseParallel)ParallelStyle)input_layouts)output_layouts)	tok_embeddingsoutputzlayers.*.attn.q_projzlayers.*.attn.k_projzlayers.*.attn.v_projzlayers.*.attn.output_projzlayers.*.mlp.w1zlayers.*.mlp.w2zlayers.*.mlp.w3returnc                   C   s   t S )z
    Helper function to get the base tensor parallel plan for Llama3 model, which will also be shared with 3.1, 3.2, and 3.3 models

    Returns:
        Dict[str, Any]: The tensor parallel plan for Llama3 model.
    )BASE_LLAMA_TP_PLAN r   r   X/home/ubuntu/.local/lib/python3.10/site-packages/torchtune/models/llama3/_parallelism.pybase_llama_tp_plan   s   r   N)typingr   torch.distributed._tensorr   !torch.distributed.tensor.parallelr   r   'torch.distributed.tensor.parallel.styler   r   strr   r   r   r   r   <module>   s   

