o
    "iG                  	   @   sT  U d dl Z d dlZd dlZd dlZd dlmZ d dlmZ d dlm	Z	m
Z
mZmZmZmZmZ d dlZd dlmZ d dlmZ d dlmZ d dlmZmZ d d	lmZ d d
lmZ d dlmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z& d dl'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z. d dl/m0Z0m1Z1m2Z2m3Z3 d dl4m5Z5 e6e7Z8ej9e:d< g dZ;G dd de,Z<G dd de)Z=G dd de=Z>	d4dee?e	f de!de@de(fddZAdee( dee( fdd ZBdee?e	f d!e@de+fd"d#ZC	d4dee+ d$e@deee+ e!f fd%d&ZDde#de!fd'd(ZEd)e d*e de@fd+d,ZFd-ejGd.e de@fd/d0ZHd1ee+ de!de@fd2d3ZIdS )5    N)ChainMap)reduce)AnycastDictListOptionalTupleUnion)narrow_tensor_by_index)DTensor)dedup_save_plans)FLATTEN_MAPPINGflatten_state_dict)_flatten_sharded_tensors)set_element)BytesStorageMetadataChunkStorageMetadataMetadataMetadataIndexSTATE_DICT_TYPESTORAGE_TYPESStorageMetaTensorStorageMetadata)LoadPlanLoadPlannerReadItemSavePlanSavePlanner	WriteItemWriteItemType)"_create_default_metadata_only_plan_create_read_items_create_write_items_init_state_dict)find_state_dict_objectlogger)DefaultSavePlannerDefaultLoadPlannercreate_default_local_load_plancreate_default_global_load_plancreate_default_local_save_plancreate_default_global_save_planc                   @   s   e Zd ZU eed< 				d"dededee ded	df
d
dZ		d#dedee	 ded	dfddZ
d	efddZdee d	eee ef fddZded	efddZded	eejejf fddZded	efddZdedefd d!ZdS )$r'   mappingsTNFr   flatten_sharded_tensorsdedup_replicated_tensorsdedup_save_to_lowest_rankreturnc                 C   s2   || _ || _i | _|| _|d urtd d S d S )NzDefaultSavePlanner's `dedup_replicated_tensors` argument is being deprecated, and no longer has any effect. Please remove this argument from your call.)r   r.   r-   r0   r&   warning)selfr   r.   r/   r0    r4   j/home/ubuntu/SoloSpeech/.venv/lib/python3.10/site-packages/torch/distributed/checkpoint/default_planner.py__init__B   s   zDefaultSavePlanner.__init__
state_dictstorage_metais_coordinatorc                 C   s2   | j r
t |\}| _| jrt|}|| _|| _d S N)r   r-   r.   r   r7   r9   )r3   r7   r8   r9   r4   r4   r5   set_up_plannerT   s   
z!DefaultSavePlanner.set_up_plannerc                 C   s0   t | j| j}| jrtj|| jd}|| _| jS )Nplanner_data)r+   r7   r9   r   dataclassesreplacer-   plan)r3   r@   r4   r4   r5   create_local_plana   s
   z$DefaultSavePlanner.create_local_plan	all_plansc                 C   sp   t || j}t|\}}| jr#dd |D }tt| }tj||d}t||s,t	d|| _
|| _| j
| jfS )Nc                 S   s   g | ]}|j qS r4   r<   ).0pr4   r4   r5   
<listcomp>u   s    z9DefaultSavePlanner.create_global_plan.<locals>.<listcomp>r<   zFailed to validate global plan)r   r0   r,   r   dictr   r>   r?   _validate_global_plan
ValueErrorglobal_planmetadata)r3   rB   rI   rJ   planner_data_dictmerged_mappingsr4   r4   r5   create_global_plani   s   
z%DefaultSavePlanner.create_global_plannew_planc                 C   s
   || _ |S r:   )r@   r3   rN   r4   r4   r5   finish_plan   s   zDefaultSavePlanner.finish_plan
write_itemc                 C      |  |j}| ||S r:   )lookup_objectindextransform_object)r3   rQ   objectr4   r4   r5   resolve_data      zDefaultSavePlanner.resolve_datarT   c                 C      t | j|S zSExtension from the planner interface to make it easy to extend the default planner.r%   r7   r3   rT   r4   r4   r5   rS         z DefaultSavePlanner.lookup_objectrV   c                 C   s(   |j tjkrt }t|| |}|S rZ   )typer    BYTE_IOioBytesIOtorchsave)r3   rQ   rV   bytesr4   r4   r5   rU      s
   z#DefaultSavePlanner.transform_object)TTNFNF)__name__
__module____qualname__r   __annotations__boolr   r6   r   r   r;   r   rA   r   r	   r   rM   rP   r   r
   rb   Tensorr`   ra   rW   r   r   rS   rU   r4   r4   r4   r5   r'   ?   sL   
 


r'   c                	   @   s  e Zd ZU dZeed< eed< 			d(dededed	d
fddZ	
	d)dede	e
 ded	d
fddZd	efddZdee d	ee fddZded	efddZdedejd	d
fddZdefddZded ejd	d
fd!d"Zd#ed	ejfd$d%Zded ejfd&d'Zd
S )*r(   ak  
    DefaultLoadPlanner that adds multiple features on top of LoadPlanner.

    In particular it adds the following:

    flatten_state_dict: Handle state_dict with nested dicts
    flatten_sharded_tensors: For FSDP in 2D parallel mode
    allow_partial_load: If False, will raise a runtime error if a key is present in state_dict, but not in the checkpoint.
    original_state_dictr-   TFr   r.   allow_partial_loadr1   Nc                 C   s"   || _ || _i | _i | _|| _d S r:   )r   r.   rl   r-   rm   )r3   r   r.   rm   r4   r4   r5   r6      s
   
zDefaultLoadPlanner.__init__r7   rJ   r9   c                 C   sF   t | || _| jrt|}| jrt|\}| _|| _|| _|| _d S r:   )	r$   rl   r.   r   r   r-   r7   rJ   r9   )r3   r7   rJ   r9   r4   r4   r5   r;      s   
z!DefaultLoadPlanner.set_up_plannerc                 C   s"   | j d usJ t| j| j | j S r:   )rJ   r)   r7   rm   r3   r4   r4   r5   rA      s   z$DefaultLoadPlanner.create_local_planrI   c                 C   s   t |S r:   )r*   )r3   rI   r4   r4   r5   rM      s   z%DefaultLoadPlanner.create_global_planrN   c                 C   s   |S r:   r4   rO   r4   r4   r5   rP         zDefaultLoadPlanner.finish_plan	read_itemvaluec                 C   sH   | j rt| j| j|jj tj|dd d S tj|dd| j|jj< d S )NF)weights_only)	r   r   rl   r-   
dest_indexfqnrb   loadr7   )r3   rp   rq   r4   r4   r5   
load_bytes   s   zDefaultLoadPlanner.load_bytesc                 C   rR   r:   )lookup_tensorrs   transform_tensorr3   rp   tensorr4   r4   r5   resolve_tensor   rX   z!DefaultLoadPlanner.resolve_tensorrz   c                 C   s   d S r:   r4   ry   r4   r4   r5   commit_tensor   ro   z DefaultLoadPlanner.commit_tensorrT   c                 C   rY   rZ   r[   r\   r4   r4   r5   rw      r]   z DefaultLoadPlanner.lookup_tensorc                 C   s   t ||j|jS rZ   )r   dest_offsetslengthsry   r4   r4   r5   rx      s   z#DefaultLoadPlanner.transform_tensor)TTFre   )rf   rg   rh   __doc__r   ri   r   rj   r6   r   r   r;   r   rA   r   rM   rP   r   r`   ra   rv   r{   rb   rk   r|   r   rw   rx   r4   r4   r4   r5   r(      sF   
 


r(   c                	       s`   e Zd ZdZd fdd	Zdededefdd	Z		
dde	de
e deddf fddZ  ZS )_EmptyStateDictLoadPlannera  
    Extension of DefaultLoadPlanner, which rebuilds state_dict from the saved metadata.
    Useful for loading in state_dict without first initializing a model, such as
    when converting a DCP checkpoint into a Torch save file.

    . N.B. `state_dict` must be an empty dictionary when used with this LoadPlanner

    .. warning::
        Because the entire state dict is initialized, It's recommended to only utilize
        this LoadPlanner on a single rank or process to avoid OOM.

    Nc                    s   || _ t j|i | d S r:   )keyssuperr6   )r3   r   argskwargs	__class__r4   r5   r6      s   z#_EmptyStateDictLoadPlanner.__init__keyrJ   r1   c                    s~    j d u rdS | j v r	 g }|j|}|D ]}|r*|d|d t|g q|| qt fdd|D r=dS dS )NT.c                 3   s    | ]}| j v V  qd S r:   )r   )rC   unflattened_keyrn   r4   r5   	<genexpr>  s    zA_EmptyStateDictLoadPlanner._should_include_key.<locals>.<genexpr>F)r   r=   getappendjoinstrany)r3   r   rJ   unflattened_keysr=   r   r4   rn   r5   _should_include_key   s   

z._EmptyStateDictLoadPlanner._should_include_keyFr7   r9   c                    s   |rJ |d us
J |j  D ]-\}}| ||sqt|tr)tj|j|jj	d}||j
v r8t||j
| | q|||< qt ||| d S )N)dtype)state_dict_metadataitemsr   
isinstancer   rb   emptysize
propertiesr   r=   r   r   r;   )r3   r7   rJ   r9   kvr   r4   r5   r;     s   


z)_EmptyStateDictLoadPlanner.set_up_plannerr:   re   )rf   rg   rh   r   r6   r   r   rj   r   r   r   r;   __classcell__r4   r4   r   r5   r      s    r   Tr7   rJ   strictr1   c                 C   s   g }	 |   D ]6\}}||jvr|rtd| dq|j| }t|tr5|j d ur4|t|||7 }q|t|||7 }qt|S )Nz&Missing key in checkpoint state_dict: r   )	r   r   RuntimeErrorr   r   device_meshget_coordinater"   r   )r7   rJ   r   requestsrt   objmdr4   r4   r5   r)   ,  s   



r)   rB   c                 C   s   | S )z
    Create global load plan used by DefaultLoadPlanner.

    The default load behavior involved no global coordination and this function
    currently doesn't change the local plans.
    r4   )rB   r4   r4   r5   r*   N  s   	r*   r9   c                 C   sT   g }|   D ]\}}t|tr|j dur|t||7 }q|t||7 }qt|S )a  
    Create the ``SavePlan`` used by DefaultSavePlanner.

    On non-coordinator ranks, this function ignores tensors and non-tensor objects,
    only producing writes for ShardedTensor objects.

    On the coordinator rank, produce writes for all values.
    N)r   r   r   r   r   r#   r   )r7   r9   r   rt   r   r4   r4   r5   r+   Z  s   
r+   rewrite_index_hintsc           
      C   s   i }g }| D ]}g }|j D ]q}|jtjks|jj|vsJ |jtjkr0t ||jj< || q|j	dus7J t
t||jjt|j	j|j	jg d}|}|ratj|jt|jd}	tj||	d}|| |j	jdusvJ d|jj d|j|j	j q|tj||d q|t|fS )a6  
    Create the global plan and metadata used by DefaultSavePlanner.

    Metadata is produced by concatenating the metadata of all ``WriteItem`` from the supplied plans.

    The only global planning change is to update index hints in all ``MetadataIndex`` objects if
    ``rewrite_index_hints`` is True.
    N)r   r   chunks)rT   zZ
                    Cannot create MD for tensor without bounds.
                    FQN: z
                )r   )r   r^   r    SHARDrT   rt   r_   r   r   tensor_datar   r   
setdefaultr   r   r>   r?   lenr   chunkr   )
rB   r   r   	new_plansr@   	new_itemsitem	tensor_mdnew_item	new_indexr4   r4   r5   r,   u  sJ   

r,   c                 C   s   t | }t|g\}}|S )zTReturn the ``Metadata`` if DefaultSavePlanner was used to checkpoint ``state_dict``.)r!   r,   )r7   r@   _r   r4   r4   r5   _create_default_local_metadata  s   r   box0box1c                 C   sd   t | j}t|D ]&}| j| |j| |j|  kr dS |j| | j| | j|  kr/ dS q	dS )z9Check if two boxes overlap. Tuples are (offset, lengths).FT)r   offsetsrangesizes)r   r   ndimsir4   r4   r5   _check_box_overlap  s   
r   outer_box_size	inner_boxc                 C   s`   t t| D ]'}|j| dk r dS |j| dk r dS |j| |j|  | | kr- dS qdS )Nr   FT)r   r   r   r   )r   r   r   r4   r4   r5   _check_box_bounds  s   r   rI   c           
   	   C   s   d}|j  D ]j\}}t|trqt|jdkrqd}t|jD ]:\}}t|j|s5t	
d||j| d}|ttj|jd7 }|j|d d  D ]}t||rYt	
d||| d}qHq ttj|jd}	||	krqt	
d||	| d}q|S )NTr   z~
                        key:%s has out of bounds chunk:
                        tensor-size:%s chunk: %s
                    F   z$key:%s has overlapping chunks: %s %szq
                    key:%s invalid fill tensor-volume:
                    %s chunks-volume: %s
                )r   r   r   r   r   r   	enumerater   r   r&   r2   r   operatormulr   r   )
rI   rJ   all_goodr   rq   chunks_volume	chunk_idxchunk0chunk1tensor_volumer4   r4   r5   rG     sH   
	
	rG   )T)Jr>   r`   loggingr   collectionsr   	functoolsr   typingr   r   r   r   r   r	   r
   rb   torch.distributed._shard._utilsr   torch.distributed._tensorr   .torch.distributed.checkpoint._dedup_save_plansr   )torch.distributed.checkpoint._nested_dictr   r   2torch.distributed.checkpoint._sharded_tensor_utilsr   &torch.distributed.checkpoint._traverser   %torch.distributed.checkpoint.metadatar   r   r   r   r   r   r   r   $torch.distributed.checkpoint.plannerr   r   r   r   r   r   r    ,torch.distributed.checkpoint.planner_helpersr!   r"   r#   r$   "torch.distributed.checkpoint.utilsr%   	getLoggerrf   r&   Loggerri   __all__r'   r(   r   r   rj   r)   r*   r+   r,   r   r   Sizer   rG   r4   r4   r4   r5   <module>   s~   
$($
	WUB

"



7
