o
    i                     @   s   d Z ddlZddlmZ ddlmZ ddlmZmZ ddlmZ ddlmZ dd	l	m	Z	 dd
l
m
Z
 ddlmZ ddlmZ ddlmZ ddlmZ ddlmZ dee dee dee dee dee dededededefddZdS )zFunctional interface.    N)Tensor   )adadelta)_make_sparseadagrad)adam)adamax)adamw)asgd)nadam)radam)rmsprop)rprop)sgdparamsgradsexp_avgsexp_avg_sqsstate_stepsepsbeta1beta2lrmaximizec                   sR  t | D ]\}
}||
  |	s n          }| dkr&q  ||
 }||
 }||
 } fdd}|  }||d| }|	|| |  }|
d|d| }|	|| |	|}|	| | 	|}~~d||  }d||  }|t| | }|	|| ||  qdS )z{Functional API that performs Sparse Adam algorithm computation.

    See :class:`~torch.optim.SparseAdam` for details.
    r   c                    s6    j } dks|  dkr|  S || S )Nr   )newdim
resize_as_)valuesconstructorgradgrad_indicessize T/home/ubuntu/veenaModal/venv/lib/python3.10/site-packages/torch/optim/_functional.pymake_sparse8   s   z sparse_adam.<locals>.make_sparser      N)	enumeratecoalesce_indices_valuesnumelr"   sparse_masksubmul_add_powsub_sqrt_mathsqrtdiv_)r   r   r   r   r   r   r   r   r   r   iparamgrad_valuesexp_avg
exp_avg_sqstepr%   old_exp_avg_valuesexp_avg_update_valuesold_exp_avg_sq_valuesexp_avg_sq_update_valuesnumerdenombias_correction1bias_correction2	step_sizer#   r   r$   sparse_adam   s:   	

rE   )__doc__r3   torchr   r   r   r   r   r   r	   r
   r   r   r   r   r   listintfloatboolrE   r#   r#   r#   r$   <module>   sF   	
