o
    Gۂid                     @   sz  d dl Z d dlZd dlZd dlZd dlZd dlmZ d dlmZ d dlm	Z	 d dl
mZ d dlmZ d dlmZ ejejeZejedgZejed	Zd
gZdae  dd Ze  dd ZG dd deZdd ZddddddZddddddZ dZ!e"e!Z#dd Z$e%dd e&d D Z'd!e'd"< d"e'd#< d#e'd!< d$d% Z(d&d' Z)G d(d) d)eZ*G d*d+ d+eZ+dS ),    N)Path)knobs)compile_module_from_src)_allocation)	GPUTarget)	GPUDriverincludeliblibcuda.so.1c                  C   s   t jj } r	| gS tddgjdd}dd | D }dd |D }td}|r7|s7d	d |	d
D }d}|rH|dt
| 7 }|d7 }n|d7 }|d7 }tdd |D s]J ||S )Nz/sbin/ldconfigz-pignore)errorsc                 S   s    g | ]}d |v r|  d qS )r
   )split).0line r   `/home/ubuntu/maya3_transcribe/venv/lib/python3.10/site-packages/triton/backends/nvidia/driver.py
<listcomp>   s     z libcuda_dirs.<locals>.<listcomp>c                 S   s   g | ]}t j|qS r   )ospathdirname)r   locr   r   r   r      s    LD_LIBRARY_PATHc                 S   s&   g | ]}t jt j|d r|qS )r
   r   r   existsjoin)r   dirr   r   r   r       s   & :zlibcuda.so cannot found!
z!Possible files are located at %s.z:Please create a symlink of libcuda.so to any of the files.z<Please make sure GPU is set up and then run "/sbin/ldconfig"z- (requires sudo) to refresh the linker cache.c                 s   s&    | ]}t jt j|d V  qdS )r
   Nr   )r   r   r   r   r   	<genexpr>(   s   $ zlibcuda_dirs.<locals>.<genexpr>)r   nvidialibcuda_path
subprocesscheck_outputdecode
splitlinesr   getenvr   strany)env_libcuda_pathlibslocsdirsenv_ld_library_pathmsgr   r   r   libcuda_dirs   s    

r.   c                   C   s   t gt S N)libdevice_dirr.   r   r   r   r   library_dirs,   s   r1   c                       s$   e Zd Z fddZdd Z  ZS )	CudaUtilsc                    s"   t | dstt| | | _| jS )Ninstance)hasattrsuperr2   __new__r3   )cls	__class__r   r   r6   8   s   
zCudaUtils.__new__c                 C   sX   t ttjtd dt tt	d}|j
a
|j| _|j| _|j| _|j| _|j| _d S )Nzdriver.c
cuda_utilssrcnamer1   include_dirs	libraries)r   r   r   r   r   r   	read_textr1   r>   r?   PyCUtensorMapload_binaryget_device_propertiescuOccupancyMaxActiveClustersset_printf_fifo_sizefill_tma_descriptor)selfmodr   r   r   __init__=   s   zCudaUtils.__init__)__name__
__module____qualname__r6   rI   __classcell__r   r   r8   r   r2   6   s    r2   c                 C   s   | d dkrdS |  drdS i ddddd	d
dddddddddddddddddddddddddd|  S )Nr   *CUdeviceptr
tensordescCUtensorMapi1int8_ti8i16int16_ti32int32_ti64int64_tu1uint8_tu8u16uint16_tu32uint32_tu64uint64_tfp16doublebf16fp32f32fp64	nvTmaDesc)
startswith)tyr   r   r   	ty_to_cppS   sN   
	
rm   r_   ra   rc   )rd   rf   rg   rh   ri   	pack_fp16	pack_bf16	pack_fp32	pack_fp64iiiKKppOOOOOOc                    s  fdd}fdd fdd fdd||  }d	d
 t|D }dfdd|  D }t| }g }|  D ]}|| q?dd
 t|D }t|dkrddddd | D  nd}	g }
| D ]&\}}|dkruql|tv r|
t|  d|  ql|
t| d|  qld|
}g }| D ];\}}|d dkr|d| d q|tv r|d| d q|dkr|d|  q|dkr|d|  qt	t|}d}dd | D }dd | D }dd | D }d d | D }|d! |d" d#t|dkrd| nd d$d| d%| fd&d| D  d'| d(|	 d)|| d|| d|| d*t|dkrZdd| nd d+}|S ),Nc           
         s  g }d}| D ]q}t |trr|drr r | nd }|d7 }td|}|d}|d}|dd }|d u rT|d|  td| D ]}	|d qF|d	 n|d
 t|D ]}	|d q]t|D ]}	|d qiq|| q r|t	 ksJ |S )Nr   rP      ztensordesc<([^[>]*)\[([^]]*)\]   ,rN   rY   rR   rj   rW   )

isinstancer&   rk   rematchgroupcountappendrangelen)
	signatureoutputtensordesc_idxsigmetarx   dtypeshapendim_)tensordesc_metar   r   _expand_signature   s0   


z(make_launcher.<locals>._expand_signaturec                    s0   t | tr| D ]} || qd S ||  d S r/   )rv   tupler{   )r   r   x)_flatten_signaturer   r   r      s
   
z)make_launcher.<locals>._flatten_signaturec                    sJ   t | trdt | }d| dS | d dkrdS | dv r!dS t| S )Nru   []r   rN   z	PyObject*	constexprrj   )rv   r   r   maprm   rl   val_extracted_typer   r   r      s   
z&make_launcher.<locals>._extracted_typec                    sr   t | trdt | }d| dS | d dkrdS | dv r!dS | dr(dS d	d
ddddddddd
t|  S )N ()r   rN   Or   rP   dlbhiLBHIK)
re   longrS   rV   rX   rZ   r\   r_   ra   rc   )rv   r   r   r   rk   rm   r   	format_ofr   r   r      s,   

z make_launcher.<locals>.format_ofc                 S      i | ]\}}||qS r   r   r   r   sr   r   r   
<dictcomp>       z!make_launcher.<locals>.<dictcomp>r   c                    s   g | ]} |qS r   r   )r   rl   r   r   r   r      s    z!make_launcher.<locals>.<listcomp>c                 S   r   r   r   r   r   r   r   r      r   r   , c                 s   s    | ]
\}}d | V  qdS )z&_argNr   r   r   rl   r   r   r   r      s    z make_launcher.<locals>.<genexpr>r   z argrN   ptr_infoz.dev_ptr_arg_storagerj   z*tma_ptrz
  c                 S   s:   g | ]\}}|d  dkrd| d| d| d| d	qS )r   rN   zDevicePtrInfo ptr_infoz = getPointer(_argr   z); if (!ptr_infoz.valid) return NULL;r   r   r   r   r   r      s
    c              	   S   s0   g | ]\}}|d krd| d| d| dqS )rj   zCUtensorMap* tma_ptrz = getTmaDesc(_argz); if (!tma_ptrz) return NULL;r   r   r   r   r   r      s
    c              
   S   s<   g | ]\}}|t v rt |  d | dt|  d| dqS ) _argz_storage = z(_argz);)FLOAT_STORAGE_TYPEFLOAT_PACK_FUNCTIONr   r   r   r   r      s
    "c                 S   s"   g | ]\}}|d krd| qS )r   z&argr   r   r   r   r   r     s   " z&global_scratchz&profile_scratcha  
#include "cuda.h"
#include <dlfcn.h>
#include <stdbool.h>
#include <stdlib.h>
#define PY_SSIZE_T_CLEAN
#include <Python.h>

typedef struct {
  PyObject_HEAD;
  _Alignas(128) CUtensorMap tensorMap;
} PyCUtensorMapObject;

static inline void gpuAssert(CUresult code, const char *file, int line)
{
   if (code != CUDA_SUCCESS)
   {
      const char* prefix = "Triton Error [CUDA]: ";
      const char* str;
      cuGetErrorString(code, &str);
      char err[1024] = {0};
      strcat(err, prefix);
      strcat(err, str);
      PyGILState_STATE gil_state;
      gil_state = PyGILState_Ensure();
      PyErr_SetString(PyExc_RuntimeError, err);
      PyGILState_Release(gil_state);
   }
}

#define CUDA_CHECK(ans) { gpuAssert((ans), __FILE__, __LINE__); }

typedef CUresult (*cuLaunchKernelEx_t)(const CUlaunchConfig* config, CUfunction f, void** kernelParams, void** extra);

static cuLaunchKernelEx_t getLaunchKernelExHandle() {
  // Open the shared library
  void* handle = dlopen("libcuda.so.1", RTLD_LAZY);
  if (!handle) {
    PyErr_SetString(PyExc_RuntimeError, "Failed to open libcuda.so.1");
    return NULL;
  }
  // Clear any existing error
  dlerror();
  cuLaunchKernelEx_t cuLaunchKernelExHandle = (cuLaunchKernelEx_t)dlsym(handle, "cuLaunchKernelEx");
  // Check for errors
  const char *dlsym_error = dlerror();
  if (dlsym_error) {
    PyErr_SetString(PyExc_RuntimeError, "Failed to retrieve cuLaunchKernelEx from libcuda.so.1");
    return NULL;
  }
  return cuLaunchKernelExHandle;
}

static void _launch(int gridX, int gridY, int gridZ, int num_warps, int num_ctas, int launch_cooperative_grid, int launch_pdl, int shared_memory, CUstream stream, CUfunction function, CUdeviceptr global_scratch, CUdeviceptr profile_scratchz) {
  void *params[] = { au   };
  if (gridX*gridY*gridZ > 0) {
    // 4 attributes that we can currently pass maximum
    CUlaunchAttribute launchAttr[4];
    static cuLaunchKernelEx_t cuLaunchKernelExHandle = NULL;
    if (cuLaunchKernelExHandle == NULL) {
      cuLaunchKernelExHandle = getLaunchKernelExHandle();
    }
    CUlaunchConfig config;
    config.gridDimX = gridX * num_ctas;
    config.gridDimY = gridY;
    config.gridDimZ = gridZ;

    config.blockDimX = 32 * num_warps;
    config.blockDimY = 1;
    config.blockDimZ = 1;
    config.sharedMemBytes = shared_memory;
    config.hStream = stream;
    config.attrs = launchAttr;
    int num_attrs = 0;

    if (launch_pdl != 0) {
      CUlaunchAttribute pdlAttr = { .id = CU_LAUNCH_ATTRIBUTE_PROGRAMMATIC_STREAM_SERIALIZATION, .value = 1};
      launchAttr[num_attrs] = pdlAttr;
      ++num_attrs;
    }

    if (launch_cooperative_grid != 0) {
      CUlaunchAttribute coopAttr = { .id = CU_LAUNCH_ATTRIBUTE_COOPERATIVE, .value = 1};
      launchAttr[num_attrs] = coopAttr;
      ++num_attrs;
    }

    if (num_ctas != 1) {
      CUlaunchAttribute clusterAttr = {};
      clusterAttr.id = CU_LAUNCH_ATTRIBUTE_CLUSTER_DIMENSION;
      clusterAttr.value.clusterDim.x = num_ctas;
      clusterAttr.value.clusterDim.y = 1;
      clusterAttr.value.clusterDim.z = 1;
      launchAttr[num_attrs] = clusterAttr;
      ++num_attrs;

      CUlaunchAttribute clusterSchedulingAttr = {};
      clusterSchedulingAttr.id = CU_LAUNCH_ATTRIBUTE_CLUSTER_SCHEDULING_POLICY_PREFERENCE;
      clusterSchedulingAttr.value.clusterSchedulingPolicyPreference = CU_CLUSTER_SCHEDULING_POLICY_SPREAD;
      launchAttr[num_attrs] = clusterSchedulingAttr;
      ++num_attrs;
    }

    // num_ctas == 16 is non-portable. Does work for H100 and B200 tho
    config.numAttrs = num_attrs;
    if (num_ctas == 16) {
      CUDA_CHECK(cuFuncSetAttribute(
          function,
          CU_FUNC_ATTRIBUTE_NON_PORTABLE_CLUSTER_SIZE_ALLOWED,
          1
      ));
    }

    CUDA_CHECK(cuLaunchKernelExHandle(&config, function, params, 0));
  }
}

typedef struct _DevicePtrInfo {
    CUdeviceptr dev_ptr;
    bool valid;
} DevicePtrInfo;

static PyObject* data_ptr_str = NULL;
static PyObject* py_tensor_map_type = NULL;

static inline DevicePtrInfo getPointer(PyObject *obj, int idx) {
  DevicePtrInfo ptr_info;
  ptr_info.dev_ptr = 0;
  ptr_info.valid = true;
  if (PyLong_Check(obj)) {
    ptr_info.dev_ptr = PyLong_AsUnsignedLongLong(obj);
    return ptr_info;
  }
  if (obj == Py_None) {
    // valid nullptr
    return ptr_info;
  }
  PyObject *ret = PyObject_CallMethodNoArgs(obj, data_ptr_str);
  if (!ret) {
    PyErr_SetString(PyExc_TypeError, "Pointer argument must be either uint64 or have data_ptr method");
    ptr_info.valid = false;
    goto cleanup;
  }
  if (!PyLong_Check(ret)) {
    PyErr_SetString(PyExc_TypeError, "data_ptr method of Pointer object must return 64-bit int");
    ptr_info.valid = false;
    goto cleanup;
  }
  ptr_info.dev_ptr = PyLong_AsUnsignedLongLong(ret);
  if(!ptr_info.dev_ptr)
    return ptr_info;
  uint64_t dev_ptr;
  int status = cuPointerGetAttribute(&dev_ptr, CU_POINTER_ATTRIBUTE_DEVICE_POINTER, ptr_info.dev_ptr);
  if (status == CUDA_ERROR_INVALID_VALUE) {
      PyErr_Format(PyExc_ValueError,
                   "Pointer argument (at %d) cannot be accessed from Triton (cpu tensor?)", idx);
      ptr_info.valid = false;
  } else if (status != CUDA_SUCCESS) {
      CUDA_CHECK(status);  // Catch any other cuda API errors
      ptr_info.valid = false;
  }
  ptr_info.dev_ptr = dev_ptr;
cleanup:
  Py_XDECREF(ret);
  return ptr_info;

}

static inline CUtensorMap* getTmaDesc(PyObject *obj) {
  if (sizeof(CUtensorMap*) != 8) {
    PyErr_SetString(PyExc_SystemError, "getTmaDesc() requires 64-bit compilation");
    return NULL;
  }

if (Py_TYPE(obj) != (PyTypeObject*)py_tensor_map_type) {
    PyErr_Format(PyExc_TypeError, "object must be of type PyCUtensorMap, got %s", Py_TYPE(obj)->tp_name);
    return NULL;
}

  CUtensorMap* map = &((PyCUtensorMapObject*)obj)->tensorMap;
  uintptr_t align_128 = (uintptr_t)map & (128 - 1);
  if (align_128 != 0) {
    PyErr_Format(PyExc_ValueError, "CUtensorMap must be aligned to 128B, but got (&map) mod 128 = %ld", align_128);
    return NULL;
  }
  return map;
}

static void ensureCudaContext() {
  CUcontext pctx;
  CUDA_CHECK(cuCtxGetCurrent(&pctx));
  if (!pctx) {
    // Ensure device context.
    CUdevice device;
    CUDA_CHECK(cuDeviceGet(&device, 0));
    CUDA_CHECK(cuDevicePrimaryCtxRetain(&pctx, device));
    CUDA_CHECK(cuCtxSetCurrent(pctx));
  }
}

static uint16_t pack_fp16(double f) {
    uint16_t result;
    // from https://github.com/python/pythoncapi-compat
#if 0x030600B1 <= PY_VERSION_HEX && PY_VERSION_HEX <= 0x030B00A1 && !defined(PYPY_VERSION)
    _PyFloat_Pack2(f, (unsigned char*)&result, 1);
#else
    PyFloat_Pack2(f, (unsigned char*)&result, 1);
#endif
    return result;
}

static uint16_t pack_bf16(double f) {
    float f32 = (float)f;
    uint32_t u32 = *(uint32_t*)&f32;
    return (uint16_t)(u32 >> 16);
}

static uint32_t pack_fp32(double f) {
    float f32 = (float)f;
    return *(uint32_t*)&f32;
}

static uint64_t pack_fp64(double f) {
    return *(uint64_t*)&f;
}

static PyObject* launch(PyObject* self, PyObject* args) {
  // ensure cuda context is valid before calling any CUDA APIs, e.g. before getPointer calls cuPointerGetAttributes
  ensureCudaContext();

  int gridX, gridY, gridZ;
  uint64_t _stream;
  uint64_t _function;
  int launch_cooperative_grid;
  int launch_pdl;
  PyObject *launch_enter_hook = NULL;
  PyObject *launch_exit_hook = NULL;
  PyObject *kernel_metadata = NULL;
  PyObject *launch_metadata = NULL;
  PyObject *global_scratch_obj = NULL;
  PyObject *profile_scratch_obj = NULL;
  c                    s$   g | ]\}} | d | dqS )r   ;r   r   r   r   r   r     s   $ z
  if(!PyArg_ParseTuple(args, "aM  ", &gridX, &gridY, &gridZ,
                                           &_stream, &_function, &launch_cooperative_grid, &launch_pdl, &global_scratch_obj, &profile_scratch_obj,
                                           &kernel_metadata, &launch_metadata,
                                           &launch_enter_hook, &launch_exit_hooka   )) {
    return NULL;
  }

  int num_warps, num_ctas, shared_memory;
  if (!PyArg_ParseTuple(kernel_metadata, "iii", &num_warps, &num_ctas, &shared_memory)) {
    PyErr_SetString(PyExc_TypeError, "kernel_metadata must be a tuple");
    return NULL;
  }

  // extract launch metadata
  if (launch_enter_hook != Py_None){
    PyObject* ret = PyObject_CallOneArg(launch_enter_hook, launch_metadata);
    if (!ret)
      return NULL;
    Py_DECREF(ret);
  }

  CUdeviceptr global_scratch = 0;
  if (global_scratch_obj != Py_None) {
    DevicePtrInfo global_scratch_info = getPointer(global_scratch_obj, -1);
    if (!global_scratch_info.valid) {
      return NULL;
    }
    global_scratch = global_scratch_info.dev_ptr;
  }

  CUdeviceptr profile_scratch = 0;
  if (profile_scratch_obj != Py_None) {
    DevicePtrInfo profile_scratch_info = getPointer(profile_scratch_obj, -1);
    if (!profile_scratch_info.valid) {
      return NULL;
    }
    profile_scratch = profile_scratch_info.dev_ptr;
  }

  // raise exception asap
  z
  Py_BEGIN_ALLOW_THREADS;
  _launch(gridX, gridY, gridZ, num_warps, num_ctas, launch_cooperative_grid, launch_pdl, shared_memory, (CUstream)_stream, (CUfunction)_function, global_scratch, profile_scratchap  );
  Py_END_ALLOW_THREADS;
  if (PyErr_Occurred()) {
    return NULL;
  }

  if(launch_exit_hook != Py_None){
    PyObject* ret = PyObject_CallOneArg(launch_exit_hook, launch_metadata);
    if (!ret)
      return NULL;
    Py_DECREF(ret);
  }

  Py_RETURN_NONE;
}

static PyMethodDef ModuleMethods[] = {
  {"launch", launch, METH_VARARGS, "Entry point for all kernels with this signature"},
  {NULL, NULL, 0, NULL} // sentinel
};

static struct PyModuleDef ModuleDef = {
  PyModuleDef_HEAD_INIT,
  "__triton_launcher",
  NULL, //documentation
  -1, //size
  ModuleMethods
};

PyMODINIT_FUNC PyInit___triton_launcher(void) {
  data_ptr_str = PyUnicode_InternFromString("data_ptr");
  if(data_ptr_str == NULL) {
    return NULL;
  }
  PyObject* driver_mod = PyImport_ImportModule("triton.backends.nvidia.driver");
  if (driver_mod == NULL) {
    return NULL;
  }
  py_tensor_map_type = PyObject_GetAttrString(driver_mod, "PyCUtensorMap");
  if (py_tensor_map_type == NULL) {
    return NULL;
  }

  PyObject *m = PyModule_Create(&ModuleDef);
  if(m == NULL) {
    return NULL;
  }
  PyModule_AddFunctions(m, ModuleMethods);
  return m;
}
)
values	enumerater   _BASE_ARGS_FORMATr}   itemsr   r{   rm   r|   )	constantsr~   r   r   expand_signatureargs_formatformatflat_signaturer   	args_listarg_decl_listr   rl   	arg_declsinternal_args_listparamsnewline	ptr_decls	tma_declsfloat_storage_declsr<   r   )r   r   r   r   r   make_launcher   s   '
,


56 r  s  v                    Sr   c                 c   s    | ]}||fV  qd S r/   r   )r   r   r   r   r   r   \  s    r      
      	   c              
   C   s   |d u r| j g| j| j| jdk| j| jS |d }|d }|d }|d }|d }| j}| j}|d dks;J | jdkrBdnd	}	|rRt|}|d  d
9  < tjjjj	
| j  ||t| ||||	}
|
g||S )Nnanswizzle	elem_size	elem_type
block_size
fp4_paddedr   rs   r   rt   )baser   stridespaddinglisttritonruntimedriveractiveutilsrF   data_ptrTMA_DTYPE_DEVICE_TO_HOST)argmetadatar   r   r   r   r   r   r   r   cu_tensor_mapr   r   r   make_tensordesc_argb  s2   *r   c                    sv   t dd | D }|s S tdd t| D r(ttks(J s1d gt  fdd}|S )Nc                 s   s$    | ]}t |to|d V  qdS )rP   Nrv   r&   rk   )r   r   r   r   r   r     s   " z)wrap_handle_tensordesc.<locals>.<genexpr>c                 S   s(   g | ]\}}t |tr|d r|qS )rP   r   )r   r   r   r   r   r   r     s   ( z*wrap_handle_tensordesc.<locals>.<listcomp>c                     sf   t | d t }d}t| td  D ]\}}|v r)|t||  |d7 }q|| q | S )Nr   rs   )r   _BASE_ARGS_FORMAT_LENr   extendr   r{   )args
final_argsr   r   r   launchertensordesc_indicesr   r   r   inner  s   
z%wrap_handle_tensordesc.<locals>.inner)r'   r   setr   r}   )r   r~   r   has_tensor_desc_argr   r   r   r   wrap_handle_tensordesc  s   r   c                   @   s   e Zd Zdd Zdd ZdS )CudaLauncherc                    s   t drjnt }fdd  fdd| D }dd j D }t|dd }t|||tdt t	t
d	}t|d
d| _t|j||| _|j| _|j| _|j| _|j| _|j| _|j| _d S )Nr   c                    s   t | tr jj| fS | S r/   )rv   r&   fn	arg_namesindex)r   )r<   r   r   <lambda>  s    z'CudaLauncher.__init__.<locals>.<lambda>c                    s   i | ]	\}} ||qS r   r   r   idxvalue)arg_idxr   r   r     s    z)CudaLauncher.__init__.<locals>.<dictcomp>c                 S   r   r   r   r   r   r   r   r     r   r   __triton_launcherr;   num_ctasrs   )r4   r   dictr   r~   getattrr   r   r1   r>   r?   r   r   launchglobal_scratch_sizeglobal_scratch_alignprofile_scratch_sizeprofile_scratch_alignlaunch_cooperative_grid
launch_pdl)rG   r<   r   r   r~   r   rH   r   )r   r<   r   rI     s*   zCudaLauncher.__init__c           
   
      sb    fdd}|j jtj}|jjtj}	j |jj	||	g	|R   d S )Nc                    s:   | dkr   }|j  |  }| }|||S d S Nr   )r   get)sizealign	allocator	grid_size
alloc_sizealloc_fngridXgridYgridZrG   streamr   r   allocate_scratch  s   z/CudaLauncher.__call__.<locals>.allocate_scratch)
r   r   r   
_allocatorr   r   _profile_allocatorr   r   r   )
rG   r	  r
  r  r  functionr   r  global_scratchprofile_scratchr   r  r   __call__  s   
zCudaLauncher.__call__N)rJ   rK   rL   rI   r  r   r   r   r   r     s    r   c                       sj   e Zd Z fddZdd Zdd Zdd Zed	d
 Zde	de	fddZ
dd Zdd Zdd Z  ZS )
CudaDriverc                    s   t  | _t| _t   d S r/   )r2   r   r   launcher_clsr5   rI   )rG   r8   r   r   rI     s   zCudaDriver.__init__c                 C   s6   |   }| |}|d d |d  }d}td||S )Nr   r   rs       cuda)get_current_deviceget_device_capabilityr   )rG   device
capability	warp_sizer   r   r   get_current_target  s
   
zCudaDriver.get_current_targetc                 C   s   dd l }|d|  S )Nr   r  )torchr  r  rG   r  r   r   r   get_active_torch_device  s   z"CudaDriver.get_active_torch_devicec                 C   s   dd l }|jS r   )r  r  r  r   r   r   get_device_interface  s   zCudaDriver.get_device_interfacec                  C   s6   zdd l } | j o| jjd u W S  ty   Y dS w )Nr   F)r  r  is_availableversionhipImportError)r  r   r   r   	is_active  s   zCudaDriver.is_activerl   returnc                 C   s   t |S r/   )rm   )rG   rl   r   r   r   map_python_to_cpp_type  s   z!CudaDriver.map_python_to_cpp_typec                 C   s   ddl m} |S )Nr   )do_bench)triton.testingr)  )rG   r)  r   r   r   get_benchmarker  s   zCudaDriver.get_benchmarkerc                 C   s&   dd l }d}|jt|d |jddS )Nr   i      r  )r   r  )r  emptyint)rG   r  
cache_sizer   r   r   get_empty_cache_for_benchmark  s   z(CudaDriver.get_empty_cache_for_benchmarkc                 C   s   |   d S r/   )zero_)rG   cacher   r   r   clear_cache  s   zCudaDriver.clear_cache)rJ   rK   rL   rI   r  r   r!  staticmethodr&  r&   r(  r+  r0  r3  rM   r   r   r8   r   r    s    
	r  ),	functoolsr   r!   r   rw   pathlibr   r   triton.runtime.buildr   triton.runtimer   triton.backends.compilerr   triton.backends.driverr   r   r   realpath__file__r   r>   r0   r?   rA   	lru_cacher.   r1   objectr2   rm   r   r   r   r}   r   r   r   r|   r   r   r   r   r  r   r   r   r   <module>   s`    

	   `'+