| 
    PyTorch
    
   | 
 
#include <cstdint>#include <cuda_runtime_api.h>#include <cusparse.h>#include <cublas_v2.h>#include <ATen/core/ATenGeneral.h>#include <ATen/Context.h>#include <c10/cuda/CUDAStream.h>#include <c10/cuda/CUDAFunctions.h>#include <ATen/cuda/Exceptions.h>Go to the source code of this file.
Namespaces | |
| namespace | at | 
| namespace | at::cuda | 
Functions | |
| int64_t | at::cuda::getNumGPUs () | 
| DEPRECATED: use device_count() instead.  More... | |
| bool | at::cuda::is_available () | 
| CUDA is available if we compiled with CUDA, and there are one or more devices.  More... | |
| cudaDeviceProp * | at::cuda::getCurrentDeviceProperties () | 
| int | at::cuda::warp_size () | 
| cudaDeviceProp * | at::cuda::getDeviceProperties (int64_t device) | 
| bool | at::cuda::canDeviceAccessPeer (int64_t device, int64_t peer_device) | 
| Allocator * | at::cuda::getCUDADeviceAllocator () | 
| cusparseHandle_t | at::cuda::getCurrentCUDASparseHandle () | 
| cublasHandle_t | at::cuda::getCurrentCUDABlasHandle () | 
| void | at::cuda::clearCublasWorkspaces () |