torch.cuda.memory.get_allocator_backend# torch.cuda.memory.get_allocator_backend()[source]# Return a string describing the active allocator backend as set by PYTORCH_CUDA_ALLOC_CONF. Currently available backends are native (PyTorch’s native caching allocator) and cudaMallocAsync` (CUDA’s built-in asynchronous allocator). Note See Memory management for details on choosing the allocator backend. Return type str