How to set max_split_size_mb?

#30
by neo-benjamin - opened

I am running this model with two GPU. I get the following error:

OutOfMemoryError: CUDA out of memory. Tried to allocate 2.37 GiB (GPU 0; 23.69 GiB total capacity; 
19.71 GiB already allocated; 914.75 MiB free; 21.86 GiB reserved in total by PyTorch) 
If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation.  
See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF

how to split workload to multiple GPUs?

Sign up or log in to comment