0

I am trying to run train pytorch model

this is my config file batch size, I have tried reducing batch_size to 1 but the same error

local batch_size = 3,
local num_batch_accumulated = 4,

This is the output of nvidia-smi

enter image description here

As we can see only 451 mib is allocated out of 6144 Mib

I have tried following different solutions mentioned in stackoverflow post here but not able to solve this.

How can I fix this strange error: "RuntimeError: CUDA error: out of memory"?

talonmies
  • 70,661
  • 34
  • 192
  • 269
Sarde
  • 658
  • 1
  • 8
  • 19
  • Please, provide some additional information. What kind of model are you trying to train? Are you running training in Jupyter notebook or as a python script? What output does torch.cuda.is_available() give you? – Sherstnyov Jul 09 '23 at 11:00
  • I am trying to run following git project https://github.com/dmirlab-group/sadga error occurs on https://github.com/dmirlab-group/sadga#step-2-training – Sarde Jul 10 '23 at 11:39

0 Answers0