Cuda out of memory

#40
by comet24082002 - opened

I followed the finetuning example on github to finetune my model for Vietnamese. Then I finetuned it using these parameter but I got error CUDA out of memory. I think that the problem is the token max length is too long ( 8192). I want to ask how to change the token max length to finetune the model to get rid of this problem. Please help me.
image.png

Beijing Academy of Artificial Intelligence org

You can set query_max_len and passage_max_len to change the max length. For OOM issue, you also can reduce the batch size per_device_train_batch_size.

Sign up or log in to comment