Out of Memory issue in sagemaker ml.g5.12xLarge instance

#41
by ChanakyaReddy - opened

model = AutoModelForCausalLM.from_pretrained(model_id) ,
Hi, Wondering if anyone else has this issue while downloading the mixtral model using the above

The target location /tmp/hub/models--mistralai--Mixtral-8x7B-v0.1/blobs only has 1808.00 MB free disk space

Hi @ChanakyaReddy !
Please consider loading your model in half-precision torch_dtype=torch.float16 or in 8-bit / 4-bit precision through the load_in_4bit / load_in_8bit arguments

Ybelkada, Thanks for the response. think I have fixed it, it appears the /tmp sub folder on AWS has some limitations, had to download it to another mounted location. Further, this is when the model is initially downloading, I don't think there is an option for downloading only the float16 of 4_bit/8_bit quantized ones, think that only applies after the initial download. Thanks for the response, appreciate it

Thank you very much for your prompt reply @ChanakyaReddy !

Sign up or log in to comment