Model Loading Error
Running your code in colab but getting error: Error loading models: .to
is not supported for 4-bit
or 8-bit
bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct dtype
.
Thanks for the report. I wonder if the library's specifications have changed...π€
I don't know the direct cause, so I've tried removing the suspicious parts for now.
@John6666
The error still seems to be occurring. Could you please check?
Running on cuda
Loading in NF4
Loading CLIP π
Loading VLM's custom vision model π
Loading tokenizer πͺ
Loading LLM: unsloth/Meta-Llama-3.1-8B-bnb-4bit π€
Error loading models: .to
is not supported for 4-bit
or 8-bit
bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct dtype
.
It seems that accelerate and bitsandbytes were fighting. It's probably because of the version upgrade of accelerate in Colab. I think the change I just made will fix this.
https://github.com/OpenBMB/MiniCPM-o/issues/379
Thank you, it worked properly!