Unable to load checkpoint shards

#21
by Tilakraj0308 - opened

Got an error something like:
.cache\huggingface\hub\models--mistralai--Mistral-7B-Instruct-v0.1\snapshots\d635d39671aaceec5ef84b745bc21625b324b7f8\pytorch_model-00001-of-00002.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.

I have the same issue, any news?

@19Peppe95 The error is because system is running out of RAM to load the model in one go.
You can use CTransformers to load the model or can try GGUF model versions of your model which is basically much smaller version of it.
Gist - Use GGUF version of this model https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF
and use CTransformers to load it from the dowloads and run the program, hopefully it should work.

@19Peppe95 The error is because system is running out of RAM to load the model in one go.
You can use CTransformers to load the model or can try GGUF model versions of your model which is basically much smaller version of it.
Gist - Use GGUF version of this model https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF
and use CTransformers to load it from the dowloads and run the program, hopefully it should work.

I also have the RAM issue. But it is werid that the model is only 14GB in totally and I have 64GB RAM and 24GB VRAM available.
Just found out that there was a download issue, the bins are broken so the memory usage when loading the files became uncontrollable.

This comment has been hidden
Tilakraj0308 changed discussion status to closed

cc @ybelkada regarding low-memory methods to load larger models

Hi everyone
In case you are facing CPU OOM issues while loading the model please consider using sharded models with small shards, for this model I would recommend using this repository: https://huggingface.co/bn22/Mistral-7B-Instruct-v0.1-sharded

@ybelkada Hi, the model you shared is not available anymore? :(

Sign up or log in to comment