Text Generation
GGUF
Russian
Edit model card

Llama.cpp compatible versions of an original 7B model.

Download one of the versions, for example model-q4_K.gguf.

wget https://huggingface.co/IlyaGusev/saiga_mistral_7b_gguf/resolve/main/model-q4_K.gguf

Download interact_mistral_llamacpp.py

wget https://raw.githubusercontent.com/IlyaGusev/rulm/master/self_instruct/src/interact_mistral_llamacpp.py

How to run:

pip install llama-cpp-python fire

python3 interact_mistral_llamacpp.py model-q4_K.gguf

System requirements:

  • 10GB RAM for q8_0 and less for smaller quantizations
Downloads last month
41,757
GGUF
Model size
7.24B params
Architecture
llama
Inference Examples
Inference API (serverless) has been turned off for this model.

Datasets used to train IlyaGusev/saiga_mistral_7b_gguf

Spaces using IlyaGusev/saiga_mistral_7b_gguf 7

Collection including IlyaGusev/saiga_mistral_7b_gguf