I want to know how gguf converts this model and why its model_family is llama.

#3
by hzjane - opened

I want to know how gguf converts this model. I tried to use baichuan2-7b-chat.Q4_0.gguf and found that its model_family is llama. Does this mean that I am running llama's model.

GGUFConfig["general.architecture"] is llama

Yes, as I understand it, llama is a model structure and gguf is a file format. gguf contains all the files needed for model inference, including the model structure, model parameters, model hyperparameters, and so on.

Sign up or log in to comment