The models can be loaded by the InferLLM project.

Chinese Alpaca model is from https://github.com/ymcui/Chinese-LLaMA-Alpaca

ggml Alpaca model is from https://huggingface.co/Sosaka/Alpaca-native-4bit-ggml/tree/main

the two models also can be loaded by the llama.cpp project.

InferLLM support the ChatGLM/ChatGLM2 model, the chatglm-q4/bin/chatglm2-q4.bin is the int4 quantized model from chatglm-6b/chatglm2-6b

InferLLM support the baichuan model, the baichuan-q4 is the int4 quantized model from baichuan

InferLLM support the llama2 model, the llama2-q4 is the int4 quantized model from llama2

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.