The models can be loaded by the InferLLM project.
Chinese Alpaca model is from https://github.com/ymcui/Chinese-LLaMA-Alpaca
ggml Alpaca model is from https://huggingface.co/Sosaka/Alpaca-native-4bit-ggml/tree/main
the two models also can be loaded by the llama.cpp project.
InferLLM support the ChatGLM/ChatGLM2 model, the chatglm-q4/bin/chatglm2-q4.bin is the int4 quantized model from chatglm-6b/chatglm2-6b
InferLLM support the baichuan model, the baichuan-q4 is the int4 quantized model from baichuan
InferLLM support the llama2 model, the llama2-q4 is the int4 quantized model from llama2
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.