This is Chinese-Llama-2-7b f16 ggml model running llama.cpp.You can run
./main -m Chinese-Llama-2-7b-f16-ggml.bin -p 'hello world'
from model see: https://huggingface.co/LinkSoul/Chinese-Llama-2-7b
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.