tags: | |
- quantized | |
- 2-bit | |
- 3-bit | |
- 4-bit | |
- 5-bit | |
- 6-bit | |
- 8-bit | |
- fp16 | |
- GGUF | |
- transformers | |
- pytorch | |
- yi | |
- text-generation | |
- conversational | |
- endpoints_compatible | |
- text-generation-inference | |
- text-generation | |
license: apache-2.0 | |
library_name: transformers | |
inference: false | |
pipeline_tag: text-generation | |
The gguf quantization of [Fi-9B](https://huggingface.co/wenbopan/Fi-9B-200K) |