--- license: apache-2.0 pipeline_tag: text-generation --- This is Migel Tissera's [Tess-XS-v1.0](https://huggingface.co/migtissera/Tess-XS-v1.0), converted to GGUF. No other changes were made. Two files are avaliable here: - Tess-XS-v1.0-**fp16**.gguf: the original model converted to GGUF without quantization - Tess-XS-v1.0-**q8_0-LOT**.gguf: the original model converted to GGUF with q8_0 quantization using the `--leave-output-tensor` command-line option From llama.cpp/quantize --help: ``` --leave-output-tensor: Will leave output.weight un(re)quantized. Increases model size but may also increase quality, especially when requantizing ``` The model was converted using `convert.py` from Georgi Gerganov's llama.cpp repo, commit `#a6fc554`. All credit belongs to [Migel Tissera](https://huggingface.co/migtissera) for fine-tuning and releasing this model. Thank you!