--- license: mit --- This is the ONNX variant of the [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) embeddings model created with the [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) integration. For ONNX export, run: ```bash pip install git+https://github.com/neuralmagic/optimum-deepsparse.git ``` ```python from optimum.deepsparse import DeepSparseModelForFeatureExtraction from transformers.onnx.utils import get_preprocessor from pathlib import Path model_id = "BAAI/bge-base-en-v1.5" # load model and convert to onnx model = DeepSparseModelForFeatureExtraction.from_pretrained(model_id, export=True) tokenizer = get_preprocessor(model_id) # save onnx checkpoint and tokenizer onnx_path = Path("bge-base-en-v1.5-dense") model.save_pretrained(onnx_path) tokenizer.save_pretrained(onnx_path) ``` Current up-to-date list of sparse and quantized bge ONNX models: [zeroshot/bge-large-en-v1.5-sparse](https://huggingface.co/zeroshot/bge-large-en-v1.5-sparse) [zeroshot/bge-large-en-v1.5-quant](https://huggingface.co/zeroshot/bge-large-en-v1.5-quant) [zeroshot/bge-base-en-v1.5-sparse](https://huggingface.co/zeroshot/bge-base-en-v1.5-sparse) [zeroshot/bge-base-en-v1.5-quant](https://huggingface.co/zeroshot/bge-base-en-v1.5-quant) [zeroshot/bge-small-en-v1.5-sparse](https://huggingface.co/zeroshot/bge-small-en-v1.5-sparse) [zeroshot/bge-small-en-v1.5-quant](https://huggingface.co/zeroshot/bge-small-en-v1.5-quant)