--- license: mit pipeline_tag: image-classification --- ONNX port of [sentence-transformers/clip-ViT-B-32](https://huggingface.co/sentence-transformers/clip-ViT-B-32). This model is intended to be used for image classification and similarity searches. ### Usage Here's an example of performing inference using the model with [FastEmbed](https://github.com/qdrant/fastembed). ```py from fastembed import ImageEmbedding images = [ "./path/to/image1.jpg", "./path/to/image2.jpg", ] model = ImageEmbedding(model_name="Qdrant/clip-ViT-B-32-vision") embeddings = list(model.embed(images)) # [ # array([-0.1115, 0.0097, 0.0052, 0.0195, ...], dtype=float32), # array([-0.1019, 0.0635, -0.0332, 0.0522, ...], dtype=float32) # ] ```