CLIP
Collection
Multi-modal models that can be used for Smart Search in Immich. Models are sorted by average recall@k in descending order.
•
5 items
•
Updated
•
16
This repo contains ONNX exports for the corresponding ViT-based CLIP model by OpenCLIP. See the OpenCLIP repo for more info. Visual and textual encoders are separated into separate models for the purpose of generating image and text embeddings.
This repo is specifically intended for use with Immich, a self-hosted photo library.