--- language: - ru tags: - PyTorch - Text2Image thumbnail: "https://github.com/sberbank-ai/ru-clip" --- # Model Card: ruCLIP Disclaimer: The code for using model you can found [here](https://github.com/sberbank-ai/ru-clip). # Model Details The ruCLIP model was developed by researchers at SberDevices and Sber AI based on origin OpenAI paper. # Model Type The model uses a ViT-B/32 Transformer architecture (initialized from OpenAI checkpoint and freezed while training) as an image encoder and uses [ruGPT3Small](https://github.com/sberbank-ai/ru-gpts) as a text encoder. These encoders are trained to maximize the similarity of (image, text) pairs via a contrastive loss. # Documents Our habr [post](https://habr.com/ru/company/sberdevices/blog/564440/). # Usage Code for using model you can obtain in our [repo](https://github.com/sberbank-ai/ru-clip). ``` from clip.evaluate.utils import ( get_text_batch, get_image_batch, get_tokenizer, show_test_images, load_weights_only ) import torch # Load model and tokenizer model, args = load_weights_only("ViT-B/32-small") model = model.cuda().float().eval() tokenizer = get_tokenizer() # Load test images and prepare for model images, texts = show_test_images(args) input_ids, attention_mask = get_text_batch(["Это " + desc for desc in texts], tokenizer, args) img_input = get_image_batch(images, args.img_transform, args) # Call model with torch.no_grad(): logits_per_image, logits_per_text = model( img_input={"x": img_input}, text_input={"x": input_ids, "attention_mask": attention_mask} ) ``` # Performance We evaluate our model on CIFAR100 and CIFAR10 datasets. zero-shot classification CIFAR100 top1 accuracy 0.4057; top5 accuracy 0.6975. zero-shot classification CIFAR10 top1 accuracy 0.7803; top5 accuracy 0.9834.