Spaces:
Running
Running
File size: 642 Bytes
3cc4a06 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 |
from .clip import clip
from PIL import Image
import torch.nn as nn
CHANNELS = {
"RN50" : 1024,
"ViT-L/14" : 768
}
class CLIPModel(nn.Module):
def __init__(self, name, num_classes=1):
super(CLIPModel, self).__init__()
self.model, self.preprocess = clip.load(name, device="cpu") # self.preprocess will not be used during training, which is handled in Dataset class
self.fc = nn.Linear( CHANNELS[name], num_classes )
def forward(self, x, return_feature=False):
features = self.model.encode_image(x)
if return_feature:
return features
return self.fc(features)
|