Dabs's picture
map location cpu
ed167eb
import gradio as gr
import models
import torch
import torchvision.transforms as transforms
import cv2
import numpy as np
# initialize the computation device
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
#intialize the model
model = models.model(pretrained=False, requires_grad=False).to(device)
# load the model checkpoint
checkpoint = torch.load('model.pth', map_location=device)
# load model weights state_dict
model.load_state_dict(checkpoint['model_state_dict'])
model.eval()
transform = transforms.Compose([
transforms.ToPILImage(),
transforms.ToTensor(),
])
genres = ['Action', 'Adventure', 'Animation', 'Biography', 'Comedy', 'Crime',
'Documentary', 'Drama', 'Family', 'Fantasy', 'History', 'Horror', 'Music',
'Musical', 'Mystery', 'N/A', 'News', 'Reality-TV', 'Romance', 'Sci-Fi', 'Short',
'Sport', 'Thriller', 'War', 'Western']
def segment(image):
image = np.asarray(image)
image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
image = transform(image)
image = torch.tensor(image, dtype=torch.float32)
image = image.to(device)
image = torch.unsqueeze(image, dim=0)
# get the predictions by passing the image through the model
outputs = model(image)
outputs = torch.sigmoid(outputs)
outputs = outputs.detach().cpu()
out_dict = {k: v for k, v in zip(genres, outputs.tolist()[0])}
return out_dict
iface = gr.Interface(fn=segment,
inputs="image",
outputs="label",
title="Poster classification",
description="classify the genre of your poster by uploading an image",
examples=[["imagenes/tt0084058.jpg"], ["imagenes/tt0084867.jpg"], ["imagenes/tt0085121.jpg"]]).launch()