File size: 1,265 Bytes
b7d7804
 
e225449
bc34c30
8444104
6ef6b1d
 
b7d7804
 
 
 
 
 
 
1931f62
 
b7d7804
 
740f729
b7d7804
 
 
 
 
55fd70d
b7d7804
 
1b17997
b7d7804
d6a700f
b7d7804
 
 
4afceb6
b7d7804
7c9b1b1
66d0a2c
af7ada6
66d0a2c
8b86a1f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
import torch
import nltk
import io
import base64
import shutil
from torchvision import transforms

from pytorch_pretrained_biggan import BigGAN, one_hot_from_names, truncated_noise_sample

class PreTrainedPipeline():
    def __init__(self, path=""):
        """
        Initialize model
        """
        nltk.download('wordnet')
        self.model = BigGAN.from_pretrained(path)
        self.truncation = 0.1

    def __call__(self, inputs: str):
        """
        Args:
            inputs (:obj:`str`):
                a string containing some text
        Return:
            A :obj:`PIL.Image` with the raw image representation as PIL.
        """
        class_vector = one_hot_from_names([inputs], batch_size=1)
        if type(class_vector) == type(None):
            raise ValueError("Input is not in ImageNet")
        noise_vector = truncated_noise_sample(truncation=self.truncation, batch_size=1)
        noise_vector = torch.from_numpy(noise_vector)
        class_vector = torch.from_numpy(class_vector)
        with torch.no_grad():
            output = self.model(noise_vector, class_vector, self.truncation)

        # Scale image
        img = output[0]
        img = (img + 1) / 2.0
        img = transforms.ToPILImage()(img)
        return img