File size: 1,988 Bytes
719be41
ce1b66b
 
 
 
 
612bcea
ce1b66b
612bcea
ce1b66b
 
 
 
 
 
 
 
 
 
 
23a07e3
ce1b66b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9a7039c
 
c7d4c37
 
 
9a7039c
 
c7d4c37
9a7039c
 
 
ce1b66b
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
from PIL import Image

import numpy as np
import torch

from transformers.tools.base import Tool, get_default_device
from transformers.utils import is_accelerate_available

from diffusers import DiffusionPipeline


IMAGE_UPSCALING_DESCRIPTION = (
    "This is a tool that upscales an image. It takes one input: `image`, which should be "
    "the image to upscale. It returns the upscaled image."
)


class ImageUpscalingTool(Tool):
    default_stable_diffusion_checkpoint = "stabilityai/sd-x2-latent-upscaler"
    description = IMAGE_UPSCALING_DESCRIPTION 
    name = "image_upscaler"
    inputs = ['image']
    outputs = ['image']

    def __init__(self, device=None, controlnet=None, stable_diffusion=None, **hub_kwargs) -> None:
        if not is_accelerate_available():
            raise ImportError("Accelerate should be installed in order to use tools.")

        super().__init__()

        self.stable_diffusion = self.default_stable_diffusion_checkpoint

        self.device = device
        self.hub_kwargs = hub_kwargs

    def setup(self):
        if self.device is None:
            self.device = get_default_device()

        self.pipeline = DiffusionPipeline.from_pretrained(self.stable_diffusion)

        self.pipeline.to(self.device)
        if self.device.type == "cuda":
            self.pipeline.to(torch_dtype=torch.float16)

        self.is_initialized = True

    def __call__(self, image):
        if not self.is_initialized:
            self.setup()

        print(f"Received image type: {type(image)}")
        
        # Convert NumPy array to PIL Image
        if isinstance(image, np.ndarray):
            image = Image.fromarray((image * 255).astype(np.uint8))
        
        print(f"Converted image type: {type(image)}")

        # After converting the image
        image.show()
        
        return self.pipeline(
            image=image,
            prompt="",
            num_inference_steps=30,
            guidance_scale=0,
        ).images[0]