Spaces:
Runtime error
Runtime error
# Prediction interface for Cog ⚙️ | |
# https://github.com/replicate/cog/blob/main/docs/python.md | |
from cog import BasePredictor, Input, Path | |
from typing import List | |
from omni_zero import OmniZeroSingle | |
from PIL import Image | |
class Predictor(BasePredictor): | |
def setup(self): | |
"""Load the model into memory to make running multiple predictions efficient""" | |
# self.model = torch.load("./weights.pth") | |
self.omni_zero = OmniZeroSingle( | |
base_model="frankjoshua/albedobaseXL_v13", | |
) | |
def predict( | |
self, | |
seed: int = Input(description="Random seed for the model", default=42), | |
prompt: str = Input(description="Prompt for the model", default="A person"), | |
negative_prompt: str = Input(description="Negative prompt for the model", default="blurry, out of focus"), | |
guidance_scale: float = Input(description="Guidance scale for the model", default=3.0, ge=0.0, le=14.0), | |
number_of_images: int = Input(description="Number of images to generate", default=1, ge=1, le=4), | |
number_of_steps: int = Input(description="Number of steps for the model", default=10, ge=1, le=50), | |
base_image: Path = Input(description="Base image for the model"), | |
base_image_strength: float = Input(description="Base image strength for the model", default=0.15, ge=0.0, le=1.0), | |
composition_image: Path = Input(description="Composition image for the model"), | |
composition_image_strength: float = Input(description="Composition image strength for the model", default=1.0, ge=0.0, le=1.0), | |
style_image: Path = Input(description="Style image for the model"), | |
style_image_strength: float = Input(description="Style image strength for the model", default=1.0, ge=0.0, le=1.0), | |
identity_image: Path = Input(description="Identity image for the model"), | |
identity_image_strength: float = Input(description="Identity image strength for the model", default=1.0, ge=0.0, le=1.0), | |
depth_image: Path = Input(description="Depth image for the model", default=None), | |
depth_image_strength: float = Input(description="Depth image strength for the model, if not supplied the composition image will be used for depth", default=0.5, ge=0.0, le=1.0), | |
) -> List[Path]: | |
"""Run a single prediction on the model""" | |
base_image = Image.open(base_image) | |
composition_image = Image.open(composition_image) | |
style_image = Image.open(style_image) | |
identity_image = Image.open(identity_image) | |
if depth_image is not None: | |
depth_image = Image.open(depth_image) | |
images = self.omni_zero.generate( | |
seed=seed, | |
prompt=prompt, | |
negative_prompt=negative_prompt, | |
guidance_scale=guidance_scale, | |
number_of_images=number_of_images, | |
number_of_steps=number_of_steps, | |
base_image=base_image, | |
base_image_strength=base_image_strength, | |
composition_image=composition_image, | |
composition_image_strength=composition_image_strength, | |
style_image=style_image, | |
style_image_strength=style_image_strength, | |
identity_image=identity_image, | |
identity_image_strength=identity_image_strength, | |
depth_image=depth_image, | |
depth_image_strength=depth_image_strength, | |
) | |
outputs = [] | |
for i, image in enumerate(images): | |
output_path = f"oz_output_{i}.jpg" | |
image.save(output_path) | |
outputs.append(Path(output_path)) | |
return outputs |