clothes_segmentation / semgent_from_folder.py
SaffalPoosh's picture
Upload folder using huggingface_hub
e5765b1
# import os
# import cv2
# import numpy as np
# import torch
# import albumentations as albu
# from iglovikov_helper_functions.utils.image_utils import load_rgb, pad, unpad
# from iglovikov_helper_functions.dl.pytorch.utils import tensor_from_rgb_image
# from cloths_segmentation.pre_trained_models import create_model
# from tqdm import tqdm
# model = create_model("Unet_2020-10-30")
# model.to("cuda")
# model.eval()
# input_dir = "../../image"
# output_dir = "../../seg_masks"
# os.makedirs(output_dir, exist_ok=True)
# for image_filename in tqdm(os.listdir(input_dir), colour="green"):
# image_path = os.path.join(input_dir, image_filename)
# image = cv2.imread(image_path)
# image_2_extract = image
# image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
# transform = albu.Compose([albu.Normalize(p=1)], p=1)
# padded_image, pads = pad(image, factor=32, border=cv2.BORDER_CONSTANT)
# x = transform(image=padded_image)["image"]
# x = torch.unsqueeze(tensor_from_rgb_image(x), 0).to("cuda")
# with torch.no_grad():
# prediction = model(x)[0][0]
# mask = (prediction > 0).cpu().numpy().astype(np.uint8)
# mask = unpad(mask, pads)
# rmask = (cv2.cvtColor(mask, cv2.COLOR_BGR2RGB) * 255).astype(np.uint8)
# mask2 = np.where((rmask < 255), 0, 1).astype('uint8')
# image_2_extract = image_2_extract * mask2[:, :, 1, np.newaxis]
# tmp = cv2.cvtColor(image_2_extract, cv2.COLOR_BGR2GRAY)
# _, alpha = cv2.threshold(tmp, 0, 255, cv2.THRESH_BINARY)
# b, g, r = cv2.split(image_2_extract)
# rgba = [b, g, r, alpha]
# dst = cv2.merge(rgba, 4)
# output_image_path = os.path.join(output_dir, image_filename.replace(".jpg", ".png"))
# cv2.imwrite(output_image_path, dst)
# # break
import os
import cv2
import numpy as np
import torch
import albumentations as albu
from iglovikov_helper_functions.utils.image_utils import load_rgb, pad, unpad
from iglovikov_helper_functions.dl.pytorch.utils import tensor_from_rgb_image
from cloths_segmentation.pre_trained_models import create_model
from tqdm import tqdm
# Create the model and wrap it with DataParallel
model = create_model("Unet_2020-10-30")
model = torch.nn.DataParallel(model)
# Move the model to CUDA devices
model.to("cuda")
model.eval()
input_dir = "../../image"
output_dir = "../../seg_masks"
os.makedirs(output_dir, exist_ok=True)
for image_filename in tqdm(os.listdir(input_dir), colour="green"):
image_path = os.path.join(input_dir, image_filename)
image = cv2.imread(image_path)
image_2_extract = image
image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
transform = albu.Compose([albu.Normalize(p=1)], p=1)
padded_image, pads = pad(image, factor=32, border=cv2.BORDER_CONSTANT)
x = transform(image=padded_image)["image"]
x = torch.unsqueeze(tensor_from_rgb_image(x), 0).to("cuda")
with torch.no_grad():
# Use DataParallel to perform inference on all 4 GPUs
prediction = model(x)[0][0]
mask = (prediction > 0).cpu().numpy().astype(np.uint8)
mask = unpad(mask, pads)
rmask = (cv2.cvtColor(mask, cv2.COLOR_BGR2RGB) * 255).astype(np.uint8)
mask2 = np.where((rmask < 255), 0, 1).astype('uint8')
image_2_extract = image_2_extract * mask2[:, :, 1, np.newaxis]
tmp = cv2.cvtColor(image_2_extract, cv2.COLOR_BGR2GRAY)
_, alpha = cv2.threshold(tmp, 0, 255, cv2.THRESH_BINARY)
b, g, r = cv2.split(image_2_extract)
rgba = [b, g, r, alpha]
dst = cv2.merge(rgba, 4)
output_image_path = os.path.join(output_dir, image_filename.replace(".jpg", ".png"))
cv2.imwrite(output_image_path, dst)
# break