segformer fp16
Browse files
app.py
CHANGED
@@ -27,7 +27,7 @@ seg_model_size = 0
|
|
27 |
seg_feature_extractor = SegformerFeatureExtractor.from_pretrained(f"nvidia/segformer-b{seg_model_size}-finetuned-cityscapes-{seg_model_img_size}-{seg_model_img_size}")
|
28 |
seg_model = SegformerForSemanticSegmentation.from_pretrained(
|
29 |
f"nvidia/segformer-b{seg_model_size}-finetuned-cityscapes-{seg_model_img_size}-{seg_model_img_size}"
|
30 |
-
).to(preferred_device)
|
31 |
|
32 |
inpainting_pipeline = StableDiffusionInpaintPipeline.from_pretrained(
|
33 |
"SimianLuo/LCM_Dreamshaper_v7",
|
@@ -54,7 +54,7 @@ ban_cars_mask = np.array(ban_cars_mask, dtype=np.uint8)
|
|
54 |
|
55 |
|
56 |
def get_seg_mask(img):
|
57 |
-
inputs = seg_feature_extractor(images=img, return_tensors="pt").to(preferred_device)
|
58 |
outputs = seg_model(**inputs)
|
59 |
logits = outputs.logits[0]
|
60 |
mask = Image.fromarray((ban_cars_mask[ torch.argmax(logits, dim=0).cpu().numpy() ]) * 255)
|
|
|
27 |
seg_feature_extractor = SegformerFeatureExtractor.from_pretrained(f"nvidia/segformer-b{seg_model_size}-finetuned-cityscapes-{seg_model_img_size}-{seg_model_img_size}")
|
28 |
seg_model = SegformerForSemanticSegmentation.from_pretrained(
|
29 |
f"nvidia/segformer-b{seg_model_size}-finetuned-cityscapes-{seg_model_img_size}-{seg_model_img_size}"
|
30 |
+
).to(preferred_device).to(preferred_dtype)
|
31 |
|
32 |
inpainting_pipeline = StableDiffusionInpaintPipeline.from_pretrained(
|
33 |
"SimianLuo/LCM_Dreamshaper_v7",
|
|
|
54 |
|
55 |
|
56 |
def get_seg_mask(img):
|
57 |
+
inputs = seg_feature_extractor(images=img, return_tensors="pt").to(preferred_device).to(preferred_dtype)
|
58 |
outputs = seg_model(**inputs)
|
59 |
logits = outputs.logits[0]
|
60 |
mask = Image.fromarray((ban_cars_mask[ torch.argmax(logits, dim=0).cpu().numpy() ]) * 255)
|