Spaces:
Runtime error
Runtime error
Commit
·
749cd3a
1
Parent(s):
33b734c
print needs parens
Browse files- assets/main.py +3 -3
assets/main.py
CHANGED
@@ -14,14 +14,14 @@ import torch
|
|
14 |
|
15 |
# multi-modal Model
|
16 |
# accepts both text and image content (or audio, etc.)
|
17 |
-
print
|
18 |
model = BlipForImageTextRetrieval.from_pretrained("Salesforce/blip-itm-base-coco")
|
19 |
processor = AutoProcessor.from_pretrained("Salesforce/blip-itm-base-coco")
|
20 |
|
21 |
-
print
|
22 |
raw_image = Image.open('./assets/pot-o-gold-my-little-pony-Derpy.jpeg').convert('RGB')
|
23 |
|
24 |
-
print
|
25 |
statements = [
|
26 |
"an image of a horse",
|
27 |
"a horse and a rainbow",
|
|
|
14 |
|
15 |
# multi-modal Model
|
16 |
# accepts both text and image content (or audio, etc.)
|
17 |
+
print("loading model ...")
|
18 |
model = BlipForImageTextRetrieval.from_pretrained("Salesforce/blip-itm-base-coco")
|
19 |
processor = AutoProcessor.from_pretrained("Salesforce/blip-itm-base-coco")
|
20 |
|
21 |
+
print("loading image ...")
|
22 |
raw_image = Image.open('./assets/pot-o-gold-my-little-pony-Derpy.jpeg').convert('RGB')
|
23 |
|
24 |
+
print("processing ...")
|
25 |
statements = [
|
26 |
"an image of a horse",
|
27 |
"a horse and a rainbow",
|