Spaces:
Paused
Paused
Update backend/main.py
Browse files- backend/main.py +3 -2
backend/main.py
CHANGED
@@ -133,11 +133,12 @@ static_files = {
|
|
133 |
},
|
134 |
}
|
135 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
136 |
-
processor = AutoProcessor.from_pretrained("facebook/seamless-m4t-v2-large",
|
|
|
137 |
|
138 |
# PM - hardcoding temporarily as my GPU doesnt have enough vram
|
139 |
# model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large").to("cpu")
|
140 |
-
model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large").to(device)
|
141 |
|
142 |
|
143 |
bytes_data = bytearray()
|
|
|
133 |
},
|
134 |
}
|
135 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
136 |
+
processor = AutoProcessor.from_pretrained("facebook/seamless-m4t-v2-large", force_download=True)
|
137 |
+
#cache_dir="/.cache"
|
138 |
|
139 |
# PM - hardcoding temporarily as my GPU doesnt have enough vram
|
140 |
# model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large").to("cpu")
|
141 |
+
model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large", force_download=True).to(device)
|
142 |
|
143 |
|
144 |
bytes_data = bytearray()
|