Spaces:
Running on Zero
Running on Zero
Commit ·
1d1ce31
1
Parent(s): 8635f79
Fix TARO GPU abort: increase load_overhead 15→120s to cover model loading time
Browse filesTARO loads CAVP extractor, onset model, MMDiT, AudioLDM2 VAE, and BigVGAN
vocoder before inference. This took >60s total, exhausting the GPU budget
before _load_taro_models could complete. 120s overhead + inference time
gives enough headroom for all model loading + feature extraction.
app.py
CHANGED
|
@@ -512,7 +512,7 @@ MODEL_CONFIGS = {
|
|
| 512 |
"window_s": TARO_MODEL_DUR, # 8.192 s
|
| 513 |
"sr": TARO_SR, # 16000 (output resampled to TARGET_SR)
|
| 514 |
"secs_per_step": 0.025, # measured 0.023 s/step on H200
|
| 515 |
-
"load_overhead":
|
| 516 |
"tab_prefix": "taro",
|
| 517 |
"label": "TARO",
|
| 518 |
"regen_fn": None, # set after function definitions (avoids forward-ref)
|
|
|
|
| 512 |
"window_s": TARO_MODEL_DUR, # 8.192 s
|
| 513 |
"sr": TARO_SR, # 16000 (output resampled to TARGET_SR)
|
| 514 |
"secs_per_step": 0.025, # measured 0.023 s/step on H200
|
| 515 |
+
"load_overhead": 120, # CAVP+onset+MMDiT+VAE+vocoder load + feature extraction
|
| 516 |
"tab_prefix": "taro",
|
| 517 |
"label": "TARO",
|
| 518 |
"regen_fn": None, # set after function definitions (avoids forward-ref)
|