Nekochu commited on
Commit
c0f2a13
·
1 Parent(s): 3c5ce9c

switch training to standard turbo (11s/epoch), auto-select standard GGUF for LoRA inference

Browse files
Files changed (2) hide show
  1. Dockerfile +5 -1
  2. app.py +3 -2
Dockerfile CHANGED
@@ -53,10 +53,14 @@ RUN mkdir -p /app/models /app/adapters /app/outputs
53
  # Download GGUF models at build time (NOT via Git LFS)
54
  # --fail ensures curl returns non-zero on HTTP errors (catches 404)
55
 
56
- # XL DiT turbo Q4_K_M (~2.8GB) - best quality/size for CPU
57
  RUN curl -fL --retry 3 --retry-delay 5 -o /app/models/acestep-v15-xl-turbo-Q4_K_M.gguf \
58
  "https://huggingface.co/Serveurperso/ACE-Step-1.5-GGUF/resolve/main/acestep-v15-xl-turbo-Q4_K_M.gguf"
59
 
 
 
 
 
60
  # LM 1.7B Q8_0 (~1.7GB) - best speed/quality on CPU
61
  RUN curl -fL --retry 3 --retry-delay 5 -o /app/models/acestep-5Hz-lm-1.7B-Q8_0.gguf \
62
  "https://huggingface.co/Serveurperso/ACE-Step-1.5-GGUF/resolve/main/acestep-5Hz-lm-1.7B-Q8_0.gguf"
 
53
  # Download GGUF models at build time (NOT via Git LFS)
54
  # --fail ensures curl returns non-zero on HTTP errors (catches 404)
55
 
56
+ # XL DiT turbo Q4_K_M (~2.8GB) - best quality for no-LoRA inference
57
  RUN curl -fL --retry 3 --retry-delay 5 -o /app/models/acestep-v15-xl-turbo-Q4_K_M.gguf \
58
  "https://huggingface.co/Serveurperso/ACE-Step-1.5-GGUF/resolve/main/acestep-v15-xl-turbo-Q4_K_M.gguf"
59
 
60
+ # Standard DiT turbo Q4_K_M (~1.1GB) - used when LoRA adapter is selected
61
+ RUN curl -fL --retry 3 --retry-delay 5 -o /app/models/acestep-v15-turbo-Q4_K_M.gguf \
62
+ "https://huggingface.co/Serveurperso/ACE-Step-1.5-GGUF/resolve/main/acestep-v15-turbo-Q4_K_M.gguf"
63
+
64
  # LM 1.7B Q8_0 (~1.7GB) - best speed/quality on CPU
65
  RUN curl -fL --retry 3 --retry-delay 5 -o /app/models/acestep-5Hz-lm-1.7B-Q8_0.gguf \
66
  "https://huggingface.co/Serveurperso/ACE-Step-1.5-GGUF/resolve/main/acestep-5Hz-lm-1.7B-Q8_0.gguf"
app.py CHANGED
@@ -148,6 +148,7 @@ def _run_pipeline(caption, lyrics, bpm, duration, seed, steps, output_format,
148
  synth_request["output_format"] = synth_fmt
149
  if adapter:
150
  synth_request["adapter"] = adapter
 
151
  if progress_cb:
152
  progress_cb("synth_submit", None)
153
  r = requests.post(f"{ACE_SERVER}/synth", json=synth_request, timeout=30)
@@ -514,7 +515,7 @@ def gradio_main():
514
  output_dir=preprocessed_dir,
515
  checkpoint_dir=ACE_CHECKPOINT_DIR,
516
  device="cpu",
517
- variant="xl-turbo",
518
  max_duration=float(MAX_AUDIO_DURATION),
519
  progress_callback=preprocess_progress,
520
  cancel_check=lambda: False,
@@ -554,7 +555,7 @@ def gradio_main():
554
  max_grad_norm=1.0,
555
  save_every_n_epochs=max(1, epochs // 2),
556
  seed=42,
557
- variant="xl-turbo",
558
  device="cpu",
559
  log_every=5,
560
  ):
 
148
  synth_request["output_format"] = synth_fmt
149
  if adapter:
150
  synth_request["adapter"] = adapter
151
+ synth_request["synth_model"] = "acestep-v15-turbo-Q4_K_M.gguf"
152
  if progress_cb:
153
  progress_cb("synth_submit", None)
154
  r = requests.post(f"{ACE_SERVER}/synth", json=synth_request, timeout=30)
 
515
  output_dir=preprocessed_dir,
516
  checkpoint_dir=ACE_CHECKPOINT_DIR,
517
  device="cpu",
518
+ variant="turbo",
519
  max_duration=float(MAX_AUDIO_DURATION),
520
  progress_callback=preprocess_progress,
521
  cancel_check=lambda: False,
 
555
  max_grad_norm=1.0,
556
  save_every_n_epochs=max(1, epochs // 2),
557
  seed=42,
558
+ variant="turbo",
559
  device="cpu",
560
  log_every=5,
561
  ):