Spaces:
Running
on
Zero
Running
on
Zero
mrfakename
commited on
Sync from GitHub repo
Browse filesThis Space is synced from the GitHub repo: https://github.com/SWivid/F5-TTS. Please submit contributions to the Space there
src/f5_tts/train/finetune_cli.py
CHANGED
@@ -47,7 +47,7 @@ def parse_args():
|
|
47 |
parser.add_argument("--num_warmup_updates", type=int, default=300, help="Warmup steps")
|
48 |
parser.add_argument("--save_per_updates", type=int, default=10000, help="Save checkpoint every X steps")
|
49 |
parser.add_argument("--last_per_steps", type=int, default=50000, help="Save last checkpoint every X steps")
|
50 |
-
parser.add_argument("--finetune",
|
51 |
parser.add_argument("--pretrain", type=str, default=None, help="the path to the checkpoint")
|
52 |
parser.add_argument(
|
53 |
"--tokenizer", type=str, default="pinyin", choices=["pinyin", "char", "custom"], help="Tokenizer type"
|
@@ -60,15 +60,13 @@ def parse_args():
|
|
60 |
)
|
61 |
parser.add_argument(
|
62 |
"--log_samples",
|
63 |
-
|
64 |
-
default=False,
|
65 |
help="Log inferenced samples per ckpt save steps",
|
66 |
)
|
67 |
parser.add_argument("--logger", type=str, default=None, choices=["wandb", "tensorboard"], help="logger")
|
68 |
parser.add_argument(
|
69 |
"--bnb_optimizer",
|
70 |
-
|
71 |
-
default=False,
|
72 |
help="Use 8-bit Adam optimizer from bitsandbytes",
|
73 |
)
|
74 |
|
|
|
47 |
parser.add_argument("--num_warmup_updates", type=int, default=300, help="Warmup steps")
|
48 |
parser.add_argument("--save_per_updates", type=int, default=10000, help="Save checkpoint every X steps")
|
49 |
parser.add_argument("--last_per_steps", type=int, default=50000, help="Save last checkpoint every X steps")
|
50 |
+
parser.add_argument("--finetune", action="store_true", help="Use Finetune")
|
51 |
parser.add_argument("--pretrain", type=str, default=None, help="the path to the checkpoint")
|
52 |
parser.add_argument(
|
53 |
"--tokenizer", type=str, default="pinyin", choices=["pinyin", "char", "custom"], help="Tokenizer type"
|
|
|
60 |
)
|
61 |
parser.add_argument(
|
62 |
"--log_samples",
|
63 |
+
action="store_true",
|
|
|
64 |
help="Log inferenced samples per ckpt save steps",
|
65 |
)
|
66 |
parser.add_argument("--logger", type=str, default=None, choices=["wandb", "tensorboard"], help="logger")
|
67 |
parser.add_argument(
|
68 |
"--bnb_optimizer",
|
69 |
+
action="store_true",
|
|
|
70 |
help="Use 8-bit Adam optimizer from bitsandbytes",
|
71 |
)
|
72 |
|
src/f5_tts/train/finetune_gradio.py
CHANGED
@@ -452,7 +452,8 @@ def start_training(
|
|
452 |
f"--dataset_name {dataset_name}"
|
453 |
)
|
454 |
|
455 |
-
|
|
|
456 |
|
457 |
if file_checkpoint_train != "":
|
458 |
cmd += f" --pretrain {file_checkpoint_train}"
|
@@ -460,12 +461,12 @@ def start_training(
|
|
460 |
if tokenizer_file != "":
|
461 |
cmd += f" --tokenizer_path {tokenizer_file}"
|
462 |
|
463 |
-
cmd += f" --tokenizer {tokenizer_type}
|
464 |
|
465 |
-
cmd += f" --log_samples
|
466 |
|
467 |
if ch_8bit_adam:
|
468 |
-
cmd += " --bnb_optimizer
|
469 |
|
470 |
print("run command : \n" + cmd + "\n")
|
471 |
|
|
|
452 |
f"--dataset_name {dataset_name}"
|
453 |
)
|
454 |
|
455 |
+
if finetune:
|
456 |
+
cmd += " --finetune"
|
457 |
|
458 |
if file_checkpoint_train != "":
|
459 |
cmd += f" --pretrain {file_checkpoint_train}"
|
|
|
461 |
if tokenizer_file != "":
|
462 |
cmd += f" --tokenizer_path {tokenizer_file}"
|
463 |
|
464 |
+
cmd += f" --tokenizer {tokenizer_type}"
|
465 |
|
466 |
+
cmd += f" --log_samples --logger {logger}"
|
467 |
|
468 |
if ch_8bit_adam:
|
469 |
+
cmd += " --bnb_optimizer"
|
470 |
|
471 |
print("run command : \n" + cmd + "\n")
|
472 |
|