> Using CUDA: True > Number of GPUs: 1 > Model has 83048620 parameters  > EPOCH: 0/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 13:52:28)   --> STEP: 0/120 -- GLOBAL_STEP: 0 | > loss_gen: 6.07645 (6.07645) | > loss_kl: 149.81517 (149.81517) | > loss_feat: 0.23809 (0.23809) | > loss_mel: 90.26639 (90.26639) | > loss_duration: 1.37561 (1.37561) | > amp_scaler: 32768.00000 (32768.00000) | > loss_0: 247.77171 (247.77171) | > grad_norm_0: 0.00000 (0.00000) | > loss_disc: 6.07647 (6.07647) | > amp_scaler-1: 16384.00000 (16384.00000) | > loss_1: 6.07647 (6.07647) | > grad_norm_1: 0.00000 (0.00000) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.76320 (12.76321) | > loader_time: 2.57090 (2.57093)  --> STEP: 25/120 -- GLOBAL_STEP: 25 | > loss_gen: 1.61151 (1.95531) | > loss_kl: 6.37796 (59.54020) | > loss_feat: 0.46147 (0.28787) | > loss_mel: 48.89629 (65.11380) | > loss_duration: 1.03971 (1.28429) | > amp_scaler: 256.00000 (839.68000) | > loss_0: 58.38694 (128.18147) | > grad_norm_0: 114.51166 (196.96564) | > loss_disc: 2.93202 (3.29342) | > amp_scaler-1: 256.00000 (839.68000) | > loss_1: 2.93202 (3.29342) | > grad_norm_1: 1.06471 (2.44188) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 7.96780 (7.71320) | > loader_time: 0.01070 (0.00861)  --> STEP: 50/120 -- GLOBAL_STEP: 50 | > loss_gen: 2.18341 (1.86862) | > loss_kl: 3.35768 (31.92324) | > loss_feat: 1.83515 (0.69160) | > loss_mel: 40.67615 (54.46701) | > loss_duration: 1.01609 (1.15537) | > amp_scaler: 256.00000 (547.84000) | > loss_0: 49.06847 (90.10585) | > grad_norm_0: 71.02003 (156.53784) | > loss_disc: 2.58882 (3.04378) | > amp_scaler-1: 256.00000 (547.84000) | > loss_1: 2.58882 (3.04378) | > grad_norm_1: 9.20920 (4.96476) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.51380 (8.70260) | > loader_time: 0.01420 (0.01023)  --> STEP: 75/120 -- GLOBAL_STEP: 75 | > loss_gen: 2.52428 (1.96619) | > loss_kl: 2.25920 (22.18461) | > loss_feat: 2.67909 (1.28699) | > loss_mel: 37.05318 (49.24829) | > loss_duration: 1.05964 (1.12522) | > amp_scaler: 256.00000 (450.56000) | > loss_0: 45.57538 (75.81130) | > grad_norm_0: 143.67207 (144.98256) | > loss_disc: 2.31854 (2.86041) | > amp_scaler-1: 256.00000 (450.56000) | > loss_1: 2.31854 (2.86041) | > grad_norm_1: 9.67988 (7.89643) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.01600 (9.45841) | > loader_time: 0.01570 (0.01204)  --> STEP: 100/120 -- GLOBAL_STEP: 100 | > loss_gen: 2.27330 (2.07862) | > loss_kl: 1.96552 (17.17153) | > loss_feat: 3.75730 (1.82306) | > loss_mel: 38.39759 (46.36456) | > loss_duration: 1.07561 (1.11295) | > amp_scaler: 256.00000 (401.92000) | > loss_0: 47.46932 (68.55072) | > grad_norm_0: 133.65538 (137.99207) | > loss_disc: 2.17987 (2.70911) | > amp_scaler-1: 256.00000 (401.92000) | > loss_1: 2.17987 (2.70911) | > grad_norm_1: 8.83030 (8.74963) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.08680 (10.04084) | > loader_time: 0.01880 (0.01329) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.97866 (2.97866) | > loss_kl: 2.06248 (2.06248) | > loss_feat: 3.92299 (3.92299) | > loss_mel: 39.12604 (39.12604) | > loss_duration: 0.99023 (0.99023) | > loss_0: 49.08041 (49.08041) | > loss_disc: 1.99337 (1.99337) | > loss_1: 1.99337 (1.99337)  --> STEP: 1 | > loss_gen: 2.88409 (2.88409) | > loss_kl: 1.27591 (1.27591) | > loss_feat: 3.80346 (3.80346) | > loss_mel: 37.44182 (37.44182) | > loss_duration: 1.03252 (1.03252) | > loss_0: 46.43779 (46.43779) | > loss_disc: 2.09228 (2.09228) | > loss_1: 2.09228 (2.09228)  --> STEP: 2 | > loss_gen: 2.84980 (2.86695) | > loss_kl: 1.60857 (1.44224) | > loss_feat: 3.69188 (3.74767) | > loss_mel: 34.98266 (36.21224) | > loss_duration: 1.04848 (1.04050) | > loss_0: 44.18139 (45.30959) | > loss_disc: 2.07510 (2.08369) | > loss_1: 2.07510 (2.08369)  --> STEP: 3 | > loss_gen: 2.89363 (2.87584) | > loss_kl: 1.61292 (1.49913) | > loss_feat: 3.49973 (3.66502) | > loss_mel: 36.05981 (36.16143) | > loss_duration: 1.11045 (1.06381) | > loss_0: 45.17653 (45.26524) | > loss_disc: 2.07775 (2.08171) | > loss_1: 2.07775 (2.08171) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.01034 (+0.00000) | > avg_loss_gen: 2.87584 (+0.00000) | > avg_loss_kl: 1.49913 (+0.00000) | > avg_loss_feat: 3.66502 (+0.00000) | > avg_loss_mel: 36.16143 (+0.00000) | > avg_loss_duration: 1.06381 (+0.00000) | > avg_loss_0: 45.26524 (+0.00000) | > avg_loss_disc: 2.08171 (+0.00000) | > avg_loss_1: 2.08171 (+0.00000) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_121.pth.tar  > EPOCH: 1/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 14:14:02)   --> STEP: 4/120 -- GLOBAL_STEP: 125 | > loss_gen: 2.17728 (2.61467) | > loss_kl: 2.37858 (2.36996) | > loss_feat: 4.63820 (4.24851) | > loss_mel: 39.43829 (39.78944) | > loss_duration: 0.99619 (1.00676) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 49.62854 (50.02934) | > grad_norm_0: 95.03241 (110.75542) | > loss_disc: 2.31090 (2.15764) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.31090 (2.15764) | > grad_norm_1: 28.31949 (20.50555) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.15410 (8.49676) | > loader_time: 0.00780 (0.00687)  --> STEP: 29/120 -- GLOBAL_STEP: 150 | > loss_gen: 2.91857 (2.73818) | > loss_kl: 1.81291 (2.03235) | > loss_feat: 5.21571 (4.29515) | > loss_mel: 36.71029 (37.44304) | > loss_duration: 1.01250 (1.11463) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 47.66997 (47.62336) | > grad_norm_0: 40.86134 (103.54919) | > loss_disc: 1.96549 (2.04659) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 1.96549 (2.04659) | > grad_norm_1: 17.91652 (13.23822) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.81390 (10.22589) | > loader_time: 0.01130 (0.00929)  --> STEP: 54/120 -- GLOBAL_STEP: 175 | > loss_gen: 2.26820 (2.72100) | > loss_kl: 2.28746 (2.02608) | > loss_feat: 4.22735 (4.11309) | > loss_mel: 33.01435 (36.37741) | > loss_duration: 1.04304 (1.07649) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 42.84041 (46.31407) | > grad_norm_0: 96.65714 (103.39897) | > loss_disc: 2.04191 (2.10411) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.04191 (2.10411) | > grad_norm_1: 20.05828 (13.52932) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93760 (10.54139) | > loader_time: 0.01340 (0.01087)  --> STEP: 79/120 -- GLOBAL_STEP: 200 | > loss_gen: 2.42211 (2.70739) | > loss_kl: 1.96419 (2.02779) | > loss_feat: 3.74082 (4.06804) | > loss_mel: 32.64783 (35.32217) | > loss_duration: 1.11678 (1.07410) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 41.89173 (45.19949) | > grad_norm_0: 67.73688 (99.74050) | > loss_disc: 2.20416 (2.10595) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.20416 (2.10595) | > grad_norm_1: 13.35448 (13.46167) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.33860 (10.75265) | > loader_time: 0.01610 (0.01307)  --> STEP: 104/120 -- GLOBAL_STEP: 225 | > loss_gen: 2.04185 (2.69329) | > loss_kl: 1.95669 (2.02969) | > loss_feat: 3.85712 (4.03353) | > loss_mel: 31.39787 (34.52400) | > loss_duration: 1.06532 (1.07504) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 40.31884 (44.35556) | > grad_norm_0: 76.48087 (101.78728) | > loss_disc: 2.17949 (2.10657) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.17949 (2.10657) | > grad_norm_1: 33.06195 (14.17479) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.13180 (11.04866) | > loader_time: 0.01940 (0.01437) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.61356 (2.61356) | > loss_kl: 1.78664 (1.78664) | > loss_feat: 4.68982 (4.68982) | > loss_mel: 31.71400 (31.71400) | > loss_duration: 0.99650 (0.99650) | > loss_0: 41.80051 (41.80051) | > loss_disc: 1.72057 (1.72057) | > loss_1: 1.72057 (1.72057)  --> STEP: 1 | > loss_gen: 2.38960 (2.38960) | > loss_kl: 2.00315 (2.00315) | > loss_feat: 4.66295 (4.66295) | > loss_mel: 32.90339 (32.90339) | > loss_duration: 1.02239 (1.02239) | > loss_0: 42.98147 (42.98147) | > loss_disc: 1.83802 (1.83802) | > loss_1: 1.83802 (1.83802)  --> STEP: 2 | > loss_gen: 2.52613 (2.45786) | > loss_kl: 2.11988 (2.06151) | > loss_feat: 4.49597 (4.57946) | > loss_mel: 32.49386 (32.69862) | > loss_duration: 1.04699 (1.03469) | > loss_0: 42.68283 (42.83215) | > loss_disc: 1.90063 (1.86933) | > loss_1: 1.90063 (1.86933)  --> STEP: 3 | > loss_gen: 2.34846 (2.42140) | > loss_kl: 1.86355 (1.99553) | > loss_feat: 4.75461 (4.63784) | > loss_mel: 31.18858 (32.19528) | > loss_duration: 1.10015 (1.05651) | > loss_0: 41.25536 (42.30656) | > loss_disc: 1.75859 (1.83242) | > loss_1: 1.75859 (1.83242) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00713 (-0.00320) | > avg_loss_gen: 2.42140 (-0.45444) | > avg_loss_kl: 1.99553 (+0.49639) | > avg_loss_feat: 4.63784 (+0.97282) | > avg_loss_mel: 32.19528 (-3.96615) | > avg_loss_duration: 1.05651 (-0.00730) | > avg_loss_0: 42.30656 (-2.95868) | > avg_loss_disc: 1.83242 (-0.24929) | > avg_loss_1: 1.83242 (-0.24929) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_242.pth.tar  > EPOCH: 2/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 14:37:06)   --> STEP: 8/120 -- GLOBAL_STEP: 250 | > loss_gen: 2.41151 (3.06750) | > loss_kl: 2.36934 (2.53108) | > loss_feat: 4.49711 (4.53257) | > loss_mel: 31.96193 (33.88518) | > loss_duration: 0.97915 (1.01729) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 42.21904 (45.03363) | > grad_norm_0: 44.20160 (109.02777) | > loss_disc: 2.10749 (2.21318) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.10749 (2.21318) | > grad_norm_1: 14.77833 (30.78584) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.24270 (9.23624) | > loader_time: 0.00830 (0.00745)  --> STEP: 33/120 -- GLOBAL_STEP: 275 | > loss_gen: 2.56905 (2.79744) | > loss_kl: 2.04527 (2.25086) | > loss_feat: 3.41434 (4.12668) | > loss_mel: 30.37102 (32.23172) | > loss_duration: 0.99966 (1.10159) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 39.39935 (42.50828) | > grad_norm_0: 84.46883 (107.59728) | > loss_disc: 2.29333 (2.13592) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.29333 (2.13592) | > grad_norm_1: 19.38472 (20.49035) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.56540 (10.26272) | > loader_time: 0.01190 (0.00963)  --> STEP: 58/120 -- GLOBAL_STEP: 300 | > loss_gen: 3.01687 (2.67491) | > loss_kl: 1.86059 (2.15229) | > loss_feat: 2.56676 (3.91783) | > loss_mel: 31.77369 (31.90849) | > loss_duration: 1.06404 (1.07459) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 40.28196 (41.72811) | > grad_norm_0: 176.02805 (109.54216) | > loss_disc: 2.68448 (2.15132) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.68448 (2.15132) | > grad_norm_1: 27.27574 (19.97150) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.67270 (10.54438) | > loader_time: 0.03750 (0.01310)  --> STEP: 83/120 -- GLOBAL_STEP: 325 | > loss_gen: 2.25881 (2.58759) | > loss_kl: 1.90891 (2.07642) | > loss_feat: 3.01875 (3.74705) | > loss_mel: 30.57160 (31.60370) | > loss_duration: 1.10598 (1.07310) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.86406 (41.08787) | > grad_norm_0: 176.51721 (109.06867) | > loss_disc: 2.33786 (2.18696) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.33786 (2.18696) | > grad_norm_1: 21.70200 (19.23659) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.23440 (10.76568) | > loader_time: 0.01610 (0.01414)  --> STEP: 108/120 -- GLOBAL_STEP: 350 | > loss_gen: 2.41083 (2.52973) | > loss_kl: 2.01434 (2.03184) | > loss_feat: 3.13783 (3.61748) | > loss_mel: 29.28912 (31.33458) | > loss_duration: 1.09095 (1.07581) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.94307 (40.58943) | > grad_norm_0: 140.01213 (107.75909) | > loss_disc: 2.23323 (2.21807) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.23323 (2.21807) | > grad_norm_1: 4.98597 (18.89663) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.27010 (11.05349) | > loader_time: 0.02040 (0.01540) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.38739 (2.38739) | > loss_kl: 1.96770 (1.96770) | > loss_feat: 3.69613 (3.69613) | > loss_mel: 29.70191 (29.70191) | > loss_duration: 0.98795 (0.98795) | > loss_0: 38.74107 (38.74107) | > loss_disc: 2.06743 (2.06743) | > loss_1: 2.06743 (2.06743)  --> STEP: 1 | > loss_gen: 2.27672 (2.27672) | > loss_kl: 1.81375 (1.81375) | > loss_feat: 4.31405 (4.31405) | > loss_mel: 30.88537 (30.88537) | > loss_duration: 1.02196 (1.02196) | > loss_0: 40.31185 (40.31185) | > loss_disc: 2.17047 (2.17047) | > loss_1: 2.17047 (2.17047)  --> STEP: 2 | > loss_gen: 2.37978 (2.32825) | > loss_kl: 1.87495 (1.84435) | > loss_feat: 2.78137 (3.54771) | > loss_mel: 28.16647 (29.52592) | > loss_duration: 1.04333 (1.03264) | > loss_0: 36.24591 (38.27888) | > loss_disc: 2.39890 (2.28469) | > loss_1: 2.39890 (2.28469)  --> STEP: 3 | > loss_gen: 2.46761 (2.37470) | > loss_kl: 1.77745 (1.82205) | > loss_feat: 3.32822 (3.47455) | > loss_mel: 29.57674 (29.54286) | > loss_duration: 1.10113 (1.05547) | > loss_0: 38.25114 (38.26963) | > loss_disc: 2.16898 (2.24612) | > loss_1: 2.16898 (2.24612) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00708 (-0.00005) | > avg_loss_gen: 2.37470 (-0.04669) | > avg_loss_kl: 1.82205 (-0.17348) | > avg_loss_feat: 3.47455 (-1.16330) | > avg_loss_mel: 29.54286 (-2.65242) | > avg_loss_duration: 1.05547 (-0.00104) | > avg_loss_0: 38.26963 (-4.03692) | > avg_loss_disc: 2.24612 (+0.41370) | > avg_loss_1: 2.24612 (+0.41370) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_363.pth.tar  > EPOCH: 3/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 15:00:06)   --> STEP: 12/120 -- GLOBAL_STEP: 375 | > loss_gen: 2.05408 (2.43668) | > loss_kl: 2.19803 (2.36718) | > loss_feat: 3.75555 (3.55495) | > loss_mel: 32.09191 (31.91812) | > loss_duration: 1.03189 (1.24154) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 41.13145 (41.51847) | > grad_norm_0: 149.08344 (134.54593) | > loss_disc: 2.16258 (2.27166) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.16258 (2.27166) | > grad_norm_1: 9.86007 (14.08873) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.43730 (9.67503) | > loader_time: 0.00890 (0.00787)  --> STEP: 37/120 -- GLOBAL_STEP: 400 | > loss_gen: 2.54336 (2.42736) | > loss_kl: 2.01295 (2.14273) | > loss_feat: 3.11324 (3.43748) | > loss_mel: 31.45818 (30.94590) | > loss_duration: 1.05574 (1.09355) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 40.18346 (40.04702) | > grad_norm_0: 102.92089 (125.86950) | > loss_disc: 2.23715 (2.31250) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.23715 (2.31250) | > grad_norm_1: 8.45452 (16.64706) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.72520 (10.33944) | > loader_time: 0.01220 (0.00986)  --> STEP: 62/120 -- GLOBAL_STEP: 425 | > loss_gen: 2.53305 (2.39754) | > loss_kl: 2.01730 (2.06015) | > loss_feat: 3.41181 (3.38610) | > loss_mel: 30.51744 (30.58490) | > loss_duration: 1.07950 (1.07440) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 39.55911 (39.50309) | > grad_norm_0: 50.21550 (123.31262) | > loss_disc: 2.19832 (2.30237) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.19832 (2.30237) | > grad_norm_1: 10.80196 (16.91200) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.67780 (10.57538) | > loader_time: 0.01450 (0.01278)  --> STEP: 87/120 -- GLOBAL_STEP: 450 | > loss_gen: 1.89645 (2.36012) | > loss_kl: 1.82410 (2.01155) | > loss_feat: 3.01717 (3.28662) | > loss_mel: 30.85844 (30.44465) | > loss_duration: 1.09433 (1.07283) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.69049 (39.17578) | > grad_norm_0: 122.32942 (124.55136) | > loss_disc: 2.51659 (2.32669) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.51659 (2.32669) | > grad_norm_1: 13.43846 (16.84394) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.98480 (10.78547) | > loader_time: 0.01720 (0.01369)  --> STEP: 112/120 -- GLOBAL_STEP: 475 | > loss_gen: 2.28836 (2.35498) | > loss_kl: 1.95240 (1.97343) | > loss_feat: 2.87333 (3.26730) | > loss_mel: 28.95652 (30.19088) | > loss_duration: 1.12050 (1.07728) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.19112 (38.86388) | > grad_norm_0: 147.80569 (128.30478) | > loss_disc: 2.43368 (2.31470) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.43368 (2.31470) | > grad_norm_1: 10.23001 (16.34824) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.32570 (11.06607) | > loader_time: 0.02120 (0.01492) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.10827 (2.10827) | > loss_kl: 2.22829 (2.22829) | > loss_feat: 2.51066 (2.51066) | > loss_mel: 29.74248 (29.74248) | > loss_duration: 0.98800 (0.98800) | > loss_0: 37.57769 (37.57769) | > loss_disc: 2.48926 (2.48926) | > loss_1: 2.48926 (2.48926)  --> STEP: 1 | > loss_gen: 2.01693 (2.01693) | > loss_kl: 1.77810 (1.77810) | > loss_feat: 3.05286 (3.05286) | > loss_mel: 30.75601 (30.75601) | > loss_duration: 1.02825 (1.02825) | > loss_0: 38.63215 (38.63215) | > loss_disc: 2.37390 (2.37390) | > loss_1: 2.37390 (2.37390)  --> STEP: 2 | > loss_gen: 2.12112 (2.06903) | > loss_kl: 2.46687 (2.12249) | > loss_feat: 2.39645 (2.72465) | > loss_mel: 28.82145 (29.78873) | > loss_duration: 1.03956 (1.03390) | > loss_0: 36.84545 (37.73880) | > loss_disc: 2.53795 (2.45592) | > loss_1: 2.53795 (2.45592)  --> STEP: 3 | > loss_gen: 1.86500 (2.00102) | > loss_kl: 1.62092 (1.95530) | > loss_feat: 3.57172 (3.00701) | > loss_mel: 28.66188 (29.41311) | > loss_duration: 1.11891 (1.06224) | > loss_0: 36.83843 (37.43868) | > loss_disc: 2.29436 (2.40207) | > loss_1: 2.29436 (2.40207) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00710 (+0.00001) | > avg_loss_gen: 2.00102 (-0.37368) | > avg_loss_kl: 1.95530 (+0.13325) | > avg_loss_feat: 3.00701 (-0.46754) | > avg_loss_mel: 29.41311 (-0.12975) | > avg_loss_duration: 1.06224 (+0.00677) | > avg_loss_0: 37.43868 (-0.83096) | > avg_loss_disc: 2.40207 (+0.15595) | > avg_loss_1: 2.40207 (+0.15595) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_484.pth.tar  > EPOCH: 4/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 15:23:01)   --> STEP: 16/120 -- GLOBAL_STEP: 500 | > loss_gen: 2.31346 (2.26632) | > loss_kl: 2.12017 (2.18468) | > loss_feat: 3.10129 (3.01701) | > loss_mel: 30.23797 (30.80548) | > loss_duration: 1.01112 (1.19132) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.78402 (39.46481) | > grad_norm_0: 137.31375 (114.91152) | > loss_disc: 2.49315 (2.43791) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.49315 (2.43791) | > grad_norm_1: 18.30975 (20.13527) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.60100 (9.81229) | > loader_time: 0.01000 (0.00814)  --> STEP: 41/120 -- GLOBAL_STEP: 525 | > loss_gen: 2.86760 (2.24635) | > loss_kl: 1.88149 (2.05563) | > loss_feat: 2.40603 (2.97777) | > loss_mel: 29.24681 (30.08907) | > loss_duration: 1.02086 (1.08701) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.42278 (38.45582) | > grad_norm_0: 133.11221 (119.41533) | > loss_disc: 2.73590 (2.41686) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.73590 (2.41686) | > grad_norm_1: 32.89259 (19.33793) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.73880 (10.33987) | > loader_time: 0.01220 (0.01010)  --> STEP: 66/120 -- GLOBAL_STEP: 550 | > loss_gen: 2.27800 (2.27911) | > loss_kl: 1.88225 (1.98857) | > loss_feat: 3.85372 (3.11250) | > loss_mel: 29.31419 (29.83815) | > loss_duration: 1.10061 (1.07320) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.42877 (38.29154) | > grad_norm_0: 56.51452 (113.06187) | > loss_disc: 2.11482 (2.35815) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.11482 (2.35815) | > grad_norm_1: 10.49149 (18.50611) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.08080 (10.57016) | > loader_time: 0.01500 (0.01294)  --> STEP: 91/120 -- GLOBAL_STEP: 575 | > loss_gen: 2.98346 (2.31364) | > loss_kl: 1.82577 (1.95009) | > loss_feat: 2.74328 (3.18946) | > loss_mel: 30.45843 (29.71671) | > loss_duration: 1.08824 (1.07363) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 39.09918 (38.24354) | > grad_norm_0: 162.18541 (113.93531) | > loss_disc: 2.46170 (2.33572) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.46170 (2.33572) | > grad_norm_1: 28.59901 (19.66549) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.97710 (10.81212) | > loader_time: 0.01990 (0.01397)  --> STEP: 116/120 -- GLOBAL_STEP: 600 | > loss_gen: 2.33470 (2.33055) | > loss_kl: 1.69320 (1.91680) | > loss_feat: 2.81211 (3.22859) | > loss_mel: 29.02093 (29.54434) | > loss_duration: 1.20066 (1.08073) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.06158 (38.10102) | > grad_norm_0: 60.24828 (115.63805) | > loss_disc: 2.48145 (2.32948) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.48145 (2.32948) | > grad_norm_1: 17.63874 (19.65808) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.13730 (11.13714) | > loader_time: 0.02370 (0.01545) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.63847 (2.63847) | > loss_kl: 2.33921 (2.33921) | > loss_feat: 3.81192 (3.81192) | > loss_mel: 30.58862 (30.58862) | > loss_duration: 0.97696 (0.97696) | > loss_0: 40.35519 (40.35519) | > loss_disc: 2.00487 (2.00487) | > loss_1: 2.00487 (2.00487)  --> STEP: 1 | > loss_gen: 2.63580 (2.63580) | > loss_kl: 2.40159 (2.40159) | > loss_feat: 4.17806 (4.17806) | > loss_mel: 30.58393 (30.58393) | > loss_duration: 1.01531 (1.01531) | > loss_0: 40.81469 (40.81469) | > loss_disc: 2.06981 (2.06981) | > loss_1: 2.06981 (2.06981)  --> STEP: 2 | > loss_gen: 2.63734 (2.63657) | > loss_kl: 1.83092 (2.11626) | > loss_feat: 3.29922 (3.73864) | > loss_mel: 29.67648 (30.13021) | > loss_duration: 1.04381 (1.02956) | > loss_0: 38.48778 (39.65124) | > loss_disc: 2.19729 (2.13355) | > loss_1: 2.19729 (2.13355)  --> STEP: 3 | > loss_gen: 2.56579 (2.61298) | > loss_kl: 1.93266 (2.05506) | > loss_feat: 3.63592 (3.70440) | > loss_mel: 27.98540 (29.41527) | > loss_duration: 1.11564 (1.05825) | > loss_0: 37.23542 (38.84596) | > loss_disc: 2.11789 (2.12833) | > loss_1: 2.11789 (2.12833) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00737 (+0.00028) | > avg_loss_gen: 2.61298 (+0.61196) | > avg_loss_kl: 2.05506 (+0.09976) | > avg_loss_feat: 3.70440 (+0.69739) | > avg_loss_mel: 29.41527 (+0.00216) | > avg_loss_duration: 1.05825 (-0.00398) | > avg_loss_0: 38.84596 (+1.40729) | > avg_loss_disc: 2.12833 (-0.27374) | > avg_loss_1: 2.12833 (-0.27374)  > EPOCH: 5/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 15:45:42)   --> STEP: 20/120 -- GLOBAL_STEP: 625 | > loss_gen: 2.37191 (2.61792) | > loss_kl: 1.77639 (2.06597) | > loss_feat: 3.34764 (4.09680) | > loss_mel: 29.60611 (30.63451) | > loss_duration: 0.99997 (1.14920) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.10202 (40.56440) | > grad_norm_0: 131.30009 (109.33488) | > loss_disc: 2.75476 (2.17062) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.75476 (2.17062) | > grad_norm_1: 50.99447 (25.57603) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.82850 (10.59106) | > loader_time: 0.01060 (0.00873)  --> STEP: 45/120 -- GLOBAL_STEP: 650 | > loss_gen: 2.41868 (2.50902) | > loss_kl: 1.71565 (1.97644) | > loss_feat: 3.80961 (3.77741) | > loss_mel: 30.49590 (30.36060) | > loss_duration: 1.04381 (1.08099) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 39.48365 (39.70447) | > grad_norm_0: 107.77903 (118.22968) | > loss_disc: 2.07519 (2.21941) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.07519 (2.21941) | > grad_norm_1: 13.88453 (22.84437) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.88450 (10.76653) | > loader_time: 0.01280 (0.01052)  --> STEP: 70/120 -- GLOBAL_STEP: 675 | > loss_gen: 2.23614 (2.47834) | > loss_kl: 1.88462 (1.93160) | > loss_feat: 3.51644 (3.68295) | > loss_mel: 29.86715 (30.00345) | > loss_duration: 1.05677 (1.07238) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.56112 (39.16873) | > grad_norm_0: 144.66542 (112.32623) | > loss_disc: 2.19633 (2.22577) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.19633 (2.22577) | > grad_norm_1: 27.60695 (22.18591) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.30300 (10.90695) | > loader_time: 0.01550 (0.01262)  --> STEP: 95/120 -- GLOBAL_STEP: 700 | > loss_gen: 2.47199 (2.44005) | > loss_kl: 1.86103 (1.89518) | > loss_feat: 4.08202 (3.57636) | > loss_mel: 28.14920 (29.72573) | > loss_duration: 1.08953 (1.07347) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.65376 (38.71080) | > grad_norm_0: 42.32550 (109.73550) | > loss_disc: 2.26454 (2.25270) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.26454 (2.25270) | > grad_norm_1: 9.32627 (21.12180) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.02720 (11.11380) | > loader_time: 0.01840 (0.01390)  --> STEP: 120/120 -- GLOBAL_STEP: 725 | > loss_gen: 2.29737 (2.41466) | > loss_kl: 1.70560 (1.86186) | > loss_feat: 3.96147 (3.50960) | > loss_mel: 30.90417 (29.51004) | > loss_duration: 1.35783 (1.08613) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 40.22644 (38.38230) | > grad_norm_0: 199.65154 (108.95415) | > loss_disc: 2.31762 (2.27271) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.31762 (2.27271) | > grad_norm_1: 11.71657 (20.13165) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.32420 (11.36812) | > loader_time: 0.02780 (0.01563) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 3.34934 (3.34934) | > loss_kl: 1.87159 (1.87159) | > loss_feat: 3.06504 (3.06504) | > loss_mel: 27.84772 (27.84772) | > loss_duration: 0.98791 (0.98791) | > loss_0: 37.12160 (37.12160) | > loss_disc: 2.36682 (2.36682) | > loss_1: 2.36682 (2.36682)  --> STEP: 1 | > loss_gen: 3.12097 (3.12097) | > loss_kl: 1.91728 (1.91728) | > loss_feat: 3.63456 (3.63456) | > loss_mel: 28.09297 (28.09297) | > loss_duration: 1.01707 (1.01707) | > loss_0: 37.78286 (37.78286) | > loss_disc: 2.30373 (2.30373) | > loss_1: 2.30373 (2.30373)  --> STEP: 2 | > loss_gen: 3.09119 (3.10608) | > loss_kl: 1.84084 (1.87906) | > loss_feat: 3.17151 (3.40304) | > loss_mel: 28.06824 (28.08061) | > loss_duration: 1.03852 (1.02780) | > loss_0: 37.21031 (37.49658) | > loss_disc: 2.43953 (2.37163) | > loss_1: 2.43953 (2.37163)  --> STEP: 3 | > loss_gen: 3.09559 (3.10258) | > loss_kl: 1.90601 (1.88804) | > loss_feat: 4.10916 (3.63841) | > loss_mel: 28.00179 (28.05433) | > loss_duration: 1.10842 (1.05467) | > loss_0: 38.22097 (37.73804) | > loss_disc: 2.01194 (2.25174) | > loss_1: 2.01194 (2.25174) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00720 (-0.00017) | > avg_loss_gen: 3.10258 (+0.48960) | > avg_loss_kl: 1.88804 (-0.16701) | > avg_loss_feat: 3.63841 (-0.06599) | > avg_loss_mel: 28.05433 (-1.36094) | > avg_loss_duration: 1.05467 (-0.00358) | > avg_loss_0: 37.73804 (-1.10792) | > avg_loss_disc: 2.25174 (+0.12340) | > avg_loss_1: 2.25174 (+0.12340)  > EPOCH: 6/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 16:08:43)   --> STEP: 24/120 -- GLOBAL_STEP: 750 | > loss_gen: 2.24251 (2.37195) | > loss_kl: 1.73680 (2.01736) | > loss_feat: 3.60781 (3.40915) | > loss_mel: 29.44890 (29.81095) | > loss_duration: 1.03637 (1.13304) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.07238 (38.74244) | > grad_norm_0: 171.29330 (95.54125) | > loss_disc: 2.46565 (2.39155) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.46565 (2.39155) | > grad_norm_1: 15.81580 (18.66249) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.66860 (10.52103) | > loader_time: 0.01070 (0.00902)  --> STEP: 49/120 -- GLOBAL_STEP: 775 | > loss_gen: 2.51355 (2.36612) | > loss_kl: 1.81736 (1.92272) | > loss_feat: 2.95596 (3.33321) | > loss_mel: 28.83391 (29.44889) | > loss_duration: 1.03298 (1.07915) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.15376 (38.15009) | > grad_norm_0: 114.37331 (109.39012) | > loss_disc: 2.32816 (2.37471) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.32816 (2.37471) | > grad_norm_1: 12.23704 (17.36994) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.87540 (10.72569) | > loader_time: 0.01370 (0.01071)  --> STEP: 74/120 -- GLOBAL_STEP: 800 | > loss_gen: 2.26161 (2.35011) | > loss_kl: 1.80096 (1.88311) | > loss_feat: 3.64513 (3.31676) | > loss_mel: 28.67740 (29.15675) | > loss_duration: 1.07985 (1.07534) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.46495 (37.78207) | > grad_norm_0: 51.43805 (114.98856) | > loss_disc: 2.15300 (2.34053) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.15300 (2.34053) | > grad_norm_1: 15.03514 (14.94958) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.73200 (10.88761) | > loader_time: 0.01570 (0.01282)  --> STEP: 99/120 -- GLOBAL_STEP: 825 | > loss_gen: 2.53341 (2.34220) | > loss_kl: 1.77131 (1.85740) | > loss_feat: 3.60116 (3.30098) | > loss_mel: 28.44342 (28.90558) | > loss_duration: 1.05732 (1.07632) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.40662 (37.48248) | > grad_norm_0: 161.10046 (113.42023) | > loss_disc: 2.30653 (2.34620) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.30653 (2.34620) | > grad_norm_1: 12.66083 (16.29262) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.17920 (11.13350) | > loader_time: 0.01880 (0.01420) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.48398 (2.48398) | > loss_kl: 1.90436 (1.90436) | > loss_feat: 3.73850 (3.73850) | > loss_mel: 29.99259 (29.99259) | > loss_duration: 0.97827 (0.97827) | > loss_0: 39.09771 (39.09771) | > loss_disc: 2.05349 (2.05349) | > loss_1: 2.05349 (2.05349)  --> STEP: 1 | > loss_gen: 2.42707 (2.42707) | > loss_kl: 1.84243 (1.84243) | > loss_feat: 3.82786 (3.82786) | > loss_mel: 29.91074 (29.91074) | > loss_duration: 1.02070 (1.02070) | > loss_0: 39.02880 (39.02880) | > loss_disc: 2.09397 (2.09397) | > loss_1: 2.09397 (2.09397)  --> STEP: 2 | > loss_gen: 2.39106 (2.40907) | > loss_kl: 2.05049 (1.94646) | > loss_feat: 3.55268 (3.69027) | > loss_mel: 29.06770 (29.48922) | > loss_duration: 1.03757 (1.02914) | > loss_0: 38.09951 (38.56416) | > loss_disc: 2.11047 (2.10222) | > loss_1: 2.11047 (2.10222)  --> STEP: 3 | > loss_gen: 2.37888 (2.39900) | > loss_kl: 1.88301 (1.92531) | > loss_feat: 3.74786 (3.70947) | > loss_mel: 26.93718 (28.63854) | > loss_duration: 1.10672 (1.05500) | > loss_0: 36.05365 (37.72732) | > loss_disc: 2.03675 (2.08040) | > loss_1: 2.03675 (2.08040) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00718 (-0.00002) | > avg_loss_gen: 2.39900 (-0.70358) | > avg_loss_kl: 1.92531 (+0.03726) | > avg_loss_feat: 3.70947 (+0.07105) | > avg_loss_mel: 28.63854 (+0.58421) | > avg_loss_duration: 1.05500 (+0.00033) | > avg_loss_0: 37.72732 (-0.01072) | > avg_loss_disc: 2.08040 (-0.17134) | > avg_loss_1: 2.08040 (-0.17134) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_847.pth.tar  > EPOCH: 7/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 16:32:05)   --> STEP: 3/120 -- GLOBAL_STEP: 850 | > loss_gen: 2.96086 (2.54180) | > loss_kl: 2.35160 (2.15326) | > loss_feat: 4.51621 (4.24243) | > loss_mel: 30.29047 (30.82790) | > loss_duration: 0.97731 (1.01717) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 41.09645 (40.78257) | > grad_norm_0: 140.65901 (160.97946) | > loss_disc: 2.09894 (2.06435) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.09894 (2.06435) | > grad_norm_1: 10.06472 (11.71507) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 8.93080 (8.32556) | > loader_time: 0.00570 (0.00726)  --> STEP: 28/120 -- GLOBAL_STEP: 875 | > loss_gen: 2.80279 (2.44815) | > loss_kl: 1.78868 (1.98197) | > loss_feat: 3.51036 (3.67748) | > loss_mel: 29.54508 (29.61072) | > loss_duration: 1.02886 (1.11603) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 38.67577 (38.83436) | > grad_norm_0: 81.67578 (118.14132) | > loss_disc: 2.30938 (2.28979) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.30938 (2.28979) | > grad_norm_1: 29.74305 (22.70913) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.05920 (10.27748) | > loader_time: 0.01140 (0.00940)  --> STEP: 53/120 -- GLOBAL_STEP: 900 | > loss_gen: 2.02171 (2.37999) | > loss_kl: 1.70452 (1.89096) | > loss_feat: 3.30066 (3.49886) | > loss_mel: 27.99948 (29.05997) | > loss_duration: 1.06929 (1.07868) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 36.09566 (37.90847) | > grad_norm_0: 160.54237 (118.49524) | > loss_disc: 2.24161 (2.29884) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.24161 (2.29884) | > grad_norm_1: 12.14684 (19.52571) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.00050 (10.59753) | > loader_time: 0.01320 (0.01098)  --> STEP: 78/120 -- GLOBAL_STEP: 925 | > loss_gen: 1.85795 (2.36086) | > loss_kl: 1.76254 (1.86153) | > loss_feat: 2.94932 (3.43601) | > loss_mel: 27.26167 (28.80310) | > loss_duration: 1.04782 (1.07600) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 34.87930 (37.53750) | > grad_norm_0: 122.99664 (117.45270) | > loss_disc: 2.39479 (2.30086) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.39479 (2.30086) | > grad_norm_1: 18.50110 (18.47225) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.22870 (10.79638) | > loader_time: 0.01580 (0.01291)  --> STEP: 103/120 -- GLOBAL_STEP: 950 | > loss_gen: 2.74876 (2.35368) | > loss_kl: 1.91818 (1.84170) | > loss_feat: 3.56451 (3.39779) | > loss_mel: 26.65822 (28.64260) | > loss_duration: 1.08518 (1.07704) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 35.97486 (37.31280) | > grad_norm_0: 111.63832 (123.95778) | > loss_disc: 2.17460 (2.31273) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.17460 (2.31273) | > grad_norm_1: 12.96475 (18.22845) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.35420 (11.06588) | > loader_time: 0.01920 (0.01412) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.55368 (2.55368) | > loss_kl: 2.29782 (2.29782) | > loss_feat: 3.71314 (3.71314) | > loss_mel: 29.41654 (29.41654) | > loss_duration: 0.97520 (0.97520) | > loss_0: 38.95638 (38.95638) | > loss_disc: 2.12055 (2.12055) | > loss_1: 2.12055 (2.12055)  --> STEP: 1 | > loss_gen: 2.49314 (2.49314) | > loss_kl: 1.88076 (1.88076) | > loss_feat: 3.02130 (3.02130) | > loss_mel: 29.57343 (29.57343) | > loss_duration: 1.01655 (1.01655) | > loss_0: 37.98518 (37.98518) | > loss_disc: 2.31490 (2.31490) | > loss_1: 2.31490 (2.31490)  --> STEP: 2 | > loss_gen: 2.45190 (2.47252) | > loss_kl: 1.83279 (1.85677) | > loss_feat: 3.27328 (3.14729) | > loss_mel: 28.00193 (28.78768) | > loss_duration: 1.03343 (1.02499) | > loss_0: 36.59332 (37.28925) | > loss_disc: 2.24278 (2.27884) | > loss_1: 2.24278 (2.27884)  --> STEP: 3 | > loss_gen: 2.31097 (2.41867) | > loss_kl: 1.93793 (1.88382) | > loss_feat: 4.01284 (3.43581) | > loss_mel: 27.33091 (28.30209) | > loss_duration: 1.12088 (1.05695) | > loss_0: 36.71353 (37.09734) | > loss_disc: 2.03898 (2.19889) | > loss_1: 2.03898 (2.19889) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00731 (+0.00013) | > avg_loss_gen: 2.41867 (+0.01967) | > avg_loss_kl: 1.88382 (-0.04149) | > avg_loss_feat: 3.43581 (-0.27366) | > avg_loss_mel: 28.30209 (-0.33645) | > avg_loss_duration: 1.05695 (+0.00195) | > avg_loss_0: 37.09734 (-0.62998) | > avg_loss_disc: 2.19889 (+0.11849) | > avg_loss_1: 2.19889 (+0.11849) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_968.pth.tar  > EPOCH: 8/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 16:55:13)   --> STEP: 7/120 -- GLOBAL_STEP: 975 | > loss_gen: 2.62054 (2.43422) | > loss_kl: 2.06450 (2.20552) | > loss_feat: 4.20265 (3.77603) | > loss_mel: 30.28827 (30.15060) | > loss_duration: 1.03735 (1.03534) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 40.21330 (39.60171) | > grad_norm_0: 141.78793 (137.46114) | > loss_disc: 2.02591 (2.23207) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.02591 (2.23207) | > grad_norm_1: 11.49835 (12.43628) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.23150 (9.17806) | > loader_time: 0.00660 (0.00747)  --> STEP: 32/120 -- GLOBAL_STEP: 1000 | > loss_gen: 2.00401 (2.46980) | > loss_kl: 1.87896 (1.99061) | > loss_feat: 4.07964 (3.78876) | > loss_mel: 28.31150 (29.27832) | > loss_duration: 0.98003 (1.10717) | > amp_scaler: 256.00000 (256.00000) | > loss_0: 37.25415 (38.63466) | > grad_norm_0: 88.11565 (101.21095) | > loss_disc: 2.25502 (2.27267) | > amp_scaler-1: 256.00000 (256.00000) | > loss_1: 2.25502 (2.27267) | > grad_norm_1: 23.49695 (18.09692) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.85870 (10.32591) | > loader_time: 0.01200 (0.00965)  --> STEP: 57/120 -- GLOBAL_STEP: 1025 | > loss_gen: 2.68628 (2.42122) | > loss_kl: 1.82858 (1.91274) | > loss_feat: 3.58508 (3.60406) | > loss_mel: 28.29495 (28.78597) | > loss_duration: 1.06898 (1.07836) | > amp_scaler: 512.00000 (345.82456) | > loss_0: 37.46386 (37.80235) | > grad_norm_0: 176.06976 (118.73025) | > loss_disc: 2.41734 (2.29285) | > amp_scaler-1: 512.00000 (345.82456) | > loss_1: 2.41734 (2.29285) | > grad_norm_1: 23.39114 (18.15245) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.03720 (10.61519) | > loader_time: 0.03840 (0.01201)  --> STEP: 82/120 -- GLOBAL_STEP: 1050 | > loss_gen: 2.42460 (2.38088) | > loss_kl: 1.61814 (1.86732) | > loss_feat: 3.10189 (3.50242) | > loss_mel: 26.82736 (28.52864) | > loss_duration: 1.03548 (1.07693) | > amp_scaler: 512.00000 (396.48780) | > loss_0: 35.00747 (37.35619) | > grad_norm_0: 42.26217 (118.68971) | > loss_disc: 2.33037 (2.30382) | > amp_scaler-1: 512.00000 (396.48780) | > loss_1: 2.33037 (2.30382) | > grad_norm_1: 9.29370 (16.39288) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.01300 (10.81826) | > loader_time: 0.01630 (0.01365)  --> STEP: 107/120 -- GLOBAL_STEP: 1075 | > loss_gen: 2.67607 (2.36352) | > loss_kl: 1.81747 (1.84954) | > loss_feat: 3.37095 (3.46028) | > loss_mel: 27.52942 (28.37171) | > loss_duration: 1.14307 (1.07982) | > amp_scaler: 512.00000 (423.47664) | > loss_0: 36.53698 (37.12486) | > grad_norm_0: 58.42699 (121.94105) | > loss_disc: 2.36161 (2.30314) | > amp_scaler-1: 512.00000 (423.47664) | > loss_1: 2.36161 (2.30314) | > grad_norm_1: 26.00139 (16.22300) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.46460 (11.11262) | > loader_time: 0.02050 (0.01498) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.32358 (2.32358) | > loss_kl: 2.00836 (2.00836) | > loss_feat: 3.45625 (3.45625) | > loss_mel: 27.26110 (27.26110) | > loss_duration: 0.98249 (0.98249) | > loss_0: 36.03176 (36.03176) | > loss_disc: 2.16736 (2.16736) | > loss_1: 2.16736 (2.16736)  --> STEP: 1 | > loss_gen: 2.25451 (2.25451) | > loss_kl: 2.04206 (2.04206) | > loss_feat: 3.40285 (3.40285) | > loss_mel: 29.43934 (29.43934) | > loss_duration: 1.02507 (1.02507) | > loss_0: 38.16383 (38.16383) | > loss_disc: 2.24321 (2.24321) | > loss_1: 2.24321 (2.24321)  --> STEP: 2 | > loss_gen: 2.24215 (2.24833) | > loss_kl: 1.86215 (1.95211) | > loss_feat: 3.12840 (3.26563) | > loss_mel: 26.73132 (28.08533) | > loss_duration: 1.04381 (1.03444) | > loss_0: 35.00783 (36.58583) | > loss_disc: 2.25882 (2.25101) | > loss_1: 2.25882 (2.25101)  --> STEP: 3 | > loss_gen: 2.32375 (2.27347) | > loss_kl: 2.08622 (1.99681) | > loss_feat: 3.48595 (3.33907) | > loss_mel: 27.51146 (27.89404) | > loss_duration: 1.11711 (1.06200) | > loss_0: 36.52450 (36.56539) | > loss_disc: 2.20424 (2.23542) | > loss_1: 2.20424 (2.23542) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00745 (+0.00014) | > avg_loss_gen: 2.27347 (-0.14520) | > avg_loss_kl: 1.99681 (+0.11299) | > avg_loss_feat: 3.33907 (-0.09674) | > avg_loss_mel: 27.89404 (-0.40805) | > avg_loss_duration: 1.06200 (+0.00504) | > avg_loss_0: 36.56539 (-0.53196) | > avg_loss_disc: 2.23542 (+0.03654) | > avg_loss_1: 2.23542 (+0.03654) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1089.pth.tar  > EPOCH: 9/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 17:18:20)   --> STEP: 11/120 -- GLOBAL_STEP: 1100 | > loss_gen: 2.45514 (2.50545) | > loss_kl: 1.98559 (2.12531) | > loss_feat: 3.11461 (3.69660) | > loss_mel: 28.98737 (28.86884) | > loss_duration: 1.00089 (1.27386) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 37.54360 (38.47006) | > grad_norm_0: 131.13751 (115.51991) | > loss_disc: 2.42826 (2.35512) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.42826 (2.35512) | > grad_norm_1: 11.17393 (23.99650) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.39450 (9.63883) | > loader_time: 0.00930 (0.00761)  --> STEP: 36/120 -- GLOBAL_STEP: 1125 | > loss_gen: 2.57945 (2.40253) | > loss_kl: 1.88863 (2.01864) | > loss_feat: 2.94399 (3.49560) | > loss_mel: 26.62782 (28.30834) | > loss_duration: 1.02153 (1.09906) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.06142 (37.32417) | > grad_norm_0: 41.71490 (106.55749) | > loss_disc: 2.38399 (2.34716) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.38399 (2.34716) | > grad_norm_1: 27.38872 (23.24472) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.94740 (10.40249) | > loader_time: 0.01460 (0.01093)  --> STEP: 61/120 -- GLOBAL_STEP: 1150 | > loss_gen: 2.14737 (2.35289) | > loss_kl: 2.01296 (1.95492) | > loss_feat: 3.36820 (3.38934) | > loss_mel: 28.05445 (28.02597) | > loss_duration: 1.07578 (1.07865) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 36.65876 (36.80177) | > grad_norm_0: 57.50605 (105.86054) | > loss_disc: 2.22824 (2.34602) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.22824 (2.34602) | > grad_norm_1: 14.32366 (21.41827) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.97850 (10.65410) | > loader_time: 0.01400 (0.01331)  --> STEP: 86/120 -- GLOBAL_STEP: 1175 | > loss_gen: 2.34603 (2.33755) | > loss_kl: 1.77753 (1.91595) | > loss_feat: 3.04983 (3.37120) | > loss_mel: 27.48634 (27.89168) | > loss_duration: 1.08060 (1.07766) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.74032 (36.59403) | > grad_norm_0: 131.31633 (111.97108) | > loss_disc: 2.48757 (2.34225) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48757 (2.34225) | > grad_norm_1: 18.94344 (19.35784) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.90350 (10.60874) | > loader_time: 0.02000 (0.01410)  --> STEP: 111/120 -- GLOBAL_STEP: 1200 | > loss_gen: 2.09664 (2.32413) | > loss_kl: 1.80383 (1.89311) | > loss_feat: 3.92477 (3.36005) | > loss_mel: 28.35120 (27.72802) | > loss_duration: 1.13062 (1.08281) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 37.30705 (36.38813) | > grad_norm_0: 36.75795 (112.11005) | > loss_disc: 2.24026 (2.34439) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.24026 (2.34439) | > grad_norm_1: 15.01256 (18.63753) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.74320 (10.79493) | > loader_time: 0.02470 (0.01545) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.52727 (2.52727) | > loss_kl: 2.19443 (2.19443) | > loss_feat: 3.50687 (3.50687) | > loss_mel: 28.33855 (28.33855) | > loss_duration: 1.01277 (1.01277) | > loss_0: 37.57989 (37.57989) | > loss_disc: 2.31226 (2.31226) | > loss_1: 2.31226 (2.31226)  --> STEP: 1 | > loss_gen: 2.40272 (2.40272) | > loss_kl: 2.12567 (2.12567) | > loss_feat: 3.46845 (3.46845) | > loss_mel: 28.74523 (28.74523) | > loss_duration: 1.05134 (1.05134) | > loss_0: 37.79342 (37.79342) | > loss_disc: 2.36362 (2.36362) | > loss_1: 2.36362 (2.36362)  --> STEP: 2 | > loss_gen: 2.52501 (2.46387) | > loss_kl: 1.93184 (2.02876) | > loss_feat: 3.37685 (3.42265) | > loss_mel: 27.57129 (28.15826) | > loss_duration: 1.06564 (1.05849) | > loss_0: 36.47063 (37.13202) | > loss_disc: 2.32272 (2.34317) | > loss_1: 2.32272 (2.34317)  --> STEP: 3 | > loss_gen: 2.58351 (2.50375) | > loss_kl: 2.12671 (2.06141) | > loss_feat: 2.97506 (3.27345) | > loss_mel: 26.22522 (27.51392) | > loss_duration: 1.14495 (1.08731) | > loss_0: 35.05545 (36.43983) | > loss_disc: 2.41878 (2.36837) | > loss_1: 2.41878 (2.36837) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00725 (-0.00020) | > avg_loss_gen: 2.50375 (+0.23028) | > avg_loss_kl: 2.06141 (+0.06460) | > avg_loss_feat: 3.27345 (-0.06562) | > avg_loss_mel: 27.51392 (-0.38013) | > avg_loss_duration: 1.08731 (+0.02531) | > avg_loss_0: 36.43983 (-0.12556) | > avg_loss_disc: 2.36837 (+0.13295) | > avg_loss_1: 2.36837 (+0.13295)  > EPOCH: 10/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 17:40:23)   --> STEP: 15/120 -- GLOBAL_STEP: 1225 | > loss_gen: 2.27680 (2.37701) | > loss_kl: 2.01479 (2.16998) | > loss_feat: 3.15024 (3.54484) | > loss_mel: 29.22594 (28.88705) | > loss_duration: 1.01855 (1.19352) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 37.68633 (38.17240) | > grad_norm_0: 42.49253 (103.73182) | > loss_disc: 2.35530 (2.37283) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.35530 (2.37283) | > grad_norm_1: 7.76435 (12.29097) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.00810 (9.82800) | > loader_time: 0.00970 (0.00810)  --> STEP: 40/120 -- GLOBAL_STEP: 1250 | > loss_gen: 2.64713 (2.33040) | > loss_kl: 1.98221 (2.02056) | > loss_feat: 3.49349 (3.49774) | > loss_mel: 27.81729 (28.24497) | > loss_duration: 0.97934 (1.08943) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 36.91946 (37.18310) | > grad_norm_0: 98.35818 (84.03394) | > loss_disc: 2.34510 (2.34196) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.34510 (2.34196) | > grad_norm_1: 16.28012 (13.21774) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.62000 (10.17611) | > loader_time: 0.01280 (0.01021)  --> STEP: 65/120 -- GLOBAL_STEP: 1275 | > loss_gen: 2.31280 (2.29487) | > loss_kl: 1.92495 (1.96778) | > loss_feat: 2.87850 (3.37133) | > loss_mel: 26.26216 (27.97326) | > loss_duration: 1.09271 (1.07855) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.47112 (36.68579) | > grad_norm_0: 68.20534 (100.85966) | > loss_disc: 2.53048 (2.37502) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53048 (2.37502) | > grad_norm_1: 19.65391 (14.53146) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.94640 (10.38550) | > loader_time: 0.01520 (0.01232)  --> STEP: 90/120 -- GLOBAL_STEP: 1300 | > loss_gen: 2.56468 (2.26994) | > loss_kl: 1.94631 (1.94200) | > loss_feat: 3.00308 (3.27466) | > loss_mel: 26.56631 (27.68278) | > loss_duration: 1.06969 (1.07949) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.15007 (36.24886) | > grad_norm_0: 43.03173 (100.51100) | > loss_disc: 2.40613 (2.39502) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.40613 (2.39502) | > grad_norm_1: 14.01895 (14.44359) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.62230 (10.61791) | > loader_time: 0.01810 (0.01352)  --> STEP: 115/120 -- GLOBAL_STEP: 1325 | > loss_gen: 2.01340 (2.25471) | > loss_kl: 1.84785 (1.92244) | > loss_feat: 3.10573 (3.23270) | > loss_mel: 28.38642 (27.54424) | > loss_duration: 1.22355 (1.08667) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 36.57695 (36.04076) | > grad_norm_0: 61.67858 (110.59438) | > loss_disc: 2.36569 (2.40428) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.36569 (2.40428) | > grad_norm_1: 19.00754 (14.55816) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.07810 (10.90350) | > loader_time: 0.02260 (0.01508) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.02684 (2.02684) | > loss_kl: 1.92597 (1.92597) | > loss_feat: 2.99879 (2.99879) | > loss_mel: 28.46849 (28.46849) | > loss_duration: 1.01295 (1.01295) | > loss_0: 36.43304 (36.43304) | > loss_disc: 2.47366 (2.47366) | > loss_1: 2.47366 (2.47366)  --> STEP: 1 | > loss_gen: 2.05462 (2.05462) | > loss_kl: 2.16871 (2.16871) | > loss_feat: 2.94889 (2.94889) | > loss_mel: 28.12627 (28.12627) | > loss_duration: 1.05807 (1.05807) | > loss_0: 36.35656 (36.35656) | > loss_disc: 2.49671 (2.49671) | > loss_1: 2.49671 (2.49671)  --> STEP: 2 | > loss_gen: 1.84034 (1.94748) | > loss_kl: 2.04392 (2.10631) | > loss_feat: 3.34398 (3.14643) | > loss_mel: 27.70117 (27.91372) | > loss_duration: 1.06594 (1.06200) | > loss_0: 35.99535 (36.17595) | > loss_disc: 2.46206 (2.47938) | > loss_1: 2.46206 (2.47938)  --> STEP: 3 | > loss_gen: 2.20319 (2.03271) | > loss_kl: 1.96972 (2.06078) | > loss_feat: 3.14746 (3.14678) | > loss_mel: 25.72879 (27.18541) | > loss_duration: 1.14603 (1.09001) | > loss_0: 34.19519 (35.51570) | > loss_disc: 2.37424 (2.44434) | > loss_1: 2.37424 (2.44434) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00756 (+0.00031) | > avg_loss_gen: 2.03271 (-0.47103) | > avg_loss_kl: 2.06078 (-0.00063) | > avg_loss_feat: 3.14678 (-0.12668) | > avg_loss_mel: 27.18541 (-0.32851) | > avg_loss_duration: 1.09001 (+0.00270) | > avg_loss_0: 35.51570 (-0.92413) | > avg_loss_disc: 2.44434 (+0.07596) | > avg_loss_1: 2.44434 (+0.07596) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1331.pth.tar  > EPOCH: 11/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 18:02:56)   --> STEP: 19/120 -- GLOBAL_STEP: 1350 | > loss_gen: 2.27255 (2.25558) | > loss_kl: 2.06941 (2.16335) | > loss_feat: 3.04420 (3.27060) | > loss_mel: 28.87636 (28.19343) | > loss_duration: 1.00184 (1.16831) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 37.26435 (37.05128) | > grad_norm_0: 99.96973 (73.65921) | > loss_disc: 2.57681 (2.49005) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57681 (2.49005) | > grad_norm_1: 14.65965 (15.16648) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.00970 (9.74874) | > loader_time: 0.00990 (0.00845)  --> STEP: 44/120 -- GLOBAL_STEP: 1375 | > loss_gen: 2.46599 (2.24589) | > loss_kl: 1.96810 (2.05764) | > loss_feat: 2.59656 (3.19126) | > loss_mel: 26.49616 (27.62371) | > loss_duration: 1.05433 (1.09246) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.58114 (36.21096) | > grad_norm_0: 205.52650 (107.43756) | > loss_disc: 2.58995 (2.47342) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58995 (2.47342) | > grad_norm_1: 17.65224 (15.14282) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.51330 (10.20679) | > loader_time: 0.01230 (0.01040)  --> STEP: 69/120 -- GLOBAL_STEP: 1400 | > loss_gen: 1.81498 (2.20546) | > loss_kl: 1.92901 (2.02189) | > loss_feat: 2.65128 (3.10179) | > loss_mel: 26.10564 (27.32141) | > loss_duration: 1.05444 (1.08280) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.55535 (35.73334) | > grad_norm_0: 60.69263 (107.93472) | > loss_disc: 2.54587 (2.47274) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54587 (2.47274) | > grad_norm_1: 20.29158 (14.17502) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.04060 (10.44261) | > loader_time: 0.01470 (0.01287)  --> STEP: 94/120 -- GLOBAL_STEP: 1425 | > loss_gen: 1.97711 (2.18409) | > loss_kl: 1.94556 (2.00261) | > loss_feat: 3.23105 (3.04335) | > loss_mel: 26.44616 (27.12835) | > loss_duration: 1.09516 (1.08313) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.69504 (35.44153) | > grad_norm_0: 88.25076 (115.80251) | > loss_disc: 2.50132 (2.48105) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50132 (2.48105) | > grad_norm_1: 15.35635 (14.83675) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.52190 (10.70341) | > loader_time: 0.02120 (0.01411)  --> STEP: 119/120 -- GLOBAL_STEP: 1450 | > loss_gen: 2.29418 (2.16446) | > loss_kl: 1.69542 (1.97658) | > loss_feat: 2.31848 (2.97484) | > loss_mel: 25.46414 (26.97634) | > loss_duration: 1.21892 (1.09355) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.99115 (35.18576) | > grad_norm_0: 97.79681 (115.19499) | > loss_disc: 2.55579 (2.49742) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55579 (2.49742) | > grad_norm_1: 19.96016 (14.30006) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.38290 (10.96281) | > loader_time: 0.02670 (0.01571) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.04515 (2.04515) | > loss_kl: 2.23935 (2.23935) | > loss_feat: 2.64235 (2.64235) | > loss_mel: 26.82300 (26.82300) | > loss_duration: 1.00514 (1.00514) | > loss_0: 34.75499 (34.75499) | > loss_disc: 2.47908 (2.47908) | > loss_1: 2.47908 (2.47908)  --> STEP: 1 | > loss_gen: 1.96671 (1.96671) | > loss_kl: 1.96778 (1.96778) | > loss_feat: 2.80994 (2.80994) | > loss_mel: 27.13933 (27.13933) | > loss_duration: 1.04228 (1.04228) | > loss_0: 34.92605 (34.92605) | > loss_disc: 2.47574 (2.47574) | > loss_1: 2.47574 (2.47574)  --> STEP: 2 | > loss_gen: 1.93189 (1.94930) | > loss_kl: 1.84750 (1.90764) | > loss_feat: 2.83376 (2.82185) | > loss_mel: 26.99764 (27.06849) | > loss_duration: 1.05651 (1.04940) | > loss_0: 34.66730 (34.79667) | > loss_disc: 2.39666 (2.43620) | > loss_1: 2.39666 (2.43620)  --> STEP: 3 | > loss_gen: 1.96476 (1.95445) | > loss_kl: 2.27228 (2.02918) | > loss_feat: 2.85994 (2.83455) | > loss_mel: 26.49423 (26.87707) | > loss_duration: 1.13343 (1.07741) | > loss_0: 34.72464 (34.77266) | > loss_disc: 2.38627 (2.41955) | > loss_1: 2.38627 (2.41955) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00718 (-0.00038) | > avg_loss_gen: 1.95445 (-0.07826) | > avg_loss_kl: 2.02918 (-0.03160) | > avg_loss_feat: 2.83455 (-0.31223) | > avg_loss_mel: 26.87707 (-0.30834) | > avg_loss_duration: 1.07741 (-0.01260) | > avg_loss_0: 34.77266 (-0.74303) | > avg_loss_disc: 2.41955 (-0.02478) | > avg_loss_1: 2.41955 (-0.02478) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1452.pth.tar  > EPOCH: 12/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 18:25:29)   --> STEP: 23/120 -- GLOBAL_STEP: 1475 | > loss_gen: 2.06198 (2.20462) | > loss_kl: 1.94797 (2.15764) | > loss_feat: 2.71214 (3.09937) | > loss_mel: 26.34801 (27.66942) | > loss_duration: 1.06380 (1.14238) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.13390 (36.27344) | > grad_norm_0: 147.04143 (95.27169) | > loss_disc: 2.55349 (2.48402) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55349 (2.48402) | > grad_norm_1: 16.45489 (15.16187) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 8.99880 (8.68877) | > loader_time: 0.01050 (0.00882)  --> STEP: 48/120 -- GLOBAL_STEP: 1500 | > loss_gen: 1.79795 (2.17515) | > loss_kl: 1.92919 (2.07063) | > loss_feat: 2.58590 (3.01862) | > loss_mel: 25.39632 (27.01056) | > loss_duration: 1.05928 (1.08496) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.76865 (35.35991) | > grad_norm_0: 145.25964 (115.42834) | > loss_disc: 2.62243 (2.48671) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.62243 (2.48671) | > grad_norm_1: 27.42083 (16.07212) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.14990 (8.89042) | > loader_time: 0.01340 (0.01059)  --> STEP: 73/120 -- GLOBAL_STEP: 1525 | > loss_gen: 2.04288 (2.15728) | > loss_kl: 2.06595 (2.02690) | > loss_feat: 2.78449 (2.95616) | > loss_mel: 24.95801 (26.82483) | > loss_duration: 1.07665 (1.07709) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.92798 (35.04225) | > grad_norm_0: 215.92270 (115.49366) | > loss_disc: 2.43191 (2.49742) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.43191 (2.49742) | > grad_norm_1: 7.97053 (15.20208) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.63090 (9.03918) | > loader_time: 0.01570 (0.01320)  --> STEP: 98/120 -- GLOBAL_STEP: 1550 | > loss_gen: 1.95609 (2.14882) | > loss_kl: 1.98876 (2.00756) | > loss_feat: 3.11076 (2.93827) | > loss_mel: 27.64729 (26.60856) | > loss_duration: 1.06670 (1.07349) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.76960 (34.77670) | > grad_norm_0: 229.50166 (112.37523) | > loss_disc: 2.33265 (2.49106) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.33265 (2.49106) | > grad_norm_1: 8.67962 (14.57195) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.64130 (9.34983) | > loader_time: 0.01840 (0.01423) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.54780 (2.54780) | > loss_kl: 1.77783 (1.77783) | > loss_feat: 2.79468 (2.79468) | > loss_mel: 27.22243 (27.22243) | > loss_duration: 0.95697 (0.95697) | > loss_0: 35.29972 (35.29972) | > loss_disc: 2.50749 (2.50749) | > loss_1: 2.50749 (2.50749)  --> STEP: 1 | > loss_gen: 2.51710 (2.51710) | > loss_kl: 1.72859 (1.72859) | > loss_feat: 2.88471 (2.88471) | > loss_mel: 28.08264 (28.08264) | > loss_duration: 1.00205 (1.00205) | > loss_0: 36.21510 (36.21510) | > loss_disc: 2.58343 (2.58343) | > loss_1: 2.58343 (2.58343)  --> STEP: 2 | > loss_gen: 2.55596 (2.53653) | > loss_kl: 1.97847 (1.85353) | > loss_feat: 2.61495 (2.74983) | > loss_mel: 25.43066 (26.75665) | > loss_duration: 1.00983 (1.00594) | > loss_0: 33.58988 (34.90249) | > loss_disc: 2.68781 (2.63562) | > loss_1: 2.68781 (2.63562)  --> STEP: 3 | > loss_gen: 2.52343 (2.53216) | > loss_kl: 1.85236 (1.85314) | > loss_feat: 2.71212 (2.73726) | > loss_mel: 25.98525 (26.49952) | > loss_duration: 1.07220 (1.02802) | > loss_0: 34.14536 (34.65011) | > loss_disc: 2.53706 (2.60277) | > loss_1: 2.53706 (2.60277) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00739 (+0.00021) | > avg_loss_gen: 2.53216 (+0.57771) | > avg_loss_kl: 1.85314 (-0.17605) | > avg_loss_feat: 2.73726 (-0.09729) | > avg_loss_mel: 26.49952 (-0.37755) | > avg_loss_duration: 1.02802 (-0.04938) | > avg_loss_0: 34.65011 (-0.12255) | > avg_loss_disc: 2.60277 (+0.18321) | > avg_loss_1: 2.60277 (+0.18321)  > EPOCH: 13/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 18:45:11)   --> STEP: 2/120 -- GLOBAL_STEP: 1575 | > loss_gen: 2.37793 (2.17856) | > loss_kl: 2.10218 (2.12072) | > loss_feat: 3.19034 (3.13582) | > loss_mel: 28.30023 (28.15055) | > loss_duration: 0.94159 (0.96975) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 36.91227 (36.55540) | > grad_norm_0: 103.72261 (101.52951) | > loss_disc: 2.45769 (2.49508) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.45769 (2.49508) | > grad_norm_1: 19.74902 (18.82166) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 8.75490 (8.51725) | > loader_time: 0.00500 (0.00765)  --> STEP: 27/120 -- GLOBAL_STEP: 1600 | > loss_gen: 2.19102 (2.13629) | > loss_kl: 2.05540 (2.09563) | > loss_feat: 3.07280 (2.98320) | > loss_mel: 27.41814 (27.00668) | > loss_duration: 0.96812 (1.09206) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.70548 (35.31386) | > grad_norm_0: 45.84895 (96.08689) | > loss_disc: 2.45188 (2.53859) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.45188 (2.53859) | > grad_norm_1: 16.65076 (15.57878) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.68040 (9.22443) | > loader_time: 0.01110 (0.00930)  --> STEP: 52/120 -- GLOBAL_STEP: 1625 | > loss_gen: 2.47265 (2.12345) | > loss_kl: 1.92858 (2.02202) | > loss_feat: 2.50843 (2.90617) | > loss_mel: 24.64459 (26.49754) | > loss_duration: 1.00432 (1.04649) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.55857 (34.59566) | > grad_norm_0: 85.58918 (106.07016) | > loss_disc: 2.76453 (2.53151) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.76453 (2.53151) | > grad_norm_1: 42.44766 (16.60475) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.97190 (9.54533) | > loader_time: 0.01370 (0.01092)  --> STEP: 77/120 -- GLOBAL_STEP: 1650 | > loss_gen: 1.77063 (2.10553) | > loss_kl: 1.72380 (1.97097) | > loss_feat: 2.56918 (2.83795) | > loss_mel: 25.12508 (26.22432) | > loss_duration: 0.98922 (1.03842) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.17790 (34.17720) | > grad_norm_0: 30.14967 (100.18408) | > loss_disc: 2.50197 (2.53145) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50197 (2.53145) | > grad_norm_1: 10.02958 (15.68937) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.48590 (9.81351) | > loader_time: 0.01580 (0.01278)  --> STEP: 102/120 -- GLOBAL_STEP: 1675 | > loss_gen: 2.07628 (2.10710) | > loss_kl: 1.74487 (1.94310) | > loss_feat: 2.70671 (2.83698) | > loss_mel: 24.57558 (26.13498) | > loss_duration: 1.00355 (1.03410) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.10698 (34.05627) | > grad_norm_0: 176.12096 (113.05856) | > loss_disc: 2.52699 (2.52522) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.52699 (2.52522) | > grad_norm_1: 6.31173 (16.03186) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.89070 (10.22322) | > loader_time: 0.01970 (0.01425) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.08583 (2.08583) | > loss_kl: 1.53776 (1.53776) | > loss_feat: 2.56156 (2.56156) | > loss_mel: 25.10919 (25.10919) | > loss_duration: 0.94355 (0.94355) | > loss_0: 32.23789 (32.23789) | > loss_disc: 2.59706 (2.59706) | > loss_1: 2.59706 (2.59706)  --> STEP: 1 | > loss_gen: 1.96251 (1.96251) | > loss_kl: 1.58222 (1.58222) | > loss_feat: 2.77848 (2.77848) | > loss_mel: 28.14298 (28.14298) | > loss_duration: 0.95706 (0.95706) | > loss_0: 35.42326 (35.42326) | > loss_disc: 2.64465 (2.64465) | > loss_1: 2.64465 (2.64465)  --> STEP: 2 | > loss_gen: 2.07085 (2.01668) | > loss_kl: 1.78237 (1.68230) | > loss_feat: 2.59364 (2.68606) | > loss_mel: 26.05655 (27.09977) | > loss_duration: 0.97807 (0.96756) | > loss_0: 33.48149 (34.45237) | > loss_disc: 2.61302 (2.62884) | > loss_1: 2.61302 (2.62884)  --> STEP: 3 | > loss_gen: 1.95116 (1.99484) | > loss_kl: 1.94617 (1.77025) | > loss_feat: 2.29987 (2.55733) | > loss_mel: 25.87872 (26.69275) | > loss_duration: 1.03489 (0.99001) | > loss_0: 33.11080 (34.00518) | > loss_disc: 2.68823 (2.64864) | > loss_1: 2.68823 (2.64864) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00725 (-0.00014) | > avg_loss_gen: 1.99484 (-0.53733) | > avg_loss_kl: 1.77025 (-0.08288) | > avg_loss_feat: 2.55733 (-0.17993) | > avg_loss_mel: 26.69275 (+0.19324) | > avg_loss_duration: 0.99001 (-0.03802) | > avg_loss_0: 34.00518 (-0.64493) | > avg_loss_disc: 2.64864 (+0.04587) | > avg_loss_1: 2.64864 (+0.04587) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1694.pth.tar  > EPOCH: 14/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 19:06:49)   --> STEP: 6/120 -- GLOBAL_STEP: 1700 | > loss_gen: 1.88560 (2.08959) | > loss_kl: 2.17133 (2.23689) | > loss_feat: 2.98793 (2.83865) | > loss_mel: 27.39992 (27.18418) | > loss_duration: 0.94477 (0.97815) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 35.38954 (35.32746) | > grad_norm_0: 111.47858 (100.38015) | > loss_disc: 2.46449 (2.62330) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.46449 (2.62330) | > grad_norm_1: 25.51287 (29.13120) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.71100 (8.83705) | > loader_time: 0.00830 (0.00717)  --> STEP: 31/120 -- GLOBAL_STEP: 1725 | > loss_gen: 1.63434 (2.07659) | > loss_kl: 1.93330 (2.07975) | > loss_feat: 2.67064 (2.76834) | > loss_mel: 24.27267 (26.16642) | > loss_duration: 0.97232 (1.06231) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.48326 (34.15342) | > grad_norm_0: 77.63988 (106.33535) | > loss_disc: 2.62736 (2.56217) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.62736 (2.56217) | > grad_norm_1: 19.03087 (16.29611) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.50850 (9.95174) | > loader_time: 0.01170 (0.00955)  --> STEP: 56/120 -- GLOBAL_STEP: 1750 | > loss_gen: 1.96259 (2.06696) | > loss_kl: 1.99194 (2.01897) | > loss_feat: 2.44419 (2.73574) | > loss_mel: 25.98207 (25.85543) | > loss_duration: 1.00757 (1.02611) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.38836 (33.70322) | > grad_norm_0: 154.75021 (101.25303) | > loss_disc: 2.60305 (2.55594) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60305 (2.55594) | > grad_norm_1: 14.66995 (15.02672) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.64690 (10.23656) | > loader_time: 0.03670 (0.01188)  --> STEP: 81/120 -- GLOBAL_STEP: 1775 | > loss_gen: 2.13723 (2.06560) | > loss_kl: 1.83339 (1.97370) | > loss_feat: 2.62865 (2.71945) | > loss_mel: 24.60712 (25.76457) | > loss_duration: 1.00368 (1.01967) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.21008 (33.54299) | > grad_norm_0: 45.23702 (102.60262) | > loss_disc: 2.56063 (2.56502) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56063 (2.56502) | > grad_norm_1: 9.26880 (16.68547) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.78040 (10.44659) | > loader_time: 0.01640 (0.01327)  --> STEP: 106/120 -- GLOBAL_STEP: 1800 | > loss_gen: 2.37480 (2.05957) | > loss_kl: 1.81398 (1.95123) | > loss_feat: 2.81907 (2.70767) | > loss_mel: 25.30836 (25.65105) | > loss_duration: 1.08690 (1.01780) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.40311 (33.38732) | > grad_norm_0: 167.01419 (113.29037) | > loss_disc: 2.51174 (2.56422) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51174 (2.56422) | > grad_norm_1: 22.85308 (15.80521) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.92990 (10.77324) | > loader_time: 0.02070 (0.01468) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.70497 (1.70497) | > loss_kl: 1.55252 (1.55252) | > loss_feat: 2.67457 (2.67457) | > loss_mel: 25.98342 (25.98342) | > loss_duration: 0.94025 (0.94025) | > loss_0: 32.85573 (32.85573) | > loss_disc: 2.46432 (2.46432) | > loss_1: 2.46432 (2.46432)  --> STEP: 1 | > loss_gen: 1.68800 (1.68800) | > loss_kl: 1.51676 (1.51676) | > loss_feat: 2.57854 (2.57854) | > loss_mel: 25.64725 (25.64725) | > loss_duration: 0.95391 (0.95391) | > loss_0: 32.38445 (32.38445) | > loss_disc: 2.53333 (2.53333) | > loss_1: 2.53333 (2.53333)  --> STEP: 2 | > loss_gen: 1.69945 (1.69373) | > loss_kl: 1.87955 (1.69815) | > loss_feat: 2.98229 (2.78042) | > loss_mel: 26.01721 (25.83223) | > loss_duration: 0.97863 (0.96627) | > loss_0: 33.55714 (32.97079) | > loss_disc: 2.43404 (2.48369) | > loss_1: 2.43404 (2.48369)  --> STEP: 3 | > loss_gen: 1.70443 (1.69730) | > loss_kl: 1.89907 (1.76513) | > loss_feat: 2.31134 (2.62406) | > loss_mel: 24.02431 (25.22959) | > loss_duration: 1.03458 (0.98904) | > loss_0: 30.97375 (32.30511) | > loss_disc: 2.56464 (2.51067) | > loss_1: 2.56464 (2.51067) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00732 (+0.00007) | > avg_loss_gen: 1.69730 (-0.29754) | > avg_loss_kl: 1.76513 (-0.00513) | > avg_loss_feat: 2.62406 (+0.06673) | > avg_loss_mel: 25.22959 (-1.46316) | > avg_loss_duration: 0.98904 (-0.00097) | > avg_loss_0: 32.30511 (-1.70007) | > avg_loss_disc: 2.51067 (-0.13796) | > avg_loss_1: 2.51067 (-0.13796) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1815.pth.tar  > EPOCH: 15/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 19:29:19)   --> STEP: 10/120 -- GLOBAL_STEP: 1825 | > loss_gen: 1.98894 (2.18036) | > loss_kl: 2.06424 (2.14248) | > loss_feat: 2.93620 (2.87777) | > loss_mel: 26.52962 (26.34199) | > loss_duration: 1.00447 (0.96331) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.52347 (34.50590) | > grad_norm_0: 37.84140 (84.32607) | > loss_disc: 2.48960 (2.55423) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48960 (2.55423) | > grad_norm_1: 12.06188 (21.50901) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.45550 (9.36316) | > loader_time: 0.00890 (0.00781)  --> STEP: 35/120 -- GLOBAL_STEP: 1850 | > loss_gen: 2.02955 (2.08614) | > loss_kl: 1.88380 (2.05455) | > loss_feat: 2.24021 (2.75372) | > loss_mel: 25.82658 (25.95886) | > loss_duration: 0.94083 (1.03871) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.92098 (33.89199) | > grad_norm_0: 79.81718 (117.80511) | > loss_disc: 2.89302 (2.58382) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.89302 (2.58382) | > grad_norm_1: 22.17787 (15.16939) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.82410 (10.18366) | > loader_time: 0.01220 (0.00993)  --> STEP: 60/120 -- GLOBAL_STEP: 1875 | > loss_gen: 1.79386 (2.05487) | > loss_kl: 1.95791 (2.01221) | > loss_feat: 3.11522 (2.67382) | > loss_mel: 26.51710 (25.74238) | > loss_duration: 0.99667 (1.01205) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.38076 (33.49532) | > grad_norm_0: 125.31999 (107.29646) | > loss_disc: 2.42112 (2.59747) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.42112 (2.59747) | > grad_norm_1: 22.16123 (17.36386) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.90410 (10.46138) | > loader_time: 0.01500 (0.01264)  --> STEP: 85/120 -- GLOBAL_STEP: 1900 | > loss_gen: 1.88401 (2.04750) | > loss_kl: 1.91284 (1.97416) | > loss_feat: 2.83309 (2.66037) | > loss_mel: 25.65963 (25.49938) | > loss_duration: 1.00327 (1.00689) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.29284 (33.18830) | > grad_norm_0: 164.34628 (117.90523) | > loss_disc: 2.72964 (2.58344) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.72964 (2.58344) | > grad_norm_1: 19.04214 (15.79286) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.74120 (10.68355) | > loader_time: 0.01630 (0.01352)  --> STEP: 110/120 -- GLOBAL_STEP: 1925 | > loss_gen: 2.07776 (2.04228) | > loss_kl: 1.84422 (1.95678) | > loss_feat: 2.34691 (2.62985) | > loss_mel: 24.27802 (25.30699) | > loss_duration: 1.02868 (1.00653) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.57559 (32.94244) | > grad_norm_0: 194.94453 (117.67485) | > loss_disc: 2.61463 (2.58205) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61463 (2.58205) | > grad_norm_1: 11.16343 (15.96848) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.20510 (10.98273) | > loader_time: 0.02420 (0.01487) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.57018 (2.57018) | > loss_kl: 1.87606 (1.87606) | > loss_feat: 2.55061 (2.55061) | > loss_mel: 26.02412 (26.02412) | > loss_duration: 0.90795 (0.90795) | > loss_0: 33.92891 (33.92891) | > loss_disc: 2.65975 (2.65975) | > loss_1: 2.65975 (2.65975)  --> STEP: 1 | > loss_gen: 2.60416 (2.60416) | > loss_kl: 1.66151 (1.66151) | > loss_feat: 2.29995 (2.29995) | > loss_mel: 23.98292 (23.98292) | > loss_duration: 0.92299 (0.92299) | > loss_0: 31.47154 (31.47154) | > loss_disc: 2.76107 (2.76107) | > loss_1: 2.76107 (2.76107)  --> STEP: 2 | > loss_gen: 2.43025 (2.51720) | > loss_kl: 2.19001 (1.92576) | > loss_feat: 2.35177 (2.32586) | > loss_mel: 22.46092 (23.22192) | > loss_duration: 0.95775 (0.94037) | > loss_0: 30.39070 (30.93112) | > loss_disc: 2.63246 (2.69676) | > loss_1: 2.63246 (2.69676)  --> STEP: 3 | > loss_gen: 2.46804 (2.50082) | > loss_kl: 1.99791 (1.94981) | > loss_feat: 2.56583 (2.40585) | > loss_mel: 23.78432 (23.40939) | > loss_duration: 1.01379 (0.96484) | > loss_0: 31.82988 (31.23071) | > loss_disc: 2.53204 (2.64186) | > loss_1: 2.53204 (2.64186) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00645 (-0.00086) | > avg_loss_gen: 2.50082 (+0.80352) | > avg_loss_kl: 1.94981 (+0.18468) | > avg_loss_feat: 2.40585 (-0.21821) | > avg_loss_mel: 23.40939 (-1.82020) | > avg_loss_duration: 0.96484 (-0.02420) | > avg_loss_0: 31.23071 (-1.07441) | > avg_loss_disc: 2.64186 (+0.13118) | > avg_loss_1: 2.64186 (+0.13118) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_1936.pth.tar  > EPOCH: 16/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 19:52:05)   --> STEP: 14/120 -- GLOBAL_STEP: 1950 | > loss_gen: 2.15697 (2.07031) | > loss_kl: 2.21293 (2.15779) | > loss_feat: 2.92509 (2.86484) | > loss_mel: 25.72082 (26.00879) | > loss_duration: 0.96301 (1.14656) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.97882 (34.24829) | > grad_norm_0: 66.68089 (89.36569) | > loss_disc: 2.45886 (2.55850) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.45886 (2.55850) | > grad_norm_1: 7.36873 (21.42453) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.19090 (9.57669) | > loader_time: 0.00940 (0.00830)  --> STEP: 39/120 -- GLOBAL_STEP: 1975 | > loss_gen: 1.75097 (2.03952) | > loss_kl: 1.84824 (2.08216) | > loss_feat: 2.96814 (2.77629) | > loss_mel: 24.98368 (25.48810) | > loss_duration: 0.94614 (1.02549) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 32.49717 (33.41157) | > grad_norm_0: 38.43781 (102.14742) | > loss_disc: 2.63870 (2.55741) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63870 (2.55741) | > grad_norm_1: 31.83186 (17.39226) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.49620 (10.14600) | > loader_time: 0.01240 (0.01019)  --> STEP: 64/120 -- GLOBAL_STEP: 2000 | > loss_gen: 1.93340 (2.04138) | > loss_kl: 1.86832 (2.02579) | > loss_feat: 2.95936 (2.73346) | > loss_mel: 25.79123 (25.23122) | > loss_duration: 1.01882 (1.00226) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.57114 (33.03410) | > grad_norm_0: 140.80792 (107.13876) | > loss_disc: 2.50244 (2.55613) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50244 (2.55613) | > grad_norm_1: 18.77151 (16.28455) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.36640 (10.39240) | > loader_time: 0.01590 (0.01314)  --> STEP: 89/120 -- GLOBAL_STEP: 2025 | > loss_gen: 1.92677 (2.03551) | > loss_kl: 1.98307 (2.00230) | > loss_feat: 2.51496 (2.68672) | > loss_mel: 23.90589 (25.04271) | > loss_duration: 0.97748 (0.99733) | > amp_scaler: 1024.00000 (627.05618) | > loss_0: 31.30817 (32.76456) | > grad_norm_0: 191.97641 (110.14297) | > loss_disc: 2.63679 (2.56472) | > amp_scaler-1: 1024.00000 (627.05618) | > loss_1: 2.63679 (2.56472) | > grad_norm_1: 9.85552 (18.30091) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.53550 (10.61124) | > loader_time: 0.01800 (0.01399)  --> STEP: 114/120 -- GLOBAL_STEP: 2050 | > loss_gen: 2.14446 (2.03183) | > loss_kl: 1.91453 (1.98976) | > loss_feat: 2.33022 (2.65555) | > loss_mel: 24.64302 (24.93383) | > loss_duration: 1.03114 (0.99780) | > amp_scaler: 1024.00000 (714.10526) | > loss_0: 32.06337 (32.60876) | > grad_norm_0: 73.51886 (111.42227) | > loss_disc: 2.72283 (2.57528) | > amp_scaler-1: 1024.00000 (714.10526) | > loss_1: 2.72283 (2.57528) | > grad_norm_1: 8.91603 (18.32761) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.01440 (10.91897) | > loader_time: 0.02120 (0.01523) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.13921 (2.13921) | > loss_kl: 1.80906 (1.80906) | > loss_feat: 2.82328 (2.82328) | > loss_mel: 24.85997 (24.85997) | > loss_duration: 0.91919 (0.91919) | > loss_0: 32.55072 (32.55072) | > loss_disc: 2.51551 (2.51551) | > loss_1: 2.51551 (2.51551)  --> STEP: 1 | > loss_gen: 2.00234 (2.00234) | > loss_kl: 2.17751 (2.17751) | > loss_feat: 2.64987 (2.64987) | > loss_mel: 25.16864 (25.16864) | > loss_duration: 0.93129 (0.93129) | > loss_0: 32.92965 (32.92965) | > loss_disc: 2.63811 (2.63811) | > loss_1: 2.63811 (2.63811)  --> STEP: 2 | > loss_gen: 2.02755 (2.01494) | > loss_kl: 1.83611 (2.00681) | > loss_feat: 2.69160 (2.67073) | > loss_mel: 23.39078 (24.27971) | > loss_duration: 0.95938 (0.94533) | > loss_0: 30.90540 (31.91753) | > loss_disc: 2.54205 (2.59008) | > loss_1: 2.54205 (2.59008)  --> STEP: 3 | > loss_gen: 2.09572 (2.04187) | > loss_kl: 1.82811 (1.94724) | > loss_feat: 2.79145 (2.71097) | > loss_mel: 24.86892 (24.47611) | > loss_duration: 1.00035 (0.96367) | > loss_0: 32.58455 (32.13987) | > loss_disc: 2.40149 (2.52722) | > loss_1: 2.40149 (2.52722) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00747 (+0.00102) | > avg_loss_gen: 2.04187 (-0.45895) | > avg_loss_kl: 1.94724 (-0.00257) | > avg_loss_feat: 2.71097 (+0.30512) | > avg_loss_mel: 24.47611 (+1.06673) | > avg_loss_duration: 0.96367 (-0.00117) | > avg_loss_0: 32.13987 (+0.90916) | > avg_loss_disc: 2.52722 (-0.11464) | > avg_loss_1: 2.52722 (-0.11464)  > EPOCH: 17/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 20:14:22)   --> STEP: 18/120 -- GLOBAL_STEP: 2075 | > loss_gen: 2.24450 (2.04196) | > loss_kl: 2.25707 (2.20859) | > loss_feat: 2.55339 (2.68852) | > loss_mel: 25.48120 (25.25744) | > loss_duration: 0.96689 (1.09774) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 33.50304 (33.29425) | > grad_norm_0: 44.78049 (70.73943) | > loss_disc: 2.64974 (2.60280) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64974 (2.60280) | > grad_norm_1: 10.39077 (15.51681) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.24480 (10.21204) | > loader_time: 0.00990 (0.00871)  --> STEP: 43/120 -- GLOBAL_STEP: 2100 | > loss_gen: 2.44989 (2.02677) | > loss_kl: 1.96148 (2.10829) | > loss_feat: 2.43504 (2.60707) | > loss_mel: 23.98089 (24.94364) | > loss_duration: 0.95766 (1.01227) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.78497 (32.69804) | > grad_norm_0: 130.92601 (99.29192) | > loss_disc: 2.64945 (2.61593) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64945 (2.61593) | > grad_norm_1: 33.42998 (18.82693) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.45690 (10.43256) | > loader_time: 0.01280 (0.01051)  --> STEP: 68/120 -- GLOBAL_STEP: 2125 | > loss_gen: 1.97830 (2.01273) | > loss_kl: 1.83693 (2.06228) | > loss_feat: 2.62780 (2.58224) | > loss_mel: 24.80831 (24.68411) | > loss_duration: 0.98940 (0.99323) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.24074 (32.33459) | > grad_norm_0: 40.68769 (96.95964) | > loss_disc: 2.56701 (2.60046) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.56701 (2.60046) | > grad_norm_1: 5.85308 (17.21379) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.90270 (10.59700) | > loader_time: 0.01530 (0.01291)  --> STEP: 93/120 -- GLOBAL_STEP: 2150 | > loss_gen: 2.03772 (2.01462) | > loss_kl: 1.83204 (2.03965) | > loss_feat: 2.63702 (2.59832) | > loss_mel: 24.74381 (24.50689) | > loss_duration: 0.95130 (0.98831) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.20190 (32.14779) | > grad_norm_0: 208.66655 (95.30539) | > loss_disc: 2.52446 (2.58628) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.52446 (2.58628) | > grad_norm_1: 9.66661 (17.73878) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.74400 (10.83310) | > loader_time: 0.01810 (0.01395)  --> STEP: 118/120 -- GLOBAL_STEP: 2175 | > loss_gen: 2.30156 (2.01309) | > loss_kl: 1.88170 (2.02119) | > loss_feat: 2.50010 (2.57373) | > loss_mel: 24.31948 (24.42845) | > loss_duration: 1.07372 (0.99266) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.07656 (32.02911) | > grad_norm_0: 220.37425 (103.90350) | > loss_disc: 2.65321 (2.59120) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.65321 (2.59120) | > grad_norm_1: 25.65397 (18.73674) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.17500 (11.11901) | > loader_time: 0.02450 (0.01538) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.21603 (2.21603) | > loss_kl: 1.73746 (1.73746) | > loss_feat: 2.35572 (2.35572) | > loss_mel: 23.75728 (23.75728) | > loss_duration: 0.88998 (0.88998) | > loss_0: 30.95647 (30.95647) | > loss_disc: 2.63019 (2.63019) | > loss_1: 2.63019 (2.63019)  --> STEP: 1 | > loss_gen: 2.10032 (2.10032) | > loss_kl: 1.99595 (1.99595) | > loss_feat: 2.83272 (2.83272) | > loss_mel: 25.67247 (25.67247) | > loss_duration: 0.90741 (0.90741) | > loss_0: 33.50886 (33.50886) | > loss_disc: 2.52839 (2.52839) | > loss_1: 2.52839 (2.52839)  --> STEP: 2 | > loss_gen: 2.15679 (2.12856) | > loss_kl: 2.20402 (2.09998) | > loss_feat: 2.56732 (2.70002) | > loss_mel: 23.75251 (24.71249) | > loss_duration: 0.94695 (0.92718) | > loss_0: 31.62759 (32.56823) | > loss_disc: 2.53680 (2.53259) | > loss_1: 2.53680 (2.53259)  --> STEP: 3 | > loss_gen: 2.13028 (2.12913) | > loss_kl: 2.08911 (2.09636) | > loss_feat: 2.73115 (2.71040) | > loss_mel: 22.96720 (24.13073) | > loss_duration: 0.99040 (0.94826) | > loss_0: 30.90814 (32.01486) | > loss_disc: 2.55720 (2.54080) | > loss_1: 2.55720 (2.54080) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00717 (-0.00030) | > avg_loss_gen: 2.12913 (+0.08726) | > avg_loss_kl: 2.09636 (+0.14911) | > avg_loss_feat: 2.71040 (-0.00058) | > avg_loss_mel: 24.13073 (-0.34539) | > avg_loss_duration: 0.94826 (-0.01542) | > avg_loss_0: 32.01486 (-0.12500) | > avg_loss_disc: 2.54080 (+0.01358) | > avg_loss_1: 2.54080 (+0.01358)  > EPOCH: 18/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 20:36:55)   --> STEP: 22/120 -- GLOBAL_STEP: 2200 | > loss_gen: 1.77316 (2.05534) | > loss_kl: 2.21088 (2.22390) | > loss_feat: 2.73119 (2.79678) | > loss_mel: 24.27327 (24.93189) | > loss_duration: 0.99240 (1.06415) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.98090 (33.07207) | > grad_norm_0: 106.71821 (119.29128) | > loss_disc: 2.61506 (2.56586) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.61506 (2.56586) | > grad_norm_1: 7.75540 (18.80066) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.76550 (10.32944) | > loader_time: 0.01030 (0.00883)  --> STEP: 47/120 -- GLOBAL_STEP: 2225 | > loss_gen: 1.84410 (2.03225) | > loss_kl: 2.10488 (2.12282) | > loss_feat: 2.56946 (2.67502) | > loss_mel: 23.96355 (24.39634) | > loss_duration: 0.90600 (0.99735) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.38799 (32.22378) | > grad_norm_0: 61.85730 (108.03471) | > loss_disc: 2.57352 (2.57285) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.57352 (2.57285) | > grad_norm_1: 30.90850 (19.16967) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.32840 (10.53640) | > loader_time: 0.01360 (0.01065)  --> STEP: 72/120 -- GLOBAL_STEP: 2250 | > loss_gen: 2.65626 (2.02539) | > loss_kl: 2.01197 (2.09564) | > loss_feat: 2.34198 (2.63870) | > loss_mel: 24.90688 (24.28001) | > loss_duration: 0.99454 (0.98640) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.91164 (32.02613) | > grad_norm_0: 113.49279 (104.59299) | > loss_disc: 2.93652 (2.58962) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.93652 (2.58962) | > grad_norm_1: 59.64648 (19.10537) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93030 (10.69059) | > loader_time: 0.01750 (0.01289)  --> STEP: 97/120 -- GLOBAL_STEP: 2275 | > loss_gen: 1.87088 (2.00761) | > loss_kl: 1.81696 (2.06427) | > loss_feat: 2.52895 (2.58276) | > loss_mel: 23.44612 (24.16022) | > loss_duration: 0.96532 (0.98144) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.62823 (31.79631) | > grad_norm_0: 142.02188 (99.23466) | > loss_disc: 2.52238 (2.60344) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.52238 (2.60344) | > grad_norm_1: 19.97903 (19.21084) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.17410 (10.93245) | > loader_time: 0.02180 (0.01439) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.17602 (2.17602) | > loss_kl: 2.28211 (2.28211) | > loss_feat: 2.34916 (2.34916) | > loss_mel: 23.10669 (23.10669) | > loss_duration: 0.88877 (0.88877) | > loss_0: 30.80275 (30.80275) | > loss_disc: 2.75321 (2.75321) | > loss_1: 2.75321 (2.75321)  --> STEP: 1 | > loss_gen: 2.16114 (2.16114) | > loss_kl: 1.95325 (1.95325) | > loss_feat: 2.31937 (2.31937) | > loss_mel: 24.18423 (24.18423) | > loss_duration: 0.88806 (0.88806) | > loss_0: 31.50605 (31.50605) | > loss_disc: 2.65579 (2.65579) | > loss_1: 2.65579 (2.65579)  --> STEP: 2 | > loss_gen: 2.02243 (2.09179) | > loss_kl: 2.16534 (2.05929) | > loss_feat: 2.43732 (2.37834) | > loss_mel: 24.04842 (24.11633) | > loss_duration: 0.92238 (0.90522) | > loss_0: 31.59589 (31.55097) | > loss_disc: 2.81314 (2.73447) | > loss_1: 2.81314 (2.73447)  --> STEP: 3 | > loss_gen: 2.05338 (2.07898) | > loss_kl: 1.83131 (1.98330) | > loss_feat: 2.39868 (2.38512) | > loss_mel: 23.62436 (23.95234) | > loss_duration: 0.99360 (0.93468) | > loss_0: 30.90132 (31.33442) | > loss_disc: 2.73492 (2.73462) | > loss_1: 2.73492 (2.73462) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00741 (+0.00024) | > avg_loss_gen: 2.07898 (-0.05015) | > avg_loss_kl: 1.98330 (-0.11306) | > avg_loss_feat: 2.38512 (-0.32527) | > avg_loss_mel: 23.95234 (-0.17839) | > avg_loss_duration: 0.93468 (-0.01358) | > avg_loss_0: 31.33442 (-0.68045) | > avg_loss_disc: 2.73462 (+0.19382) | > avg_loss_1: 2.73462 (+0.19382)  > EPOCH: 19/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 20:59:36)   --> STEP: 1/120 -- GLOBAL_STEP: 2300 | > loss_gen: 1.93914 (1.93914) | > loss_kl: 2.50777 (2.50777) | > loss_feat: 2.99779 (2.99779) | > loss_mel: 26.02859 (26.02859) | > loss_duration: 0.91882 (0.91882) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 34.39212 (34.39212) | > grad_norm_0: 108.30675 (108.30675) | > loss_disc: 2.51548 (2.51548) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.51548 (2.51548) | > grad_norm_1: 29.66887 (29.66887) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.64550 (9.64545) | > loader_time: 0.00510 (0.00508)  --> STEP: 26/120 -- GLOBAL_STEP: 2325 | > loss_gen: 1.75330 (2.01108) | > loss_kl: 2.22729 (2.30570) | > loss_feat: 2.76231 (2.69161) | > loss_mel: 24.73343 (24.50172) | > loss_duration: 0.92536 (1.04463) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.40168 (32.55474) | > grad_norm_0: 77.45090 (104.32900) | > loss_disc: 2.63535 (2.58201) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.63535 (2.58201) | > grad_norm_1: 14.71698 (20.30330) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.84170 (10.40916) | > loader_time: 0.01150 (0.00928)  --> STEP: 51/120 -- GLOBAL_STEP: 2350 | > loss_gen: 1.87651 (2.02217) | > loss_kl: 1.90854 (2.18801) | > loss_feat: 2.53165 (2.66674) | > loss_mel: 23.61066 (24.11368) | > loss_duration: 0.94620 (0.99159) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.87355 (31.98219) | > grad_norm_0: 56.18275 (90.59438) | > loss_disc: 2.60855 (2.58025) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60855 (2.58025) | > grad_norm_1: 39.21887 (22.50038) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.19440 (10.61284) | > loader_time: 0.01340 (0.01095)  --> STEP: 76/120 -- GLOBAL_STEP: 2375 | > loss_gen: 1.89766 (2.00691) | > loss_kl: 1.96273 (2.14043) | > loss_feat: 2.28548 (2.61500) | > loss_mel: 23.43146 (24.03933) | > loss_duration: 0.97863 (0.98029) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.55596 (31.78195) | > grad_norm_0: 182.14769 (101.83534) | > loss_disc: 2.58711 (2.59291) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58711 (2.59291) | > grad_norm_1: 6.34521 (22.26155) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.62060 (10.77081) | > loader_time: 0.01620 (0.01323)  --> STEP: 101/120 -- GLOBAL_STEP: 2400 | > loss_gen: 1.89469 (2.00769) | > loss_kl: 1.99750 (2.11088) | > loss_feat: 2.56665 (2.59183) | > loss_mel: 22.68061 (23.87289) | > loss_duration: 0.97937 (0.97591) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.11883 (31.55919) | > grad_norm_0: 147.37845 (108.33665) | > loss_disc: 2.58874 (2.59471) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58874 (2.59471) | > grad_norm_1: 14.74237 (22.84088) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.40340 (11.03431) | > loader_time: 0.02020 (0.01437) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.20395 (2.20395) | > loss_kl: 1.94410 (1.94410) | > loss_feat: 2.56941 (2.56941) | > loss_mel: 23.88837 (23.88837) | > loss_duration: 0.89725 (0.89725) | > loss_0: 31.50308 (31.50308) | > loss_disc: 2.55467 (2.55467) | > loss_1: 2.55467 (2.55467)  --> STEP: 1 | > loss_gen: 2.06573 (2.06573) | > loss_kl: 1.69973 (1.69973) | > loss_feat: 2.11161 (2.11161) | > loss_mel: 23.01453 (23.01453) | > loss_duration: 0.89901 (0.89901) | > loss_0: 29.79060 (29.79060) | > loss_disc: 2.78122 (2.78122) | > loss_1: 2.78122 (2.78122)  --> STEP: 2 | > loss_gen: 2.04827 (2.05700) | > loss_kl: 1.97101 (1.83537) | > loss_feat: 2.40034 (2.25598) | > loss_mel: 22.62585 (22.82019) | > loss_duration: 0.93398 (0.91649) | > loss_0: 29.97945 (29.88503) | > loss_disc: 2.68338 (2.73230) | > loss_1: 2.68338 (2.73230)  --> STEP: 3 | > loss_gen: 2.02694 (2.04698) | > loss_kl: 2.31319 (1.99464) | > loss_feat: 2.55582 (2.35592) | > loss_mel: 21.16570 (22.26869) | > loss_duration: 0.98119 (0.93806) | > loss_0: 29.04284 (29.60430) | > loss_disc: 2.66289 (2.70916) | > loss_1: 2.66289 (2.70916) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00737 (-0.00004) | > avg_loss_gen: 2.04698 (-0.03200) | > avg_loss_kl: 1.99464 (+0.01135) | > avg_loss_feat: 2.35592 (-0.02920) | > avg_loss_mel: 22.26869 (-1.68364) | > avg_loss_duration: 0.93806 (+0.00338) | > avg_loss_0: 29.60430 (-1.73012) | > avg_loss_disc: 2.70916 (-0.02545) | > avg_loss_1: 2.70916 (-0.02545) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_2420.pth.tar  > EPOCH: 20/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 21:22:41)   --> STEP: 5/120 -- GLOBAL_STEP: 2425 | > loss_gen: 1.86009 (2.04177) | > loss_kl: 2.20213 (2.37042) | > loss_feat: 3.10585 (2.90508) | > loss_mel: 24.18397 (24.53030) | > loss_duration: 0.91171 (0.91908) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 32.26375 (32.76664) | > grad_norm_0: 43.70682 (105.12679) | > loss_disc: 2.50960 (2.54818) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.50960 (2.54818) | > grad_norm_1: 8.84496 (22.83078) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.82610 (8.77349) | > loader_time: 0.00820 (0.00660)  --> STEP: 30/120 -- GLOBAL_STEP: 2450 | > loss_gen: 2.19317 (2.02584) | > loss_kl: 2.15129 (2.26797) | > loss_feat: 2.95408 (2.73927) | > loss_mel: 23.09026 (23.94747) | > loss_duration: 0.90033 (1.01727) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.28912 (31.99781) | > grad_norm_0: 185.16513 (118.74539) | > loss_disc: 2.54913 (2.57828) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54913 (2.57828) | > grad_norm_1: 12.25850 (19.72057) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.68510 (10.15235) | > loader_time: 0.01210 (0.00950)  --> STEP: 55/120 -- GLOBAL_STEP: 2475 | > loss_gen: 1.87984 (2.00812) | > loss_kl: 2.10367 (2.18477) | > loss_feat: 2.38815 (2.63676) | > loss_mel: 23.61930 (23.71951) | > loss_duration: 0.94492 (0.97995) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.93589 (31.52911) | > grad_norm_0: 46.16772 (125.98110) | > loss_disc: 2.64981 (2.59503) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64981 (2.59503) | > grad_norm_1: 29.01280 (20.20756) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.01220 (10.46894) | > loader_time: 0.03670 (0.01149)  --> STEP: 80/120 -- GLOBAL_STEP: 2500 | > loss_gen: 1.99441 (2.01123) | > loss_kl: 2.04764 (2.13869) | > loss_feat: 2.58173 (2.63295) | > loss_mel: 22.25992 (23.56712) | > loss_duration: 0.94931 (0.97265) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.83302 (31.32264) | > grad_norm_0: 58.50789 (120.71889) | > loss_disc: 2.58296 (2.57942) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58296 (2.57942) | > grad_norm_1: 17.99738 (20.95134) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.74220 (10.68078) | > loader_time: 0.01650 (0.01321)  --> STEP: 105/120 -- GLOBAL_STEP: 2525 | > loss_gen: 2.02141 (2.00680) | > loss_kl: 2.12104 (2.11942) | > loss_feat: 2.25916 (2.60655) | > loss_mel: 22.95879 (23.49711) | > loss_duration: 0.99496 (0.96945) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.35536 (31.19933) | > grad_norm_0: 197.04805 (116.88618) | > loss_disc: 2.68042 (2.59295) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.68042 (2.59295) | > grad_norm_1: 13.04183 (22.92958) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.82140 (10.95313) | > loader_time: 0.02000 (0.01438) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.03309 (2.03309) | > loss_kl: 1.91331 (1.91331) | > loss_feat: 2.41887 (2.41887) | > loss_mel: 23.57221 (23.57221) | > loss_duration: 0.88503 (0.88503) | > loss_0: 30.82251 (30.82251) | > loss_disc: 2.55461 (2.55461) | > loss_1: 2.55461 (2.55461)  --> STEP: 1 | > loss_gen: 1.96914 (1.96914) | > loss_kl: 2.17661 (2.17661) | > loss_feat: 2.51836 (2.51836) | > loss_mel: 24.02989 (24.02989) | > loss_duration: 0.88849 (0.88849) | > loss_0: 31.58250 (31.58250) | > loss_disc: 2.74734 (2.74734) | > loss_1: 2.74734 (2.74734)  --> STEP: 2 | > loss_gen: 2.01437 (1.99176) | > loss_kl: 2.02229 (2.09945) | > loss_feat: 2.72814 (2.62325) | > loss_mel: 23.45414 (23.74201) | > loss_duration: 0.92469 (0.90659) | > loss_0: 31.14363 (31.36306) | > loss_disc: 2.59046 (2.66890) | > loss_1: 2.59046 (2.66890)  --> STEP: 3 | > loss_gen: 1.98910 (1.99087) | > loss_kl: 1.97510 (2.05800) | > loss_feat: 2.67204 (2.63952) | > loss_mel: 23.89078 (23.79160) | > loss_duration: 0.97661 (0.92993) | > loss_0: 31.50364 (31.40992) | > loss_disc: 2.56344 (2.63375) | > loss_1: 2.56344 (2.63375) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00721 (-0.00016) | > avg_loss_gen: 1.99087 (-0.05611) | > avg_loss_kl: 2.05800 (+0.06336) | > avg_loss_feat: 2.63952 (+0.28359) | > avg_loss_mel: 23.79160 (+1.52291) | > avg_loss_duration: 0.92993 (-0.00813) | > avg_loss_0: 31.40992 (+1.80562) | > avg_loss_disc: 2.63375 (-0.07542) | > avg_loss_1: 2.63375 (-0.07542)  > EPOCH: 21/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 21:45:13)   --> STEP: 9/120 -- GLOBAL_STEP: 2550 | > loss_gen: 1.80128 (2.02908) | > loss_kl: 2.29527 (2.30180) | > loss_feat: 2.40511 (2.71075) | > loss_mel: 24.10293 (24.43659) | > loss_duration: 0.92345 (0.93191) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.52805 (32.41013) | > grad_norm_0: 45.92817 (97.80465) | > loss_disc: 2.62489 (2.64348) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.62489 (2.64348) | > grad_norm_1: 11.45447 (17.48542) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.09910 (10.13507) | > loader_time: 0.00860 (0.00763)  --> STEP: 34/120 -- GLOBAL_STEP: 2575 | > loss_gen: 2.08593 (1.99834) | > loss_kl: 1.98500 (2.21797) | > loss_feat: 2.43478 (2.57360) | > loss_mel: 22.78493 (23.78931) | > loss_duration: 0.92337 (1.00581) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.21402 (31.58502) | > grad_norm_0: 24.99683 (97.31084) | > loss_disc: 2.70912 (2.65741) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.70912 (2.65741) | > grad_norm_1: 10.69697 (20.41560) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.01090 (10.43080) | > loader_time: 0.01190 (0.00990)  --> STEP: 59/120 -- GLOBAL_STEP: 2600 | > loss_gen: 2.04640 (1.98987) | > loss_kl: 2.15396 (2.16407) | > loss_feat: 2.72714 (2.55170) | > loss_mel: 23.27196 (23.50798) | > loss_duration: 0.92495 (0.97478) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.12440 (31.18840) | > grad_norm_0: 111.67448 (114.00628) | > loss_disc: 2.50896 (2.62986) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.50896 (2.62986) | > grad_norm_1: 10.87454 (19.65756) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.87040 (10.59123) | > loader_time: 0.01430 (0.01175)  --> STEP: 84/120 -- GLOBAL_STEP: 2625 | > loss_gen: 1.62116 (1.98547) | > loss_kl: 2.09338 (2.13415) | > loss_feat: 2.72789 (2.54004) | > loss_mel: 23.22311 (23.29274) | > loss_duration: 0.93282 (0.96797) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.59836 (30.92037) | > grad_norm_0: 48.22216 (111.43661) | > loss_disc: 2.59095 (2.62753) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.59095 (2.62753) | > grad_norm_1: 20.08624 (21.17866) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.86820 (10.75673) | > loader_time: 0.01660 (0.01288)  --> STEP: 109/120 -- GLOBAL_STEP: 2650 | > loss_gen: 1.81024 (1.99043) | > loss_kl: 2.04472 (2.11486) | > loss_feat: 2.39846 (2.54393) | > loss_mel: 21.55443 (23.14116) | > loss_duration: 1.01451 (0.96719) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.82236 (30.75758) | > grad_norm_0: 33.11410 (108.77967) | > loss_disc: 2.54355 (2.61146) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54355 (2.61146) | > grad_norm_1: 32.39971 (20.55058) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.02130 (11.03462) | > loader_time: 0.02040 (0.01430) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.39614 (2.39614) | > loss_kl: 2.33201 (2.33201) | > loss_feat: 2.74972 (2.74972) | > loss_mel: 23.41900 (23.41900) | > loss_duration: 0.88705 (0.88705) | > loss_0: 31.78392 (31.78392) | > loss_disc: 2.53665 (2.53665) | > loss_1: 2.53665 (2.53665)  --> STEP: 1 | > loss_gen: 2.33487 (2.33487) | > loss_kl: 1.88451 (1.88451) | > loss_feat: 2.75002 (2.75002) | > loss_mel: 23.54505 (23.54505) | > loss_duration: 0.87798 (0.87798) | > loss_0: 31.39243 (31.39243) | > loss_disc: 2.66687 (2.66687) | > loss_1: 2.66687 (2.66687)  --> STEP: 2 | > loss_gen: 2.48948 (2.41217) | > loss_kl: 2.31435 (2.09943) | > loss_feat: 2.74341 (2.74671) | > loss_mel: 21.95555 (22.75030) | > loss_duration: 0.91394 (0.89596) | > loss_0: 30.41672 (30.90458) | > loss_disc: 2.46136 (2.56412) | > loss_1: 2.46136 (2.56412)  --> STEP: 3 | > loss_gen: 2.30348 (2.37594) | > loss_kl: 2.21217 (2.13701) | > loss_feat: 3.28260 (2.92534) | > loss_mel: 23.34452 (22.94838) | > loss_duration: 0.96601 (0.91931) | > loss_0: 32.10878 (31.30598) | > loss_disc: 2.32344 (2.48389) | > loss_1: 2.32344 (2.48389) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00702 (-0.00019) | > avg_loss_gen: 2.37594 (+0.38507) | > avg_loss_kl: 2.13701 (+0.07901) | > avg_loss_feat: 2.92534 (+0.28582) | > avg_loss_mel: 22.94838 (-0.84323) | > avg_loss_duration: 0.91931 (-0.01062) | > avg_loss_0: 31.30598 (-0.10395) | > avg_loss_disc: 2.48389 (-0.14985) | > avg_loss_1: 2.48389 (-0.14985)  > EPOCH: 22/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 22:07:48)   --> STEP: 13/120 -- GLOBAL_STEP: 2675 | > loss_gen: 1.97429 (1.99145) | > loss_kl: 2.10105 (2.29934) | > loss_feat: 2.67010 (2.72058) | > loss_mel: 23.15377 (23.46982) | > loss_duration: 0.92849 (1.13298) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.82771 (31.61417) | > grad_norm_0: 104.00014 (95.02117) | > loss_disc: 2.63337 (2.62623) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.63337 (2.62623) | > grad_norm_1: 11.16364 (20.83611) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.40160 (10.26564) | > loader_time: 0.00960 (0.00816)  --> STEP: 38/120 -- GLOBAL_STEP: 2700 | > loss_gen: 2.22639 (1.99887) | > loss_kl: 1.97281 (2.21576) | > loss_feat: 2.09651 (2.63539) | > loss_mel: 22.10769 (23.16575) | > loss_duration: 0.92983 (0.99791) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.33323 (31.01367) | > grad_norm_0: 157.70927 (116.22647) | > loss_disc: 2.76925 (2.61237) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.76925 (2.61237) | > grad_norm_1: 27.53605 (24.84649) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.09850 (10.49636) | > loader_time: 0.01240 (0.01012)  --> STEP: 63/120 -- GLOBAL_STEP: 2725 | > loss_gen: 2.00892 (1.98268) | > loss_kl: 2.18089 (2.17869) | > loss_feat: 2.44232 (2.57245) | > loss_mel: 23.09795 (23.12497) | > loss_duration: 0.97547 (0.97126) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.70554 (30.83005) | > grad_norm_0: 99.52990 (128.26999) | > loss_disc: 2.61505 (2.61930) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.61505 (2.61930) | > grad_norm_1: 12.07510 (23.02051) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.03290 (10.64354) | > loader_time: 0.01450 (0.01267)  --> STEP: 88/120 -- GLOBAL_STEP: 2750 | > loss_gen: 2.14520 (1.98640) | > loss_kl: 2.11104 (2.14825) | > loss_feat: 2.53513 (2.57520) | > loss_mel: 23.80670 (23.09378) | > loss_duration: 0.96214 (0.96481) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.56021 (30.76844) | > grad_norm_0: 149.41585 (131.60559) | > loss_disc: 2.53936 (2.60698) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.53936 (2.60698) | > grad_norm_1: 13.19633 (22.62239) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.59590 (10.83325) | > loader_time: 0.01830 (0.01382)  --> STEP: 113/120 -- GLOBAL_STEP: 2775 | > loss_gen: 2.11018 (1.98471) | > loss_kl: 1.89821 (2.12195) | > loss_feat: 2.42467 (2.56697) | > loss_mel: 22.74148 (23.01318) | > loss_duration: 0.99457 (0.96509) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.16911 (30.65189) | > grad_norm_0: 97.04692 (121.13400) | > loss_disc: 2.64347 (2.60450) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64347 (2.60450) | > grad_norm_1: 44.02097 (22.19899) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.11080 (11.09547) | > loader_time: 0.02080 (0.01518) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.49418 (2.49418) | > loss_kl: 2.12572 (2.12572) | > loss_feat: 2.53007 (2.53007) | > loss_mel: 22.32837 (22.32837) | > loss_duration: 0.89059 (0.89059) | > loss_0: 30.36893 (30.36893) | > loss_disc: 2.64257 (2.64257) | > loss_1: 2.64257 (2.64257)  --> STEP: 1 | > loss_gen: 2.35220 (2.35220) | > loss_kl: 2.23908 (2.23908) | > loss_feat: 2.55416 (2.55416) | > loss_mel: 24.18766 (24.18766) | > loss_duration: 0.88054 (0.88054) | > loss_0: 32.21363 (32.21363) | > loss_disc: 2.62240 (2.62240) | > loss_1: 2.62240 (2.62240)  --> STEP: 2 | > loss_gen: 2.49681 (2.42451) | > loss_kl: 1.61791 (1.92849) | > loss_feat: 2.32183 (2.43800) | > loss_mel: 21.55275 (22.87021) | > loss_duration: 0.91072 (0.89563) | > loss_0: 28.90003 (30.55683) | > loss_disc: 2.60435 (2.61338) | > loss_1: 2.60435 (2.61338)  --> STEP: 3 | > loss_gen: 2.42613 (2.42505) | > loss_kl: 2.05056 (1.96918) | > loss_feat: 2.35523 (2.41041) | > loss_mel: 20.71416 (22.15152) | > loss_duration: 0.98272 (0.92466) | > loss_0: 28.52880 (29.88082) | > loss_disc: 2.65640 (2.62772) | > loss_1: 2.65640 (2.62772) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00707 (+0.00005) | > avg_loss_gen: 2.42505 (+0.04910) | > avg_loss_kl: 1.96918 (-0.16782) | > avg_loss_feat: 2.41041 (-0.51493) | > avg_loss_mel: 22.15152 (-0.79685) | > avg_loss_duration: 0.92466 (+0.00535) | > avg_loss_0: 29.88082 (-1.42516) | > avg_loss_disc: 2.62772 (+0.14383) | > avg_loss_1: 2.62772 (+0.14383)  > EPOCH: 23/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 22:30:26)   --> STEP: 17/120 -- GLOBAL_STEP: 2800 | > loss_gen: 1.75768 (2.00471) | > loss_kl: 1.94175 (2.33741) | > loss_feat: 2.39977 (2.70986) | > loss_mel: 23.25772 (23.52885) | > loss_duration: 0.93892 (1.08335) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.29584 (31.66419) | > grad_norm_0: 31.88881 (96.41141) | > loss_disc: 2.58524 (2.65542) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58524 (2.65542) | > grad_norm_1: 16.10292 (22.17141) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.58920 (10.28765) | > loader_time: 0.01040 (0.00860)  --> STEP: 42/120 -- GLOBAL_STEP: 2825 | > loss_gen: 1.87849 (1.98613) | > loss_kl: 1.99344 (2.24956) | > loss_feat: 2.58557 (2.59671) | > loss_mel: 23.30325 (23.13958) | > loss_duration: 0.92838 (0.98823) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.68913 (30.96020) | > grad_norm_0: 67.87041 (89.13699) | > loss_disc: 2.51987 (2.62447) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.51987 (2.62447) | > grad_norm_1: 15.21855 (21.89924) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.68920 (10.49044) | > loader_time: 0.01280 (0.01033)  --> STEP: 67/120 -- GLOBAL_STEP: 2850 | > loss_gen: 2.09826 (1.98855) | > loss_kl: 2.12889 (2.19705) | > loss_feat: 2.75063 (2.58917) | > loss_mel: 22.74165 (22.92101) | > loss_duration: 0.92221 (0.96759) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.64163 (30.66337) | > grad_norm_0: 66.52359 (97.39475) | > loss_disc: 2.52844 (2.61390) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.52844 (2.61390) | > grad_norm_1: 9.63703 (22.45764) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.77680 (10.63387) | > loader_time: 0.01520 (0.01233)  --> STEP: 92/120 -- GLOBAL_STEP: 2875 | > loss_gen: 2.13433 (1.98898) | > loss_kl: 2.00510 (2.15864) | > loss_feat: 2.27314 (2.57152) | > loss_mel: 21.40586 (22.83079) | > loss_duration: 0.94684 (0.96232) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.76527 (30.51225) | > grad_norm_0: 107.54881 (109.13636) | > loss_disc: 2.70299 (2.62146) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.70299 (2.62146) | > grad_norm_1: 19.98232 (23.72852) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.50740 (10.82505) | > loader_time: 0.01820 (0.01345)  --> STEP: 117/120 -- GLOBAL_STEP: 2900 | > loss_gen: 1.81073 (1.98321) | > loss_kl: 2.09941 (2.13943) | > loss_feat: 2.54664 (2.57034) | > loss_mel: 23.45234 (22.77127) | > loss_duration: 1.06879 (0.96555) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.97790 (30.42980) | > grad_norm_0: 125.76331 (104.18092) | > loss_disc: 2.71435 (2.61566) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.71435 (2.61566) | > grad_norm_1: 13.58759 (22.39750) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.03790 (11.08424) | > loader_time: 0.02320 (0.01489) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.26253 (2.26253) | > loss_kl: 2.16580 (2.16580) | > loss_feat: 2.24365 (2.24365) | > loss_mel: 22.72692 (22.72692) | > loss_duration: 0.88276 (0.88276) | > loss_0: 30.28166 (30.28166) | > loss_disc: 2.66421 (2.66421) | > loss_1: 2.66421 (2.66421)  --> STEP: 1 | > loss_gen: 2.23831 (2.23831) | > loss_kl: 2.20830 (2.20830) | > loss_feat: 2.22179 (2.22179) | > loss_mel: 23.48986 (23.48986) | > loss_duration: 0.88654 (0.88654) | > loss_0: 31.04482 (31.04482) | > loss_disc: 2.72261 (2.72261) | > loss_1: 2.72261 (2.72261)  --> STEP: 2 | > loss_gen: 2.16367 (2.20099) | > loss_kl: 1.86727 (2.03779) | > loss_feat: 2.27178 (2.24679) | > loss_mel: 22.51155 (23.00071) | > loss_duration: 0.92315 (0.90485) | > loss_0: 29.73742 (30.39112) | > loss_disc: 2.70077 (2.71169) | > loss_1: 2.70077 (2.71169)  --> STEP: 3 | > loss_gen: 2.07288 (2.15828) | > loss_kl: 1.98065 (2.01874) | > loss_feat: 2.50624 (2.33327) | > loss_mel: 22.32326 (22.77489) | > loss_duration: 0.96905 (0.92625) | > loss_0: 29.85207 (30.21144) | > loss_disc: 2.58732 (2.67023) | > loss_1: 2.58732 (2.67023) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00649 (-0.00058) | > avg_loss_gen: 2.15828 (-0.26676) | > avg_loss_kl: 2.01874 (+0.04956) | > avg_loss_feat: 2.33327 (-0.07714) | > avg_loss_mel: 22.77489 (+0.62337) | > avg_loss_duration: 0.92625 (+0.00159) | > avg_loss_0: 30.21144 (+0.33061) | > avg_loss_disc: 2.67023 (+0.04251) | > avg_loss_1: 2.67023 (+0.04251)  > EPOCH: 24/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 22:52:57)   --> STEP: 21/120 -- GLOBAL_STEP: 2925 | > loss_gen: 1.98298 (2.00300) | > loss_kl: 1.96361 (2.33258) | > loss_feat: 2.33163 (2.69832) | > loss_mel: 23.66275 (23.35597) | > loss_duration: 0.91499 (1.04157) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.85595 (31.43144) | > grad_norm_0: 70.67574 (93.31784) | > loss_disc: 2.72354 (2.63374) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.72354 (2.63374) | > grad_norm_1: 17.20971 (19.10550) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.23140 (10.18855) | > loader_time: 0.01080 (0.00897)  --> STEP: 46/120 -- GLOBAL_STEP: 2950 | > loss_gen: 2.10405 (1.97156) | > loss_kl: 2.06425 (2.26147) | > loss_feat: 2.31242 (2.56562) | > loss_mel: 22.73384 (23.02824) | > loss_duration: 0.91470 (0.97582) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.12926 (30.80270) | > grad_norm_0: 128.74690 (101.49056) | > loss_disc: 2.61855 (2.63914) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.61855 (2.63914) | > grad_norm_1: 16.84636 (20.51381) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.61650 (10.37968) | > loader_time: 0.01320 (0.01067)  --> STEP: 71/120 -- GLOBAL_STEP: 2975 | > loss_gen: 2.18626 (1.96958) | > loss_kl: 2.05288 (2.19721) | > loss_feat: 2.60307 (2.55720) | > loss_mel: 22.02198 (22.84810) | > loss_duration: 0.97367 (0.96309) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.83787 (30.53518) | > grad_norm_0: 76.06900 (112.85116) | > loss_disc: 2.53007 (2.62717) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.53007 (2.62717) | > grad_norm_1: 13.65319 (20.68434) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.88820 (10.54281) | > loader_time: 0.01570 (0.01316)  --> STEP: 96/120 -- GLOBAL_STEP: 3000 | > loss_gen: 2.20199 (1.97379) | > loss_kl: 2.11617 (2.16337) | > loss_feat: 2.24189 (2.56768) | > loss_mel: 22.14034 (22.69145) | > loss_duration: 0.92858 (0.95831) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.62895 (30.35460) | > grad_norm_0: 179.28160 (107.20855) | > loss_disc: 2.68600 (2.62055) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.68600 (2.62055) | > grad_norm_1: 28.42599 (21.56733) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.89250 (10.77662) | > loader_time: 0.01860 (0.01414) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.98364 (1.98364) | > loss_kl: 1.92958 (1.92958) | > loss_feat: 2.28932 (2.28932) | > loss_mel: 22.62321 (22.62321) | > loss_duration: 0.88074 (0.88074) | > loss_0: 29.70649 (29.70649) | > loss_disc: 2.63650 (2.63650) | > loss_1: 2.63650 (2.63650)  --> STEP: 1 | > loss_gen: 1.97084 (1.97084) | > loss_kl: 2.08741 (2.08741) | > loss_feat: 2.48626 (2.48626) | > loss_mel: 23.50900 (23.50900) | > loss_duration: 0.88150 (0.88150) | > loss_0: 30.93501 (30.93501) | > loss_disc: 2.64781 (2.64781) | > loss_1: 2.64781 (2.64781)  --> STEP: 2 | > loss_gen: 1.91678 (1.94381) | > loss_kl: 1.97985 (2.03363) | > loss_feat: 2.18547 (2.33587) | > loss_mel: 21.23184 (22.37042) | > loss_duration: 0.90712 (0.89431) | > loss_0: 28.22105 (29.57803) | > loss_disc: 2.62885 (2.63833) | > loss_1: 2.62885 (2.63833)  --> STEP: 3 | > loss_gen: 1.98427 (1.95729) | > loss_kl: 1.96603 (2.01110) | > loss_feat: 3.00365 (2.55846) | > loss_mel: 21.79935 (22.18006) | > loss_duration: 0.96093 (0.91651) | > loss_0: 29.71422 (29.62343) | > loss_disc: 2.47447 (2.58371) | > loss_1: 2.47447 (2.58371) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00730 (+0.00081) | > avg_loss_gen: 1.95729 (-0.20099) | > avg_loss_kl: 2.01110 (-0.00764) | > avg_loss_feat: 2.55846 (+0.22519) | > avg_loss_mel: 22.18006 (-0.59483) | > avg_loss_duration: 0.91651 (-0.00973) | > avg_loss_0: 29.62343 (-0.58801) | > avg_loss_disc: 2.58371 (-0.08652) | > avg_loss_1: 2.58371 (-0.08652) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_3025.pth.tar  > EPOCH: 25/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 23:15:41)   --> STEP: 0/120 -- GLOBAL_STEP: 3025 | > loss_gen: 1.99074 (1.99074) | > loss_kl: 2.16567 (2.16567) | > loss_feat: 2.87701 (2.87701) | > loss_mel: 23.78377 (23.78377) | > loss_duration: 0.89541 (0.89541) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.71261 (31.71261) | > grad_norm_0: 53.22063 (53.22063) | > loss_disc: 2.46008 (2.46008) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.46008 (2.46008) | > grad_norm_1: 12.51229 (12.51229) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.13450 (10.13451) | > loader_time: 0.66280 (0.66277)  --> STEP: 25/120 -- GLOBAL_STEP: 3050 | > loss_gen: 1.93305 (2.00610) | > loss_kl: 2.23381 (2.30764) | > loss_feat: 2.50545 (2.74807) | > loss_mel: 22.67993 (22.96561) | > loss_duration: 0.97525 (1.02947) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.32748 (31.05689) | > grad_norm_0: 39.94299 (80.49714) | > loss_disc: 2.64070 (2.59216) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64070 (2.59216) | > grad_norm_1: 11.35603 (16.99766) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34540 (9.95995) | > loader_time: 0.01110 (0.00917)  --> STEP: 50/120 -- GLOBAL_STEP: 3075 | > loss_gen: 1.96298 (1.98320) | > loss_kl: 2.20457 (2.22755) | > loss_feat: 2.20619 (2.64603) | > loss_mel: 21.12311 (22.73360) | > loss_duration: 0.94551 (0.97230) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.44235 (30.56268) | > grad_norm_0: 161.22580 (107.65612) | > loss_disc: 2.63001 (2.61660) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.63001 (2.61660) | > grad_norm_1: 20.38883 (18.16842) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.91220 (10.31184) | > loader_time: 0.01340 (0.01075)  --> STEP: 75/120 -- GLOBAL_STEP: 3100 | > loss_gen: 2.08695 (1.98308) | > loss_kl: 2.12545 (2.17728) | > loss_feat: 2.50560 (2.61428) | > loss_mel: 21.97912 (22.58253) | > loss_duration: 0.94590 (0.96154) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.64302 (30.31871) | > grad_norm_0: 44.31750 (105.88803) | > loss_disc: 2.64269 (2.62030) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64269 (2.62030) | > grad_norm_1: 30.47633 (20.18381) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93220 (10.52595) | > loader_time: 0.01530 (0.01325)  --> STEP: 100/120 -- GLOBAL_STEP: 3125 | > loss_gen: 2.59125 (1.97899) | > loss_kl: 1.91052 (2.15479) | > loss_feat: 2.34524 (2.58911) | > loss_mel: 22.32879 (22.47717) | > loss_duration: 0.96139 (0.95709) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.13719 (30.15715) | > grad_norm_0: 152.91248 (107.18635) | > loss_disc: 2.85260 (2.62485) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.85260 (2.62485) | > grad_norm_1: 92.33680 (21.51366) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.92920 (10.81305) | > loader_time: 0.01920 (0.01430) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.10335 (2.10335) | > loss_kl: 2.51368 (2.51368) | > loss_feat: 2.72272 (2.72272) | > loss_mel: 22.59284 (22.59284) | > loss_duration: 0.87842 (0.87842) | > loss_0: 30.81100 (30.81100) | > loss_disc: 2.65030 (2.65030) | > loss_1: 2.65030 (2.65030)  --> STEP: 1 | > loss_gen: 2.09741 (2.09741) | > loss_kl: 2.40314 (2.40314) | > loss_feat: 2.40615 (2.40615) | > loss_mel: 22.53186 (22.53186) | > loss_duration: 0.87801 (0.87801) | > loss_0: 30.31657 (30.31657) | > loss_disc: 2.70102 (2.70102) | > loss_1: 2.70102 (2.70102)  --> STEP: 2 | > loss_gen: 2.13829 (2.11785) | > loss_kl: 1.85177 (2.12746) | > loss_feat: 2.56177 (2.48396) | > loss_mel: 22.73317 (22.63252) | > loss_duration: 0.90367 (0.89084) | > loss_0: 30.18868 (30.25263) | > loss_disc: 2.61973 (2.66037) | > loss_1: 2.61973 (2.66037)  --> STEP: 3 | > loss_gen: 2.01064 (2.08211) | > loss_kl: 1.97188 (2.07560) | > loss_feat: 2.85001 (2.60598) | > loss_mel: 22.06631 (22.44378) | > loss_duration: 0.95885 (0.91351) | > loss_0: 29.85768 (30.12098) | > loss_disc: 2.59837 (2.63971) | > loss_1: 2.59837 (2.63971) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00750 (+0.00020) | > avg_loss_gen: 2.08211 (+0.12482) | > avg_loss_kl: 2.07560 (+0.06450) | > avg_loss_feat: 2.60598 (+0.04752) | > avg_loss_mel: 22.44378 (+0.26372) | > avg_loss_duration: 0.91351 (-0.00300) | > avg_loss_0: 30.12098 (+0.49755) | > avg_loss_disc: 2.63971 (+0.05600) | > avg_loss_1: 2.63971 (+0.05600)  > EPOCH: 26/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-25 23:38:04)   --> STEP: 4/120 -- GLOBAL_STEP: 3150 | > loss_gen: 1.90928 (2.03872) | > loss_kl: 2.69447 (2.52310) | > loss_feat: 2.57938 (2.90441) | > loss_mel: 23.49303 (23.53625) | > loss_duration: 0.91166 (0.92229) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.58781 (31.92478) | > grad_norm_0: 146.77560 (145.23610) | > loss_disc: 2.63537 (2.60144) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.63537 (2.60144) | > grad_norm_1: 18.38223 (13.88734) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.05470 (9.90467) | > loader_time: 0.00730 (0.00714)  --> STEP: 29/120 -- GLOBAL_STEP: 3175 | > loss_gen: 1.92449 (1.98019) | > loss_kl: 2.27880 (2.32128) | > loss_feat: 2.39214 (2.67696) | > loss_mel: 22.50918 (23.02608) | > loss_duration: 0.90984 (1.01142) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.01445 (31.01592) | > grad_norm_0: 39.79612 (102.41511) | > loss_disc: 2.66782 (2.64387) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.66782 (2.64387) | > grad_norm_1: 24.20507 (18.77365) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.70970 (10.37630) | > loader_time: 0.01110 (0.00946)  --> STEP: 54/120 -- GLOBAL_STEP: 3200 | > loss_gen: 1.86318 (1.97379) | > loss_kl: 2.20967 (2.24418) | > loss_feat: 2.76782 (2.62750) | > loss_mel: 21.60345 (22.79577) | > loss_duration: 0.92958 (0.96838) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.37369 (30.60962) | > grad_norm_0: 184.02969 (109.31689) | > loss_disc: 2.48848 (2.61827) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.48848 (2.61827) | > grad_norm_1: 20.68377 (19.78138) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.92640 (10.55146) | > loader_time: 0.01350 (0.01096)  --> STEP: 79/120 -- GLOBAL_STEP: 3225 | > loss_gen: 2.29687 (1.98579) | > loss_kl: 2.25330 (2.20504) | > loss_feat: 2.51871 (2.64457) | > loss_mel: 21.42514 (22.54425) | > loss_duration: 0.91397 (0.96081) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.40799 (30.34046) | > grad_norm_0: 62.49751 (103.23522) | > loss_disc: 2.69443 (2.60830) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.69443 (2.60830) | > grad_norm_1: 61.61788 (22.63335) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.47690 (10.69656) | > loader_time: 0.01620 (0.01332)  --> STEP: 104/120 -- GLOBAL_STEP: 3250 | > loss_gen: 1.96737 (1.98368) | > loss_kl: 2.04965 (2.18211) | > loss_feat: 2.57325 (2.64525) | > loss_mel: 21.88130 (22.38816) | > loss_duration: 0.95496 (0.95588) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.42653 (30.15508) | > grad_norm_0: 182.19740 (107.47253) | > loss_disc: 2.54596 (2.60715) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54596 (2.60715) | > grad_norm_1: 5.54028 (23.62615) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.08530 (10.97039) | > loader_time: 0.01950 (0.01444) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.96431 (1.96431) | > loss_kl: 2.10818 (2.10818) | > loss_feat: 2.44092 (2.44092) | > loss_mel: 21.64871 (21.64871) | > loss_duration: 0.86605 (0.86605) | > loss_0: 29.02817 (29.02817) | > loss_disc: 2.62066 (2.62066) | > loss_1: 2.62066 (2.62066)  --> STEP: 1 | > loss_gen: 1.95831 (1.95831) | > loss_kl: 2.64883 (2.64883) | > loss_feat: 2.89798 (2.89798) | > loss_mel: 23.25413 (23.25413) | > loss_duration: 0.87090 (0.87090) | > loss_0: 31.63016 (31.63016) | > loss_disc: 2.47544 (2.47544) | > loss_1: 2.47544 (2.47544)  --> STEP: 2 | > loss_gen: 2.09534 (2.02683) | > loss_kl: 1.72487 (2.18685) | > loss_feat: 2.27077 (2.58437) | > loss_mel: 21.86158 (22.55785) | > loss_duration: 0.90429 (0.88760) | > loss_0: 28.85685 (30.24350) | > loss_disc: 2.52586 (2.50065) | > loss_1: 2.52586 (2.50065)  --> STEP: 3 | > loss_gen: 1.99196 (2.01520) | > loss_kl: 2.14382 (2.17251) | > loss_feat: 2.12341 (2.43072) | > loss_mel: 19.72013 (21.61195) | > loss_duration: 0.95068 (0.90862) | > loss_0: 26.93000 (29.13900) | > loss_disc: 2.65151 (2.55094) | > loss_1: 2.65151 (2.55094) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00755 (+0.00005) | > avg_loss_gen: 2.01520 (-0.06691) | > avg_loss_kl: 2.17251 (+0.09691) | > avg_loss_feat: 2.43072 (-0.17526) | > avg_loss_mel: 21.61195 (-0.83183) | > avg_loss_duration: 0.90862 (-0.00489) | > avg_loss_0: 29.13900 (-0.98198) | > avg_loss_disc: 2.55094 (-0.08877) | > avg_loss_1: 2.55094 (-0.08877) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_3267.pth.tar  > EPOCH: 27/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 00:00:57)   --> STEP: 8/120 -- GLOBAL_STEP: 3275 | > loss_gen: 2.19781 (2.11868) | > loss_kl: 2.48910 (2.35472) | > loss_feat: 2.94050 (3.06624) | > loss_mel: 22.73483 (23.42357) | > loss_duration: 0.97063 (0.91980) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.33287 (31.88301) | > grad_norm_0: 135.53157 (151.71065) | > loss_disc: 2.55270 (2.54943) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.55270 (2.54943) | > grad_norm_1: 14.47751 (19.56244) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.81670 (9.20238) | > loader_time: 0.00830 (0.00765)  --> STEP: 33/120 -- GLOBAL_STEP: 3300 | > loss_gen: 1.91720 (2.02835) | > loss_kl: 1.99762 (2.31050) | > loss_feat: 2.61691 (2.79901) | > loss_mel: 21.94576 (22.54910) | > loss_duration: 0.94237 (0.99577) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.41986 (30.68272) | > grad_norm_0: 183.90556 (115.96446) | > loss_disc: 2.56134 (2.59595) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.56134 (2.59595) | > grad_norm_1: 9.23303 (25.63844) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.78910 (10.17530) | > loader_time: 0.01180 (0.00977)  --> STEP: 58/120 -- GLOBAL_STEP: 3325 | > loss_gen: 2.14705 (2.01731) | > loss_kl: 2.18442 (2.22538) | > loss_feat: 2.68397 (2.75649) | > loss_mel: 22.19375 (22.40125) | > loss_duration: 0.91915 (0.96377) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.12834 (30.36421) | > grad_norm_0: 121.55753 (113.58915) | > loss_disc: 2.54708 (2.58305) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54708 (2.58305) | > grad_norm_1: 28.69600 (24.17995) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.26060 (10.44192) | > loader_time: 0.03720 (0.01277)  --> STEP: 83/120 -- GLOBAL_STEP: 3350 | > loss_gen: 2.42204 (2.01841) | > loss_kl: 1.97908 (2.18319) | > loss_feat: 2.82837 (2.74420) | > loss_mel: 21.81708 (22.26267) | > loss_duration: 0.93211 (0.95674) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.97869 (30.16520) | > grad_norm_0: 152.67329 (111.64124) | > loss_disc: 2.57691 (2.58230) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.57691 (2.58230) | > grad_norm_1: 22.96962 (24.89342) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.40180 (10.63452) | > loader_time: 0.01660 (0.01378)  --> STEP: 108/120 -- GLOBAL_STEP: 3375 | > loss_gen: 1.81261 (2.00524) | > loss_kl: 2.09192 (2.16220) | > loss_feat: 2.44355 (2.70911) | > loss_mel: 22.16196 (22.17921) | > loss_duration: 0.98774 (0.95455) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.49778 (30.01030) | > grad_norm_0: 106.56955 (115.19798) | > loss_disc: 2.71081 (2.59896) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.71081 (2.59896) | > grad_norm_1: 10.96913 (26.50097) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.01490 (10.90781) | > loader_time: 0.01990 (0.01486) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.08915 (2.08915) | > loss_kl: 2.21403 (2.21403) | > loss_feat: 2.65432 (2.65432) | > loss_mel: 22.57592 (22.57592) | > loss_duration: 0.88241 (0.88241) | > loss_0: 30.41582 (30.41582) | > loss_disc: 2.52954 (2.52954) | > loss_1: 2.52954 (2.52954)  --> STEP: 1 | > loss_gen: 2.01309 (2.01309) | > loss_kl: 2.23032 (2.23032) | > loss_feat: 2.65555 (2.65555) | > loss_mel: 22.59864 (22.59864) | > loss_duration: 0.88650 (0.88650) | > loss_0: 30.38409 (30.38409) | > loss_disc: 2.56271 (2.56271) | > loss_1: 2.56271 (2.56271)  --> STEP: 2 | > loss_gen: 2.00207 (2.00758) | > loss_kl: 2.17597 (2.20314) | > loss_feat: 2.58391 (2.61973) | > loss_mel: 21.57744 (22.08804) | > loss_duration: 0.90261 (0.89455) | > loss_0: 29.24199 (29.81304) | > loss_disc: 2.64172 (2.60221) | > loss_1: 2.64172 (2.60221)  --> STEP: 3 | > loss_gen: 1.95099 (1.98871) | > loss_kl: 2.46492 (2.29040) | > loss_feat: 2.28738 (2.50895) | > loss_mel: 21.85523 (22.01044) | > loss_duration: 0.95197 (0.91369) | > loss_0: 29.51049 (29.71219) | > loss_disc: 2.74280 (2.64908) | > loss_1: 2.74280 (2.64908) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00709 (-0.00046) | > avg_loss_gen: 1.98871 (-0.02649) | > avg_loss_kl: 2.29040 (+0.11789) | > avg_loss_feat: 2.50895 (+0.07823) | > avg_loss_mel: 22.01044 (+0.39849) | > avg_loss_duration: 0.91369 (+0.00507) | > avg_loss_0: 29.71219 (+0.57319) | > avg_loss_disc: 2.64908 (+0.09814) | > avg_loss_1: 2.64908 (+0.09814)  > EPOCH: 28/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 00:23:19)   --> STEP: 12/120 -- GLOBAL_STEP: 3400 | > loss_gen: 1.91154 (2.05216) | > loss_kl: 2.39434 (2.42413) | > loss_feat: 3.05899 (2.91405) | > loss_mel: 22.49871 (22.91421) | > loss_duration: 0.99196 (1.13818) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.85554 (31.44273) | > grad_norm_0: 117.16492 (99.85922) | > loss_disc: 2.58863 (2.62144) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58863 (2.62144) | > grad_norm_1: 38.37440 (42.55246) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.07470 (10.05616) | > loader_time: 0.00930 (0.00819)  --> STEP: 37/120 -- GLOBAL_STEP: 3425 | > loss_gen: 1.88674 (2.00461) | > loss_kl: 2.11780 (2.26799) | > loss_feat: 2.36538 (2.73516) | > loss_mel: 22.04415 (22.53803) | > loss_duration: 0.94058 (0.98763) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.35464 (30.53342) | > grad_norm_0: 51.08620 (110.10889) | > loss_disc: 2.61359 (2.61340) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.61359 (2.61340) | > grad_norm_1: 11.01593 (28.44666) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.66250 (10.28815) | > loader_time: 0.01260 (0.01016)  --> STEP: 62/120 -- GLOBAL_STEP: 3450 | > loss_gen: 2.14521 (1.99711) | > loss_kl: 1.92870 (2.20477) | > loss_feat: 2.57938 (2.70298) | > loss_mel: 22.13530 (22.30755) | > loss_duration: 0.94976 (0.96022) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.73835 (30.17262) | > grad_norm_0: 180.18724 (109.69316) | > loss_disc: 2.65055 (2.59902) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.65055 (2.59902) | > grad_norm_1: 18.00291 (25.23169) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.67110 (10.45731) | > loader_time: 0.01480 (0.01257)  --> STEP: 87/120 -- GLOBAL_STEP: 3475 | > loss_gen: 1.76935 (1.99417) | > loss_kl: 2.18723 (2.17957) | > loss_feat: 2.79790 (2.68805) | > loss_mel: 21.91294 (22.15782) | > loss_duration: 0.95516 (0.95448) | > amp_scaler: 512.00000 (959.26437) | > loss_0: 29.62259 (29.97409) | > grad_norm_0: 86.20770 (104.66888) | > loss_disc: 2.60375 (2.60965) | > amp_scaler-1: 512.00000 (953.37931) | > loss_1: 2.60375 (2.60965) | > grad_norm_1: 31.00877 (27.73828) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.49760 (10.66595) | > loader_time: 0.01710 (0.01352)  --> STEP: 112/120 -- GLOBAL_STEP: 3500 | > loss_gen: 1.96887 (1.99092) | > loss_kl: 2.00239 (2.15513) | > loss_feat: 2.46668 (2.66964) | > loss_mel: 21.25905 (22.05758) | > loss_duration: 0.97438 (0.95412) | > amp_scaler: 512.00000 (859.42857) | > loss_0: 28.67137 (29.82738) | > grad_norm_0: 39.24903 (109.27670) | > loss_disc: 2.67567 (2.61151) | > amp_scaler-1: 512.00000 (854.85714) | > loss_1: 2.67567 (2.61151) | > grad_norm_1: 12.30999 (25.51719) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.04210 (10.94263) | > loader_time: 0.02160 (0.01484) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.87579 (1.87579) | > loss_kl: 2.25574 (2.25574) | > loss_feat: 2.72023 (2.72023) | > loss_mel: 22.46352 (22.46352) | > loss_duration: 0.86784 (0.86784) | > loss_0: 30.18312 (30.18312) | > loss_disc: 2.55584 (2.55584) | > loss_1: 2.55584 (2.55584)  --> STEP: 1 | > loss_gen: 1.85486 (1.85486) | > loss_kl: 2.25010 (2.25010) | > loss_feat: 2.90910 (2.90910) | > loss_mel: 24.14951 (24.14951) | > loss_duration: 0.87333 (0.87333) | > loss_0: 32.03690 (32.03690) | > loss_disc: 2.64653 (2.64653) | > loss_1: 2.64653 (2.64653)  --> STEP: 2 | > loss_gen: 1.84684 (1.85085) | > loss_kl: 2.04759 (2.14884) | > loss_feat: 2.51306 (2.71108) | > loss_mel: 21.53978 (22.84464) | > loss_duration: 0.89908 (0.88621) | > loss_0: 28.84636 (30.44163) | > loss_disc: 2.69180 (2.66917) | > loss_1: 2.69180 (2.66917)  --> STEP: 3 | > loss_gen: 1.85296 (1.85156) | > loss_kl: 2.30981 (2.20250) | > loss_feat: 2.62651 (2.68289) | > loss_mel: 20.96625 (22.21851) | > loss_duration: 0.95241 (0.90828) | > loss_0: 28.70794 (29.86373) | > loss_disc: 2.67918 (2.67251) | > loss_1: 2.67918 (2.67251) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00716 (+0.00008) | > avg_loss_gen: 1.85156 (-0.13716) | > avg_loss_kl: 2.20250 (-0.08790) | > avg_loss_feat: 2.68289 (+0.17394) | > avg_loss_mel: 22.21851 (+0.20807) | > avg_loss_duration: 0.90828 (-0.00542) | > avg_loss_0: 29.86373 (+0.15154) | > avg_loss_disc: 2.67251 (+0.02343) | > avg_loss_1: 2.67251 (+0.02343)  > EPOCH: 29/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 00:45:41)   --> STEP: 16/120 -- GLOBAL_STEP: 3525 | > loss_gen: 1.77320 (2.05532) | > loss_kl: 2.26992 (2.35636) | > loss_feat: 2.67630 (2.89312) | > loss_mel: 22.74987 (22.75477) | > loss_duration: 0.90637 (1.08361) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.37567 (31.14317) | > grad_norm_0: 97.35812 (124.63197) | > loss_disc: 2.54911 (2.57114) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54911 (2.57114) | > grad_norm_1: 36.16368 (21.79252) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.40140 (10.17092) | > loader_time: 0.01010 (0.00843)  --> STEP: 41/120 -- GLOBAL_STEP: 3550 | > loss_gen: 1.91520 (1.99713) | > loss_kl: 2.07103 (2.28126) | > loss_feat: 2.54506 (2.70849) | > loss_mel: 21.50701 (22.28740) | > loss_duration: 0.90894 (0.98208) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.94725 (30.25637) | > grad_norm_0: 195.66043 (101.58131) | > loss_disc: 2.59184 (2.60669) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59184 (2.60669) | > grad_norm_1: 9.31598 (24.64473) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.89900 (10.39720) | > loader_time: 0.01290 (0.01029)  --> STEP: 66/120 -- GLOBAL_STEP: 3575 | > loss_gen: 1.90623 (1.99464) | > loss_kl: 2.14306 (2.23138) | > loss_feat: 2.88732 (2.70041) | > loss_mel: 22.11435 (22.10148) | > loss_duration: 0.94981 (0.96012) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.00077 (29.98802) | > grad_norm_0: 186.44086 (113.18224) | > loss_disc: 2.54067 (2.60509) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54067 (2.60509) | > grad_norm_1: 9.40803 (26.58343) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.81130 (10.55768) | > loader_time: 0.01540 (0.01231)  --> STEP: 91/120 -- GLOBAL_STEP: 3600 | > loss_gen: 1.95091 (2.00080) | > loss_kl: 2.06613 (2.19749) | > loss_feat: 2.68318 (2.71790) | > loss_mel: 21.17777 (21.97229) | > loss_duration: 0.95430 (0.95379) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.83229 (29.84226) | > grad_norm_0: 37.21359 (117.70719) | > loss_disc: 2.55455 (2.59076) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55455 (2.59076) | > grad_norm_1: 32.46734 (25.77485) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.74760 (10.75323) | > loader_time: 0.02010 (0.01354)  --> STEP: 116/120 -- GLOBAL_STEP: 3625 | > loss_gen: 1.95912 (2.00094) | > loss_kl: 2.02210 (2.16999) | > loss_feat: 2.57899 (2.71700) | > loss_mel: 22.50714 (21.88957) | > loss_duration: 1.04307 (0.95598) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.11042 (29.73348) | > grad_norm_0: 75.91589 (113.58191) | > loss_disc: 2.54313 (2.59207) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54313 (2.59207) | > grad_norm_1: 7.65579 (26.56100) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.37490 (11.04359) | > loader_time: 0.02540 (0.01498) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.07756 (2.07756) | > loss_kl: 2.20461 (2.20461) | > loss_feat: 2.46286 (2.46286) | > loss_mel: 21.84925 (21.84925) | > loss_duration: 0.88686 (0.88686) | > loss_0: 29.48113 (29.48113) | > loss_disc: 2.56891 (2.56891) | > loss_1: 2.56891 (2.56891)  --> STEP: 1 | > loss_gen: 1.96558 (1.96558) | > loss_kl: 2.01371 (2.01371) | > loss_feat: 2.71374 (2.71374) | > loss_mel: 22.90121 (22.90121) | > loss_duration: 0.88884 (0.88884) | > loss_0: 30.48307 (30.48307) | > loss_disc: 2.54720 (2.54720) | > loss_1: 2.54720 (2.54720)  --> STEP: 2 | > loss_gen: 1.97149 (1.96853) | > loss_kl: 1.93146 (1.97258) | > loss_feat: 2.72404 (2.71889) | > loss_mel: 21.46461 (22.18291) | > loss_duration: 0.91673 (0.90278) | > loss_0: 29.00832 (29.74569) | > loss_disc: 2.54169 (2.54445) | > loss_1: 2.54169 (2.54445)  --> STEP: 3 | > loss_gen: 2.13386 (2.02364) | > loss_kl: 2.34543 (2.09686) | > loss_feat: 3.37937 (2.93905) | > loss_mel: 23.68835 (22.68472) | > loss_duration: 0.97449 (0.92668) | > loss_0: 32.52149 (30.67096) | > loss_disc: 2.34607 (2.47832) | > loss_1: 2.34607 (2.47832) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00721 (+0.00005) | > avg_loss_gen: 2.02364 (+0.17209) | > avg_loss_kl: 2.09686 (-0.10564) | > avg_loss_feat: 2.93905 (+0.25616) | > avg_loss_mel: 22.68472 (+0.46621) | > avg_loss_duration: 0.92668 (+0.01841) | > avg_loss_0: 30.67096 (+0.80723) | > avg_loss_disc: 2.47832 (-0.19418) | > avg_loss_1: 2.47832 (-0.19418)  > EPOCH: 30/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 01:08:11)   --> STEP: 20/120 -- GLOBAL_STEP: 3650 | > loss_gen: 2.06866 (2.06554) | > loss_kl: 2.17492 (2.36920) | > loss_feat: 2.97503 (2.99549) | > loss_mel: 21.76583 (22.44086) | > loss_duration: 0.93415 (1.04206) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.91859 (30.91316) | > grad_norm_0: 63.67721 (98.90393) | > loss_disc: 2.57061 (2.55445) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57061 (2.55445) | > grad_norm_1: 46.14771 (20.10648) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.71810 (10.28608) | > loader_time: 0.01040 (0.00836)  --> STEP: 45/120 -- GLOBAL_STEP: 3675 | > loss_gen: 1.71856 (2.00951) | > loss_kl: 2.02721 (2.24011) | > loss_feat: 2.43537 (2.79974) | > loss_mel: 21.94695 (22.12260) | > loss_duration: 0.91153 (0.97091) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.03963 (30.14287) | > grad_norm_0: 126.16852 (101.45955) | > loss_disc: 2.87033 (2.61490) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.87033 (2.61490) | > grad_norm_1: 62.77252 (24.73312) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.79270 (10.47495) | > loader_time: 0.01340 (0.01039)  --> STEP: 70/120 -- GLOBAL_STEP: 3700 | > loss_gen: 2.03009 (1.99463) | > loss_kl: 2.05219 (2.21285) | > loss_feat: 2.89381 (2.73473) | > loss_mel: 21.51394 (21.98171) | > loss_duration: 0.97980 (0.95567) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.46982 (29.87958) | > grad_norm_0: 117.82404 (100.81229) | > loss_disc: 2.63327 (2.62242) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63327 (2.62242) | > grad_norm_1: 8.30476 (24.60615) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.89400 (10.62499) | > loader_time: 0.01550 (0.01274)  --> STEP: 95/120 -- GLOBAL_STEP: 3725 | > loss_gen: 2.05194 (1.98993) | > loss_kl: 2.09452 (2.18408) | > loss_feat: 2.82102 (2.70065) | > loss_mel: 21.98601 (21.84624) | > loss_duration: 0.95158 (0.95145) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.90507 (29.67235) | > grad_norm_0: 181.25702 (102.49059) | > loss_disc: 2.50223 (2.62236) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50223 (2.62236) | > grad_norm_1: 16.65539 (25.18871) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.01480 (10.83944) | > loader_time: 0.01840 (0.01385)  --> STEP: 120/120 -- GLOBAL_STEP: 3750 | > loss_gen: 1.80883 (1.98699) | > loss_kl: 1.93479 (2.15863) | > loss_feat: 2.65048 (2.68201) | > loss_mel: 21.33092 (21.80352) | > loss_duration: 1.16302 (0.95888) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.88804 (29.59003) | > grad_norm_0: 87.77691 (106.62135) | > loss_disc: 2.56804 (2.62529) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56804 (2.62529) | > grad_norm_1: 32.11041 (25.56760) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.34210 (11.11068) | > loader_time: 0.02460 (0.01540) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.06153 (2.06153) | > loss_kl: 2.27473 (2.27473) | > loss_feat: 2.53538 (2.53538) | > loss_mel: 22.20810 (22.20810) | > loss_duration: 0.87054 (0.87054) | > loss_0: 29.95028 (29.95028) | > loss_disc: 2.59909 (2.59909) | > loss_1: 2.59909 (2.59909)  --> STEP: 1 | > loss_gen: 1.95360 (1.95360) | > loss_kl: 2.22898 (2.22898) | > loss_feat: 2.45204 (2.45204) | > loss_mel: 21.78622 (21.78622) | > loss_duration: 0.88605 (0.88605) | > loss_0: 29.30690 (29.30690) | > loss_disc: 2.65180 (2.65180) | > loss_1: 2.65180 (2.65180)  --> STEP: 2 | > loss_gen: 1.93188 (1.94274) | > loss_kl: 2.17990 (2.20444) | > loss_feat: 2.68990 (2.57097) | > loss_mel: 22.25370 (22.01996) | > loss_duration: 0.90541 (0.89573) | > loss_0: 29.96080 (29.63385) | > loss_disc: 2.63359 (2.64270) | > loss_1: 2.63359 (2.64270)  --> STEP: 3 | > loss_gen: 2.11773 (2.00107) | > loss_kl: 2.34549 (2.25146) | > loss_feat: 2.68025 (2.60740) | > loss_mel: 20.97304 (21.67099) | > loss_duration: 0.95975 (0.91707) | > loss_0: 29.07627 (29.44799) | > loss_disc: 2.57657 (2.62065) | > loss_1: 2.57657 (2.62065) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00718 (-0.00003) | > avg_loss_gen: 2.00107 (-0.02257) | > avg_loss_kl: 2.25146 (+0.15459) | > avg_loss_feat: 2.60740 (-0.33165) | > avg_loss_mel: 21.67099 (-1.01373) | > avg_loss_duration: 0.91707 (-0.00961) | > avg_loss_0: 29.44799 (-1.22297) | > avg_loss_disc: 2.62065 (+0.14233) | > avg_loss_1: 2.62065 (+0.14233)  > EPOCH: 31/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 01:30:41)   --> STEP: 24/120 -- GLOBAL_STEP: 3775 | > loss_gen: 2.02000 (2.01778) | > loss_kl: 2.13390 (2.33188) | > loss_feat: 3.07126 (2.87292) | > loss_mel: 21.36464 (22.13988) | > loss_duration: 0.87371 (1.02420) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.46351 (30.38667) | > grad_norm_0: 128.38573 (95.22142) | > loss_disc: 2.48281 (2.56180) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48281 (2.56180) | > grad_norm_1: 12.57643 (21.00452) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.49500 (10.33500) | > loader_time: 0.01060 (0.00909)  --> STEP: 49/120 -- GLOBAL_STEP: 3800 | > loss_gen: 1.72978 (1.99299) | > loss_kl: 2.23935 (2.27009) | > loss_feat: 2.60824 (2.75235) | > loss_mel: 21.19009 (21.86419) | > loss_duration: 0.91051 (0.96556) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.67798 (29.84518) | > grad_norm_0: 45.71208 (100.71974) | > loss_disc: 2.83232 (2.59696) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.83232 (2.59696) | > grad_norm_1: 16.31346 (21.51818) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.94440 (10.51560) | > loader_time: 0.01320 (0.01075)  --> STEP: 74/120 -- GLOBAL_STEP: 3825 | > loss_gen: 1.76075 (1.98897) | > loss_kl: 2.04148 (2.22846) | > loss_feat: 2.68423 (2.71246) | > loss_mel: 21.57390 (21.75758) | > loss_duration: 0.91118 (0.95511) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.97154 (29.64258) | > grad_norm_0: 43.51136 (97.99520) | > loss_disc: 2.69908 (2.60943) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.69908 (2.60943) | > grad_norm_1: 85.76363 (22.53372) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.95300 (10.64729) | > loader_time: 0.01580 (0.01325)  --> STEP: 99/120 -- GLOBAL_STEP: 3850 | > loss_gen: 1.97922 (1.98262) | > loss_kl: 2.14326 (2.20820) | > loss_feat: 2.87505 (2.69118) | > loss_mel: 21.11580 (21.64970) | > loss_duration: 0.91632 (0.95090) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.02966 (29.48261) | > grad_norm_0: 51.80465 (96.56002) | > loss_disc: 2.57632 (2.61319) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57632 (2.61319) | > grad_norm_1: 22.21549 (23.40926) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.73390 (10.85870) | > loader_time: 0.01880 (0.01445) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.07090 (2.07090) | > loss_kl: 2.14797 (2.14797) | > loss_feat: 2.69377 (2.69377) | > loss_mel: 21.67334 (21.67334) | > loss_duration: 0.87326 (0.87326) | > loss_0: 29.45923 (29.45923) | > loss_disc: 2.60525 (2.60525) | > loss_1: 2.60525 (2.60525)  --> STEP: 1 | > loss_gen: 2.03861 (2.03861) | > loss_kl: 2.22646 (2.22646) | > loss_feat: 2.55461 (2.55461) | > loss_mel: 21.85898 (21.85898) | > loss_duration: 0.89173 (0.89173) | > loss_0: 29.57040 (29.57040) | > loss_disc: 2.73401 (2.73401) | > loss_1: 2.73401 (2.73401)  --> STEP: 2 | > loss_gen: 1.92482 (1.98171) | > loss_kl: 2.10832 (2.16739) | > loss_feat: 2.55347 (2.55404) | > loss_mel: 21.08400 (21.47149) | > loss_duration: 0.90440 (0.89807) | > loss_0: 28.57501 (29.07270) | > loss_disc: 2.65721 (2.69561) | > loss_1: 2.65721 (2.69561)  --> STEP: 3 | > loss_gen: 1.95468 (1.97270) | > loss_kl: 2.38548 (2.24009) | > loss_feat: 2.88577 (2.66462) | > loss_mel: 20.97976 (21.30758) | > loss_duration: 0.95774 (0.91796) | > loss_0: 29.16343 (29.10295) | > loss_disc: 2.70231 (2.69785) | > loss_1: 2.70231 (2.69785) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00759 (+0.00041) | > avg_loss_gen: 1.97270 (-0.02837) | > avg_loss_kl: 2.24009 (-0.01137) | > avg_loss_feat: 2.66462 (+0.05722) | > avg_loss_mel: 21.30758 (-0.36341) | > avg_loss_duration: 0.91796 (+0.00089) | > avg_loss_0: 29.10295 (-0.34504) | > avg_loss_disc: 2.69785 (+0.07719) | > avg_loss_1: 2.69785 (+0.07719)  > EPOCH: 32/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 01:53:08)   --> STEP: 3/120 -- GLOBAL_STEP: 3875 | > loss_gen: 2.06021 (2.02578) | > loss_kl: 2.47473 (2.47317) | > loss_feat: 2.86754 (2.87393) | > loss_mel: 22.70070 (22.98901) | > loss_duration: 0.93363 (0.89651) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.03680 (31.25841) | > grad_norm_0: 188.71465 (182.90790) | > loss_disc: 2.73345 (2.75102) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.73345 (2.75102) | > grad_norm_1: 34.47157 (40.06476) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.88370 (9.69453) | > loader_time: 0.00750 (0.00668)  --> STEP: 28/120 -- GLOBAL_STEP: 3900 | > loss_gen: 1.91371 (1.96531) | > loss_kl: 2.12050 (2.31613) | > loss_feat: 2.57840 (2.68741) | > loss_mel: 21.77761 (22.21150) | > loss_duration: 0.92665 (1.00260) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.31686 (30.18295) | > grad_norm_0: 94.46928 (140.59659) | > loss_disc: 2.70239 (2.66398) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.70239 (2.66398) | > grad_norm_1: 15.70415 (23.37495) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.75160 (10.23519) | > loader_time: 0.01130 (0.00947)  --> STEP: 53/120 -- GLOBAL_STEP: 3925 | > loss_gen: 1.85060 (1.96739) | > loss_kl: 2.11057 (2.24631) | > loss_feat: 2.66577 (2.65755) | > loss_mel: 21.61623 (21.83645) | > loss_duration: 0.95376 (0.96101) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.19693 (29.66871) | > grad_norm_0: 41.05355 (116.30633) | > loss_disc: 2.54969 (2.63497) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54969 (2.63497) | > grad_norm_1: 13.85962 (23.59471) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.85940 (10.42656) | > loader_time: 0.01350 (0.01104)  --> STEP: 78/120 -- GLOBAL_STEP: 3950 | > loss_gen: 1.72214 (1.97306) | > loss_kl: 2.14669 (2.21374) | > loss_feat: 2.65911 (2.67207) | > loss_mel: 21.45486 (21.72648) | > loss_duration: 0.92911 (0.95278) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.91191 (29.53813) | > grad_norm_0: 112.55666 (111.05109) | > loss_disc: 2.61570 (2.62292) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61570 (2.62292) | > grad_norm_1: 25.06273 (22.16667) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.18810 (10.58780) | > loader_time: 0.01580 (0.01333)  --> STEP: 103/120 -- GLOBAL_STEP: 3975 | > loss_gen: 1.97034 (1.96748) | > loss_kl: 1.96503 (2.19053) | > loss_feat: 2.67997 (2.65696) | > loss_mel: 20.86658 (21.64318) | > loss_duration: 0.94890 (0.94946) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.43082 (29.40762) | > grad_norm_0: 101.40095 (115.38724) | > loss_disc: 2.58906 (2.62322) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58906 (2.62322) | > grad_norm_1: 9.36043 (21.61521) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.29800 (10.86207) | > loader_time: 0.01950 (0.01444) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.77821 (1.77821) | > loss_kl: 2.39209 (2.39209) | > loss_feat: 2.84728 (2.84728) | > loss_mel: 21.55539 (21.55539) | > loss_duration: 0.85556 (0.85556) | > loss_0: 29.42854 (29.42854) | > loss_disc: 2.46433 (2.46433) | > loss_1: 2.46433 (2.46433)  --> STEP: 1 | > loss_gen: 1.78525 (1.78525) | > loss_kl: 2.35132 (2.35132) | > loss_feat: 2.73463 (2.73463) | > loss_mel: 21.79124 (21.79124) | > loss_duration: 0.87432 (0.87432) | > loss_0: 29.53676 (29.53676) | > loss_disc: 2.49020 (2.49020) | > loss_1: 2.49020 (2.49020)  --> STEP: 2 | > loss_gen: 1.69235 (1.73880) | > loss_kl: 1.89816 (2.12474) | > loss_feat: 3.11416 (2.92440) | > loss_mel: 21.58265 (21.68695) | > loss_duration: 0.90272 (0.88852) | > loss_0: 29.19005 (29.36340) | > loss_disc: 2.42295 (2.45658) | > loss_1: 2.42295 (2.45658)  --> STEP: 3 | > loss_gen: 1.71715 (1.73158) | > loss_kl: 2.00330 (2.08426) | > loss_feat: 2.92397 (2.92426) | > loss_mel: 21.40838 (21.59409) | > loss_duration: 0.96414 (0.91373) | > loss_0: 29.01694 (29.24792) | > loss_disc: 2.52523 (2.47946) | > loss_1: 2.52523 (2.47946) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00734 (-0.00026) | > avg_loss_gen: 1.73158 (-0.24112) | > avg_loss_kl: 2.08426 (-0.15583) | > avg_loss_feat: 2.92426 (+0.25964) | > avg_loss_mel: 21.59409 (+0.28651) | > avg_loss_duration: 0.91373 (-0.00423) | > avg_loss_0: 29.24792 (+0.14497) | > avg_loss_disc: 2.47946 (-0.21839) | > avg_loss_1: 2.47946 (-0.21839)  > EPOCH: 33/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 02:15:33)   --> STEP: 7/120 -- GLOBAL_STEP: 4000 | > loss_gen: 2.10658 (2.08141) | > loss_kl: 2.38079 (2.36540) | > loss_feat: 3.10482 (3.05905) | > loss_mel: 22.61423 (22.33731) | > loss_duration: 0.95796 (0.91033) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.16439 (30.75350) | > grad_norm_0: 41.29216 (75.06958) | > loss_disc: 2.55590 (2.57018) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55590 (2.57018) | > grad_norm_1: 17.66285 (15.23385) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34570 (10.01974) | > loader_time: 0.00620 (0.00749)  --> STEP: 32/120 -- GLOBAL_STEP: 4025 | > loss_gen: 1.79593 (2.03033) | > loss_kl: 2.32222 (2.36185) | > loss_feat: 2.81131 (2.85704) | > loss_mel: 21.40349 (21.85960) | > loss_duration: 0.94239 (0.99065) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.27534 (30.09946) | > grad_norm_0: 74.82696 (79.38161) | > loss_disc: 2.53715 (2.59668) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53715 (2.59668) | > grad_norm_1: 8.77031 (25.53888) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.65910 (10.35343) | > loader_time: 0.01210 (0.00963)  --> STEP: 57/120 -- GLOBAL_STEP: 4050 | > loss_gen: 2.16509 (2.00326) | > loss_kl: 2.28502 (2.28236) | > loss_feat: 2.63869 (2.76401) | > loss_mel: 21.36808 (21.73582) | > loss_duration: 0.90466 (0.95681) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.36154 (29.74226) | > grad_norm_0: 130.97670 (83.75862) | > loss_disc: 2.59648 (2.61443) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59648 (2.61443) | > grad_norm_1: 35.90284 (25.01398) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.95210 (10.51957) | > loader_time: 0.01400 (0.01205)  --> STEP: 82/120 -- GLOBAL_STEP: 4075 | > loss_gen: 2.29338 (1.99432) | > loss_kl: 1.95341 (2.22666) | > loss_feat: 2.63384 (2.73043) | > loss_mel: 21.94455 (21.63038) | > loss_duration: 0.95963 (0.95107) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.78481 (29.53286) | > grad_norm_0: 105.35587 (89.37148) | > loss_disc: 2.63529 (2.61981) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63529 (2.61981) | > grad_norm_1: 21.26574 (25.01335) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93190 (10.67386) | > loader_time: 0.01690 (0.01307)  --> STEP: 107/120 -- GLOBAL_STEP: 4100 | > loss_gen: 1.56974 (1.98791) | > loss_kl: 2.25730 (2.20802) | > loss_feat: 2.85104 (2.72012) | > loss_mel: 21.03845 (21.51003) | > loss_duration: 0.96176 (0.94880) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.67830 (29.37489) | > grad_norm_0: 152.08794 (90.45998) | > loss_disc: 2.74372 (2.61257) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.74372 (2.61257) | > grad_norm_1: 93.27912 (24.45824) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.92340 (10.95714) | > loader_time: 0.02030 (0.01431) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.94783 (1.94783) | > loss_kl: 2.12522 (2.12522) | > loss_feat: 2.82200 (2.82200) | > loss_mel: 21.04194 (21.04194) | > loss_duration: 0.87423 (0.87423) | > loss_0: 28.81123 (28.81123) | > loss_disc: 2.54549 (2.54549) | > loss_1: 2.54549 (2.54549)  --> STEP: 1 | > loss_gen: 1.90574 (1.90574) | > loss_kl: 2.17765 (2.17765) | > loss_feat: 2.70252 (2.70252) | > loss_mel: 21.99121 (21.99121) | > loss_duration: 0.88210 (0.88210) | > loss_0: 29.65922 (29.65922) | > loss_disc: 2.63814 (2.63814) | > loss_1: 2.63814 (2.63814)  --> STEP: 2 | > loss_gen: 2.09173 (1.99873) | > loss_kl: 2.26991 (2.22378) | > loss_feat: 2.34867 (2.52559) | > loss_mel: 20.71191 (21.35156) | > loss_duration: 0.89994 (0.89102) | > loss_0: 28.32215 (28.99068) | > loss_disc: 2.67158 (2.65486) | > loss_1: 2.67158 (2.65486)  --> STEP: 3 | > loss_gen: 1.87407 (1.95718) | > loss_kl: 1.67568 (2.04108) | > loss_feat: 2.64466 (2.56528) | > loss_mel: 20.13849 (20.94720) | > loss_duration: 0.95931 (0.91378) | > loss_0: 27.29221 (28.42453) | > loss_disc: 2.63801 (2.64924) | > loss_1: 2.63801 (2.64924) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00729 (-0.00005) | > avg_loss_gen: 1.95718 (+0.22560) | > avg_loss_kl: 2.04108 (-0.04318) | > avg_loss_feat: 2.56528 (-0.35897) | > avg_loss_mel: 20.94720 (-0.64689) | > avg_loss_duration: 0.91378 (+0.00006) | > avg_loss_0: 28.42453 (-0.82339) | > avg_loss_disc: 2.64924 (+0.16978) | > avg_loss_1: 2.64924 (+0.16978) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_4114.pth.tar  > EPOCH: 34/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 02:38:21)   --> STEP: 11/120 -- GLOBAL_STEP: 4125 | > loss_gen: 2.11473 (2.07088) | > loss_kl: 2.16404 (2.37955) | > loss_feat: 2.90522 (2.97569) | > loss_mel: 21.43282 (22.07624) | > loss_duration: 0.91369 (0.91311) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.53049 (30.41547) | > grad_norm_0: 87.53516 (79.22961) | > loss_disc: 2.55288 (2.52846) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55288 (2.52846) | > grad_norm_1: 18.96548 (18.52687) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.12230 (9.36637) | > loader_time: 0.00920 (0.00780)  --> STEP: 36/120 -- GLOBAL_STEP: 4150 | > loss_gen: 2.00650 (2.02113) | > loss_kl: 2.12469 (2.31235) | > loss_feat: 2.72637 (2.89643) | > loss_mel: 21.30373 (21.78076) | > loss_duration: 0.95290 (0.98411) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.11419 (29.99478) | > grad_norm_0: 56.01918 (89.38639) | > loss_disc: 2.59779 (2.56417) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59779 (2.56417) | > grad_norm_1: 13.03418 (19.36339) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.71740 (10.15384) | > loader_time: 0.01240 (0.00992)  --> STEP: 61/120 -- GLOBAL_STEP: 4175 | > loss_gen: 1.92967 (2.00590) | > loss_kl: 2.00707 (2.24924) | > loss_feat: 2.73795 (2.79778) | > loss_mel: 21.37437 (21.58937) | > loss_duration: 0.96285 (0.95648) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.01191 (29.59877) | > grad_norm_0: 45.75160 (102.61932) | > loss_disc: 2.56668 (2.59441) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56668 (2.59441) | > grad_norm_1: 24.51849 (22.68747) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.52330 (10.40947) | > loader_time: 0.01460 (0.01218)  --> STEP: 86/120 -- GLOBAL_STEP: 4200 | > loss_gen: 1.87195 (1.99770) | > loss_kl: 2.12267 (2.21941) | > loss_feat: 2.76561 (2.77353) | > loss_mel: 21.35392 (21.45243) | > loss_duration: 0.94506 (0.94972) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.05920 (29.39278) | > grad_norm_0: 79.16200 (91.84236) | > loss_disc: 2.51365 (2.59201) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51365 (2.59201) | > grad_norm_1: 12.15878 (21.10740) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.59580 (10.62364) | > loader_time: 0.01710 (0.01318)  --> STEP: 111/120 -- GLOBAL_STEP: 4225 | > loss_gen: 1.89175 (1.99827) | > loss_kl: 2.20676 (2.18933) | > loss_feat: 2.69568 (2.77654) | > loss_mel: 21.78174 (21.39993) | > loss_duration: 0.94037 (0.94906) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.51630 (29.31313) | > grad_norm_0: 136.02924 (103.95626) | > loss_disc: 2.54068 (2.59116) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54068 (2.59116) | > grad_norm_1: 11.65331 (20.23597) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.96530 (10.92613) | > loader_time: 0.02200 (0.01451) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.78703 (1.78703) | > loss_kl: 2.25228 (2.25228) | > loss_feat: 2.77532 (2.77532) | > loss_mel: 21.84038 (21.84038) | > loss_duration: 0.86474 (0.86474) | > loss_0: 29.51974 (29.51974) | > loss_disc: 2.86060 (2.86060) | > loss_1: 2.86060 (2.86060)  --> STEP: 1 | > loss_gen: 1.70638 (1.70638) | > loss_kl: 2.37385 (2.37385) | > loss_feat: 2.76481 (2.76481) | > loss_mel: 21.63498 (21.63498) | > loss_duration: 0.87261 (0.87261) | > loss_0: 29.35262 (29.35262) | > loss_disc: 2.93411 (2.93411) | > loss_1: 2.93411 (2.93411)  --> STEP: 2 | > loss_gen: 1.70526 (1.70582) | > loss_kl: 1.94862 (2.16124) | > loss_feat: 2.52890 (2.64685) | > loss_mel: 21.20033 (21.41765) | > loss_duration: 0.90288 (0.88774) | > loss_0: 28.28599 (28.81931) | > loss_disc: 2.98079 (2.95745) | > loss_1: 2.98079 (2.95745)  --> STEP: 3 | > loss_gen: 1.66958 (1.69374) | > loss_kl: 2.13195 (2.15148) | > loss_feat: 2.64665 (2.64678) | > loss_mel: 19.81364 (20.88298) | > loss_duration: 0.96271 (0.91273) | > loss_0: 27.22452 (28.28771) | > loss_disc: 3.03123 (2.98204) | > loss_1: 3.03123 (2.98204) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00758 (+0.00030) | > avg_loss_gen: 1.69374 (-0.26344) | > avg_loss_kl: 2.15148 (+0.11039) | > avg_loss_feat: 2.64678 (+0.08150) | > avg_loss_mel: 20.88298 (-0.06422) | > avg_loss_duration: 0.91273 (-0.00105) | > avg_loss_0: 28.28771 (-0.13682) | > avg_loss_disc: 2.98204 (+0.33280) | > avg_loss_1: 2.98204 (+0.33280)  > EPOCH: 35/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 03:00:41)   --> STEP: 15/120 -- GLOBAL_STEP: 4250 | > loss_gen: 1.46644 (2.11233) | > loss_kl: 2.44284 (2.35921) | > loss_feat: 3.10387 (3.55411) | > loss_mel: 23.50249 (22.83669) | > loss_duration: 0.89465 (1.07499) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.41030 (31.93733) | > grad_norm_0: 27.91072 (117.91173) | > loss_disc: 4.50366 (3.30716) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 4.50366 (3.30716) | > grad_norm_1: 51.86836 (47.56118) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.29150 (10.22860) | > loader_time: 0.00980 (0.00815)  --> STEP: 40/120 -- GLOBAL_STEP: 4275 | > loss_gen: 2.57413 (2.25123) | > loss_kl: 2.19978 (2.26994) | > loss_feat: 5.06007 (3.99989) | > loss_mel: 23.39640 (22.91694) | > loss_duration: 0.91581 (0.97393) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 34.14619 (32.41193) | > grad_norm_0: 85.28009 (108.27154) | > loss_disc: 2.00863 (3.23029) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.00863 (3.23029) | > grad_norm_1: 14.03118 (36.64974) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.62340 (10.41339) | > loader_time: 0.01250 (0.01014)  --> STEP: 65/120 -- GLOBAL_STEP: 4300 | > loss_gen: 1.84703 (2.23896) | > loss_kl: 1.99519 (2.21620) | > loss_feat: 3.36549 (4.00579) | > loss_mel: 21.99473 (22.73130) | > loss_duration: 0.94115 (0.95301) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.14359 (32.14525) | > grad_norm_0: 57.21507 (103.33719) | > loss_disc: 2.32487 (3.12149) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.32487 (3.12149) | > grad_norm_1: 6.96942 (33.81175) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.56080 (10.53540) | > loader_time: 0.01480 (0.01273)  --> STEP: 90/120 -- GLOBAL_STEP: 4325 | > loss_gen: 1.69906 (2.15127) | > loss_kl: 2.16123 (2.19398) | > loss_feat: 3.53437 (3.75168) | > loss_mel: 21.66276 (22.51306) | > loss_duration: 0.90123 (0.94761) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.95865 (31.55760) | > grad_norm_0: 80.19377 (96.36784) | > loss_disc: 2.26630 (3.04558) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.26630 (3.04558) | > grad_norm_1: 37.02437 (29.45267) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.54440 (10.70381) | > loader_time: 0.02080 (0.01386)  --> STEP: 115/120 -- GLOBAL_STEP: 4350 | > loss_gen: 1.64215 (2.09675) | > loss_kl: 2.21204 (2.17438) | > loss_feat: 1.39423 (3.39601) | > loss_mel: 21.57599 (22.32142) | > loss_duration: 1.03973 (0.94981) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.86413 (30.93838) | > grad_norm_0: 109.93201 (93.57071) | > loss_disc: 2.86694 (2.99925) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.86694 (2.99925) | > grad_norm_1: 7.37049 (28.64848) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.24460 (10.95945) | > loader_time: 0.02230 (0.01525) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.64689 (1.64689) | > loss_kl: 1.82133 (1.82133) | > loss_feat: 1.31715 (1.31715) | > loss_mel: 20.64140 (20.64140) | > loss_duration: 0.85550 (0.85550) | > loss_0: 26.28227 (26.28227) | > loss_disc: 2.84105 (2.84105) | > loss_1: 2.84105 (2.84105)  --> STEP: 1 | > loss_gen: 1.61707 (1.61707) | > loss_kl: 2.32108 (2.32108) | > loss_feat: 1.49289 (1.49289) | > loss_mel: 21.58148 (21.58148) | > loss_duration: 0.87514 (0.87514) | > loss_0: 27.88765 (27.88765) | > loss_disc: 2.80894 (2.80894) | > loss_1: 2.80894 (2.80894)  --> STEP: 2 | > loss_gen: 1.57203 (1.59455) | > loss_kl: 2.17472 (2.24790) | > loss_feat: 1.72539 (1.60914) | > loss_mel: 20.42621 (21.00385) | > loss_duration: 0.89460 (0.88487) | > loss_0: 26.79295 (27.34030) | > loss_disc: 2.80313 (2.80604) | > loss_1: 2.80313 (2.80604)  --> STEP: 3 | > loss_gen: 1.58350 (1.59087) | > loss_kl: 2.17933 (2.22504) | > loss_feat: 1.33036 (1.51621) | > loss_mel: 20.64345 (20.88371) | > loss_duration: 0.95102 (0.90692) | > loss_0: 26.68766 (27.12276) | > loss_disc: 2.85400 (2.82202) | > loss_1: 2.85400 (2.82202) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00723 (-0.00036) | > avg_loss_gen: 1.59087 (-0.10287) | > avg_loss_kl: 2.22504 (+0.07357) | > avg_loss_feat: 1.51621 (-1.13057) | > avg_loss_mel: 20.88371 (+0.00073) | > avg_loss_duration: 0.90692 (-0.00581) | > avg_loss_0: 27.12276 (-1.16496) | > avg_loss_disc: 2.82202 (-0.16002) | > avg_loss_1: 2.82202 (-0.16002) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_4356.pth.tar  > EPOCH: 36/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 03:23:21)   --> STEP: 19/120 -- GLOBAL_STEP: 4375 | > loss_gen: 1.71237 (1.74754) | > loss_kl: 2.26103 (2.33245) | > loss_feat: 1.85675 (1.79016) | > loss_mel: 21.27382 (21.56930) | > loss_duration: 0.86461 (1.04225) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.96857 (28.48170) | > grad_norm_0: 110.23311 (81.87847) | > loss_disc: 2.83013 (2.79207) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.83013 (2.79207) | > grad_norm_1: 4.23600 (5.65846) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.21340 (9.61362) | > loader_time: 0.01020 (0.00878)  --> STEP: 44/120 -- GLOBAL_STEP: 4400 | > loss_gen: 1.64200 (1.74721) | > loss_kl: 2.14196 (2.26374) | > loss_feat: 1.77023 (1.77659) | > loss_mel: 20.71947 (21.11936) | > loss_duration: 0.91327 (0.96850) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.18692 (27.87540) | > grad_norm_0: 84.31176 (91.67954) | > loss_disc: 2.78010 (2.76851) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.78010 (2.76851) | > grad_norm_1: 19.50669 (6.88149) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.56350 (10.06524) | > loader_time: 0.01270 (0.01052)  --> STEP: 69/120 -- GLOBAL_STEP: 4425 | > loss_gen: 1.90473 (1.77191) | > loss_kl: 2.16325 (2.23185) | > loss_feat: 2.02677 (1.85092) | > loss_mel: 20.73808 (20.98351) | > loss_duration: 0.93346 (0.95266) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.76629 (27.79084) | > grad_norm_0: 176.37062 (91.57868) | > loss_disc: 2.65123 (2.74182) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65123 (2.74182) | > grad_norm_1: 11.26585 (10.61821) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.71060 (10.29340) | > loader_time: 0.01510 (0.01319)  --> STEP: 94/120 -- GLOBAL_STEP: 4450 | > loss_gen: 1.82120 (1.79598) | > loss_kl: 2.08632 (2.20022) | > loss_feat: 2.34479 (1.94127) | > loss_mel: 20.93416 (20.94012) | > loss_duration: 0.93271 (0.94822) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.11919 (27.82581) | > grad_norm_0: 92.46744 (90.14009) | > loss_disc: 2.65602 (2.71484) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65602 (2.71484) | > grad_norm_1: 12.93112 (12.31545) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.40340 (10.56815) | > loader_time: 0.01820 (0.01410)  --> STEP: 119/120 -- GLOBAL_STEP: 4475 | > loss_gen: 2.22304 (1.82572) | > loss_kl: 1.89101 (2.17138) | > loss_feat: 2.47373 (2.03086) | > loss_mel: 21.28638 (20.91234) | > loss_duration: 1.08166 (0.95368) | > amp_scaler: 1024.00000 (559.32773) | > loss_0: 28.95582 (27.89398) | > grad_norm_0: 69.02879 (83.30838) | > loss_disc: 2.59881 (2.69383) | > amp_scaler-1: 1024.00000 (563.63025) | > loss_1: 2.59881 (2.69383) | > grad_norm_1: 43.81775 (14.94527) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.22580 (10.88430) | > loader_time: 0.02370 (0.01549) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.86731 (1.86731) | > loss_kl: 1.86466 (1.86466) | > loss_feat: 2.46950 (2.46950) | > loss_mel: 21.85362 (21.85362) | > loss_duration: 0.88148 (0.88148) | > loss_0: 28.93657 (28.93657) | > loss_disc: 2.64030 (2.64030) | > loss_1: 2.64030 (2.64030)  --> STEP: 1 | > loss_gen: 1.79862 (1.79862) | > loss_kl: 2.29108 (2.29108) | > loss_feat: 2.36004 (2.36004) | > loss_mel: 21.44844 (21.44844) | > loss_duration: 0.87952 (0.87952) | > loss_0: 28.77771 (28.77771) | > loss_disc: 2.73456 (2.73456) | > loss_1: 2.73456 (2.73456)  --> STEP: 2 | > loss_gen: 1.93333 (1.86598) | > loss_kl: 2.12679 (2.20893) | > loss_feat: 2.96394 (2.66199) | > loss_mel: 21.94752 (21.69798) | > loss_duration: 0.91057 (0.89504) | > loss_0: 29.88214 (29.32993) | > loss_disc: 2.53582 (2.63519) | > loss_1: 2.53582 (2.63519)  --> STEP: 3 | > loss_gen: 2.02541 (1.91912) | > loss_kl: 2.04492 (2.15426) | > loss_feat: 2.68441 (2.66946) | > loss_mel: 20.83260 (21.40952) | > loss_duration: 0.96450 (0.91820) | > loss_0: 28.55184 (29.07056) | > loss_disc: 2.53907 (2.60315) | > loss_1: 2.53907 (2.60315) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00733 (+0.00010) | > avg_loss_gen: 1.91912 (+0.32826) | > avg_loss_kl: 2.15426 (-0.07078) | > avg_loss_feat: 2.66946 (+1.15325) | > avg_loss_mel: 21.40952 (+0.52580) | > avg_loss_duration: 0.91820 (+0.01128) | > avg_loss_0: 29.07056 (+1.94781) | > avg_loss_disc: 2.60315 (-0.21888) | > avg_loss_1: 2.60315 (-0.21888)  > EPOCH: 37/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 03:45:25)   --> STEP: 23/120 -- GLOBAL_STEP: 4500 | > loss_gen: 1.79536 (1.99869) | > loss_kl: 2.24048 (2.32178) | > loss_feat: 2.87372 (2.76789) | > loss_mel: 21.40358 (21.75157) | > loss_duration: 0.95425 (1.02407) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.26739 (29.86400) | > grad_norm_0: 167.39088 (76.53824) | > loss_disc: 2.58406 (2.58479) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58406 (2.58479) | > grad_norm_1: 36.73252 (19.49712) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.25610 (10.16582) | > loader_time: 0.01080 (0.00887)  --> STEP: 48/120 -- GLOBAL_STEP: 4525 | > loss_gen: 1.90566 (1.98582) | > loss_kl: 2.11188 (2.26077) | > loss_feat: 2.30847 (2.68548) | > loss_mel: 21.43847 (21.37740) | > loss_duration: 0.89473 (0.96323) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.65921 (29.27269) | > grad_norm_0: 197.98611 (72.96942) | > loss_disc: 2.64616 (2.59153) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.64616 (2.59153) | > grad_norm_1: 5.36177 (18.67839) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.92510 (10.38494) | > loader_time: 0.01300 (0.01063)  --> STEP: 73/120 -- GLOBAL_STEP: 4550 | > loss_gen: 1.81531 (1.98814) | > loss_kl: 2.10276 (2.21413) | > loss_feat: 2.64253 (2.69499) | > loss_mel: 20.83069 (21.21972) | > loss_duration: 0.94833 (0.95171) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.33962 (29.06868) | > grad_norm_0: 25.44579 (72.70032) | > loss_disc: 2.52998 (2.58300) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.52998 (2.58300) | > grad_norm_1: 11.48787 (18.60485) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.94680 (10.54459) | > loader_time: 0.01570 (0.01361)  --> STEP: 98/120 -- GLOBAL_STEP: 4575 | > loss_gen: 1.71663 (1.98986) | > loss_kl: 2.01831 (2.18538) | > loss_feat: 2.79354 (2.69432) | > loss_mel: 20.60783 (21.09892) | > loss_duration: 0.94495 (0.94733) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.08126 (28.91581) | > grad_norm_0: 99.83982 (73.28949) | > loss_disc: 2.69794 (2.58466) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.69794 (2.58466) | > grad_norm_1: 78.20098 (19.50519) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.65960 (10.77332) | > loader_time: 0.01880 (0.01464) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.94353 (1.94353) | > loss_kl: 2.24139 (2.24139) | > loss_feat: 2.86365 (2.86365) | > loss_mel: 21.29778 (21.29778) | > loss_duration: 0.85080 (0.85080) | > loss_0: 29.19714 (29.19714) | > loss_disc: 2.47074 (2.47074) | > loss_1: 2.47074 (2.47074)  --> STEP: 1 | > loss_gen: 1.95426 (1.95426) | > loss_kl: 2.37494 (2.37494) | > loss_feat: 2.27895 (2.27895) | > loss_mel: 21.90219 (21.90219) | > loss_duration: 0.87630 (0.87630) | > loss_0: 29.38664 (29.38664) | > loss_disc: 2.68622 (2.68622) | > loss_1: 2.68622 (2.68622)  --> STEP: 2 | > loss_gen: 1.83420 (1.89423) | > loss_kl: 2.03269 (2.20382) | > loss_feat: 2.66156 (2.47025) | > loss_mel: 21.00229 (21.45224) | > loss_duration: 0.89134 (0.88382) | > loss_0: 28.42208 (28.90436) | > loss_disc: 2.59784 (2.64203) | > loss_1: 2.59784 (2.64203)  --> STEP: 3 | > loss_gen: 1.81188 (1.86678) | > loss_kl: 2.13797 (2.18187) | > loss_feat: 2.51505 (2.48518) | > loss_mel: 20.23989 (21.04812) | > loss_duration: 0.95476 (0.90747) | > loss_0: 27.65955 (28.48942) | > loss_disc: 2.63221 (2.63876) | > loss_1: 2.63221 (2.63876) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00723 (-0.00009) | > avg_loss_gen: 1.86678 (-0.05234) | > avg_loss_kl: 2.18187 (+0.02761) | > avg_loss_feat: 2.48518 (-0.18428) | > avg_loss_mel: 21.04812 (-0.36140) | > avg_loss_duration: 0.90747 (-0.01073) | > avg_loss_0: 28.48942 (-0.58114) | > avg_loss_disc: 2.63876 (+0.03561) | > avg_loss_1: 2.63876 (+0.03561)  > EPOCH: 38/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 04:07:46)   --> STEP: 2/120 -- GLOBAL_STEP: 4600 | > loss_gen: 2.44641 (2.19146) | > loss_kl: 2.49223 (2.47796) | > loss_feat: 2.75645 (3.00210) | > loss_mel: 22.14102 (22.42084) | > loss_duration: 0.94482 (0.92487) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.78094 (31.01724) | > grad_norm_0: 35.19532 (105.68960) | > loss_disc: 2.75093 (2.67890) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.75093 (2.67890) | > grad_norm_1: 12.66860 (11.18926) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.84050 (9.69498) | > loader_time: 0.00580 (0.00606)  --> STEP: 27/120 -- GLOBAL_STEP: 4625 | > loss_gen: 2.00137 (1.98992) | > loss_kl: 2.30221 (2.33041) | > loss_feat: 2.82950 (2.78248) | > loss_mel: 20.73552 (21.58467) | > loss_duration: 0.90653 (1.00583) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.77513 (29.69332) | > grad_norm_0: 21.43820 (65.98312) | > loss_disc: 2.54417 (2.62725) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54417 (2.62725) | > grad_norm_1: 8.82514 (14.52762) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.69860 (10.26063) | > loader_time: 0.01080 (0.00927)  --> STEP: 52/120 -- GLOBAL_STEP: 4650 | > loss_gen: 2.19229 (1.98537) | > loss_kl: 2.26271 (2.26016) | > loss_feat: 2.76303 (2.73454) | > loss_mel: 20.38483 (21.27678) | > loss_duration: 0.91193 (0.95948) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.51479 (29.21632) | > grad_norm_0: 87.87191 (81.21205) | > loss_disc: 2.54153 (2.62066) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.54153 (2.62066) | > grad_norm_1: 28.07656 (19.19453) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.94470 (10.44394) | > loader_time: 0.01380 (0.01086)  --> STEP: 77/120 -- GLOBAL_STEP: 4675 | > loss_gen: 1.66579 (1.98950) | > loss_kl: 2.18171 (2.22665) | > loss_feat: 2.73858 (2.74006) | > loss_mel: 20.88960 (21.15004) | > loss_duration: 0.92734 (0.95089) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.40302 (29.05713) | > grad_norm_0: 35.79607 (82.44253) | > loss_disc: 2.61324 (2.61547) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.61324 (2.61547) | > grad_norm_1: 57.42747 (19.71548) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.03010 (10.58573) | > loader_time: 0.01620 (0.01326)  --> STEP: 102/120 -- GLOBAL_STEP: 4700 | > loss_gen: 2.18344 (1.99439) | > loss_kl: 2.13470 (2.19921) | > loss_feat: 3.05401 (2.75665) | > loss_mel: 20.24219 (21.06660) | > loss_duration: 0.93391 (0.94677) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.54826 (28.96363) | > grad_norm_0: 38.47262 (78.51486) | > loss_disc: 2.50538 (2.60633) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.50538 (2.60633) | > grad_norm_1: 6.59873 (19.00121) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.88230 (10.86514) | > loader_time: 0.01940 (0.01441) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.07880 (2.07880) | > loss_kl: 2.25148 (2.25148) | > loss_feat: 2.85208 (2.85208) | > loss_mel: 20.98053 (20.98053) | > loss_duration: 0.85029 (0.85029) | > loss_0: 29.01318 (29.01318) | > loss_disc: 2.52366 (2.52366) | > loss_1: 2.52366 (2.52366)  --> STEP: 1 | > loss_gen: 2.04309 (2.04309) | > loss_kl: 2.25957 (2.25957) | > loss_feat: 2.89163 (2.89163) | > loss_mel: 20.87606 (20.87606) | > loss_duration: 0.88976 (0.88976) | > loss_0: 28.96011 (28.96011) | > loss_disc: 2.52368 (2.52368) | > loss_1: 2.52368 (2.52368)  --> STEP: 2 | > loss_gen: 2.05524 (2.04917) | > loss_kl: 1.99674 (2.12816) | > loss_feat: 2.55900 (2.72531) | > loss_mel: 20.20062 (20.53834) | > loss_duration: 0.90025 (0.89500) | > loss_0: 27.71184 (28.33597) | > loss_disc: 2.60224 (2.56296) | > loss_1: 2.60224 (2.56296)  --> STEP: 3 | > loss_gen: 1.91508 (2.00447) | > loss_kl: 2.20170 (2.15267) | > loss_feat: 2.65276 (2.70113) | > loss_mel: 20.00644 (20.36104) | > loss_duration: 0.95454 (0.91485) | > loss_0: 27.73053 (28.13416) | > loss_disc: 2.64815 (2.59136) | > loss_1: 2.64815 (2.59136) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00651 (-0.00073) | > avg_loss_gen: 2.00447 (+0.13769) | > avg_loss_kl: 2.15267 (-0.02920) | > avg_loss_feat: 2.70113 (+0.21594) | > avg_loss_mel: 20.36104 (-0.68708) | > avg_loss_duration: 0.91485 (+0.00738) | > avg_loss_0: 28.13416 (-0.35526) | > avg_loss_disc: 2.59136 (-0.04740) | > avg_loss_1: 2.59136 (-0.04740)  > EPOCH: 39/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 04:30:11)   --> STEP: 6/120 -- GLOBAL_STEP: 4725 | > loss_gen: 2.25630 (2.15863) | > loss_kl: 2.48135 (2.39749) | > loss_feat: 3.58855 (3.32969) | > loss_mel: 22.47413 (21.92003) | > loss_duration: 0.96309 (0.91939) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.76342 (30.72524) | > grad_norm_0: 105.24962 (108.64659) | > loss_disc: 2.45191 (2.43790) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.45191 (2.43790) | > grad_norm_1: 9.37238 (17.60475) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.01860 (9.94875) | > loader_time: 0.00840 (0.00753)  --> STEP: 31/120 -- GLOBAL_STEP: 4750 | > loss_gen: 1.91350 (2.02746) | > loss_kl: 2.29873 (2.30746) | > loss_feat: 2.97613 (2.97588) | > loss_mel: 20.53222 (21.22991) | > loss_duration: 0.89194 (0.99488) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.61252 (29.53560) | > grad_norm_0: 47.05797 (90.00369) | > loss_disc: 2.60017 (2.57462) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60017 (2.57462) | > grad_norm_1: 9.36053 (19.25965) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.43670 (10.29214) | > loader_time: 0.01190 (0.00964)  --> STEP: 56/120 -- GLOBAL_STEP: 4775 | > loss_gen: 1.97075 (2.01781) | > loss_kl: 2.19732 (2.26477) | > loss_feat: 2.51723 (2.90111) | > loss_mel: 20.98663 (21.07009) | > loss_duration: 0.92239 (0.95867) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.59431 (29.21244) | > grad_norm_0: 56.07928 (98.49906) | > loss_disc: 2.67971 (2.59708) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.67971 (2.59708) | > grad_norm_1: 9.38756 (19.29654) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.87570 (10.45848) | > loader_time: 0.03590 (0.01195)  --> STEP: 81/120 -- GLOBAL_STEP: 4800 | > loss_gen: 1.98602 (2.02302) | > loss_kl: 2.05160 (2.22976) | > loss_feat: 2.82503 (2.88465) | > loss_mel: 20.46974 (20.98134) | > loss_duration: 0.92134 (0.95123) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.25373 (29.07000) | > grad_norm_0: 34.15067 (105.02464) | > loss_disc: 2.62073 (2.58658) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.62073 (2.58658) | > grad_norm_1: 13.01594 (19.48283) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.57980 (10.61140) | > loader_time: 0.01660 (0.01351)  --> STEP: 106/120 -- GLOBAL_STEP: 4825 | > loss_gen: 1.98953 (2.02266) | > loss_kl: 2.01756 (2.20279) | > loss_feat: 2.94642 (2.88666) | > loss_mel: 20.24099 (20.92529) | > loss_duration: 0.99243 (0.94833) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.18694 (28.98573) | > grad_norm_0: 54.94044 (96.59637) | > loss_disc: 2.55184 (2.58206) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.55184 (2.58206) | > grad_norm_1: 21.02821 (19.46139) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.92060 (10.85734) | > loader_time: 0.02020 (0.01464) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.13348 (2.13348) | > loss_kl: 1.82470 (1.82470) | > loss_feat: 2.77431 (2.77431) | > loss_mel: 22.02114 (22.02114) | > loss_duration: 0.84971 (0.84971) | > loss_0: 29.60334 (29.60334) | > loss_disc: 2.59650 (2.59650) | > loss_1: 2.59650 (2.59650)  --> STEP: 1 | > loss_gen: 1.89064 (1.89064) | > loss_kl: 2.30896 (2.30896) | > loss_feat: 2.76243 (2.76243) | > loss_mel: 21.97668 (21.97668) | > loss_duration: 0.87936 (0.87936) | > loss_0: 29.81806 (29.81806) | > loss_disc: 2.71508 (2.71508) | > loss_1: 2.71508 (2.71508)  --> STEP: 2 | > loss_gen: 2.06643 (1.97854) | > loss_kl: 2.25014 (2.27955) | > loss_feat: 2.65773 (2.71008) | > loss_mel: 21.62083 (21.79875) | > loss_duration: 0.89248 (0.88592) | > loss_0: 29.48762 (29.65284) | > loss_disc: 2.70788 (2.71148) | > loss_1: 2.70788 (2.71148)  --> STEP: 3 | > loss_gen: 2.00343 (1.98684) | > loss_kl: 2.27497 (2.27802) | > loss_feat: 3.00930 (2.80982) | > loss_mel: 20.73313 (21.44354) | > loss_duration: 0.95159 (0.90781) | > loss_0: 28.97242 (29.42603) | > loss_disc: 2.61635 (2.67977) | > loss_1: 2.61635 (2.67977) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00748 (+0.00097) | > avg_loss_gen: 1.98684 (-0.01763) | > avg_loss_kl: 2.27802 (+0.12535) | > avg_loss_feat: 2.80982 (+0.10869) | > avg_loss_mel: 21.44354 (+1.08250) | > avg_loss_duration: 0.90781 (-0.00704) | > avg_loss_0: 29.42603 (+1.29187) | > avg_loss_disc: 2.67977 (+0.08841) | > avg_loss_1: 2.67977 (+0.08841)  > EPOCH: 40/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 04:52:30)   --> STEP: 10/120 -- GLOBAL_STEP: 4850 | > loss_gen: 2.20703 (2.11399) | > loss_kl: 2.40001 (2.37554) | > loss_feat: 2.88429 (3.16689) | > loss_mel: 21.23331 (21.90886) | > loss_duration: 0.85731 (0.90791) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.58195 (30.47319) | > grad_norm_0: 183.85461 (143.05222) | > loss_disc: 2.60523 (2.53153) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60523 (2.53153) | > grad_norm_1: 26.12987 (18.21742) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.00770 (9.85166) | > loader_time: 0.00900 (0.00745)  --> STEP: 35/120 -- GLOBAL_STEP: 4875 | > loss_gen: 1.92518 (2.05407) | > loss_kl: 2.31102 (2.30308) | > loss_feat: 2.59367 (3.04778) | > loss_mel: 21.26779 (21.33574) | > loss_duration: 0.88782 (0.98235) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.98547 (29.72302) | > grad_norm_0: 170.47096 (99.77834) | > loss_disc: 2.78550 (2.59153) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.78550 (2.59153) | > grad_norm_1: 20.87645 (24.68087) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.49280 (10.17000) | > loader_time: 0.01230 (0.00979)  --> STEP: 60/120 -- GLOBAL_STEP: 4900 | > loss_gen: 1.74899 (2.02091) | > loss_kl: 2.14955 (2.25390) | > loss_feat: 2.91365 (2.91249) | > loss_mel: 21.16617 (21.14808) | > loss_duration: 0.92877 (0.95432) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.90712 (29.28970) | > grad_norm_0: 75.16118 (85.65840) | > loss_disc: 2.66336 (2.60504) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.66336 (2.60504) | > grad_norm_1: 20.60901 (21.67756) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.71760 (10.33531) | > loader_time: 0.01420 (0.01334)  --> STEP: 85/120 -- GLOBAL_STEP: 4925 | > loss_gen: 2.16477 (2.02051) | > loss_kl: 2.07419 (2.21781) | > loss_feat: 2.75960 (2.90177) | > loss_mel: 21.20639 (21.02071) | > loss_duration: 0.95744 (0.94875) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.16240 (29.10955) | > grad_norm_0: 178.33411 (90.23669) | > loss_disc: 2.71477 (2.59353) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.71477 (2.59353) | > grad_norm_1: 13.14041 (20.11563) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.07850 (10.50530) | > loader_time: 0.01690 (0.01402)  --> STEP: 110/120 -- GLOBAL_STEP: 4950 | > loss_gen: 2.07916 (2.01548) | > loss_kl: 2.03227 (2.18856) | > loss_feat: 2.64210 (2.86937) | > loss_mel: 21.10248 (20.92690) | > loss_duration: 0.98988 (0.94787) | > amp_scaler: 512.00000 (912.29091) | > loss_0: 28.84589 (28.94819) | > grad_norm_0: 53.39550 (90.37463) | > loss_disc: 2.61791 (2.59996) | > amp_scaler-1: 512.00000 (912.29091) | > loss_1: 2.61791 (2.59996) | > grad_norm_1: 9.50029 (20.99452) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.95640 (10.79487) | > loader_time: 0.02170 (0.01514) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.11513 (2.11513) | > loss_kl: 2.33154 (2.33154) | > loss_feat: 2.87122 (2.87122) | > loss_mel: 21.21385 (21.21385) | > loss_duration: 0.86636 (0.86636) | > loss_0: 29.39809 (29.39809) | > loss_disc: 2.60805 (2.60805) | > loss_1: 2.60805 (2.60805)  --> STEP: 1 | > loss_gen: 2.16451 (2.16451) | > loss_kl: 2.10998 (2.10998) | > loss_feat: 2.86467 (2.86467) | > loss_mel: 20.11412 (20.11412) | > loss_duration: 0.88752 (0.88752) | > loss_0: 28.14081 (28.14081) | > loss_disc: 2.63076 (2.63076) | > loss_1: 2.63076 (2.63076)  --> STEP: 2 | > loss_gen: 2.11374 (2.13913) | > loss_kl: 2.12491 (2.11744) | > loss_feat: 2.84772 (2.85619) | > loss_mel: 19.81467 (19.96440) | > loss_duration: 0.90972 (0.89862) | > loss_0: 27.81076 (27.97579) | > loss_disc: 2.54509 (2.58793) | > loss_1: 2.54509 (2.58793)  --> STEP: 3 | > loss_gen: 2.04701 (2.10842) | > loss_kl: 2.11154 (2.11548) | > loss_feat: 3.10607 (2.93948) | > loss_mel: 20.54047 (20.15642) | > loss_duration: 0.96584 (0.92103) | > loss_0: 28.77093 (28.24083) | > loss_disc: 2.61700 (2.59762) | > loss_1: 2.61700 (2.59762) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00724 (-0.00024) | > avg_loss_gen: 2.10842 (+0.12159) | > avg_loss_kl: 2.11548 (-0.16255) | > avg_loss_feat: 2.93948 (+0.12966) | > avg_loss_mel: 20.15642 (-1.28712) | > avg_loss_duration: 0.92103 (+0.01322) | > avg_loss_0: 28.24083 (-1.18520) | > avg_loss_disc: 2.59762 (-0.08215) | > avg_loss_1: 2.59762 (-0.08215)  > EPOCH: 41/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 05:14:37)   --> STEP: 14/120 -- GLOBAL_STEP: 4975 | > loss_gen: 2.23437 (2.09985) | > loss_kl: 2.36629 (2.37195) | > loss_feat: 3.12348 (3.18240) | > loss_mel: 21.25629 (21.29639) | > loss_duration: 0.90560 (1.08873) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.88604 (30.03932) | > grad_norm_0: 96.71295 (83.44917) | > loss_disc: 2.55567 (2.50275) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55567 (2.50275) | > grad_norm_1: 9.07856 (9.87687) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.14260 (9.99677) | > loader_time: 0.00930 (0.00812)  --> STEP: 39/120 -- GLOBAL_STEP: 5000 | > loss_gen: 1.87761 (2.07115) | > loss_kl: 2.31673 (2.31285) | > loss_feat: 3.25974 (3.10296) | > loss_mel: 21.78297 (21.08128) | > loss_duration: 0.92805 (0.97426) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.16510 (29.54250) | > grad_norm_0: 44.58147 (86.84602) | > loss_disc: 2.51541 (2.53168) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51541 (2.53168) | > grad_norm_1: 18.96988 (15.98345) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34040 (10.23605) | > loader_time: 0.01180 (0.01012)  --> STEP: 64/120 -- GLOBAL_STEP: 5025 | > loss_gen: 2.12098 (2.05623) | > loss_kl: 2.18059 (2.25300) | > loss_feat: 3.15482 (3.02280) | > loss_mel: 20.71492 (21.01583) | > loss_duration: 0.94584 (0.95169) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.11715 (29.29954) | > grad_norm_0: 93.81972 (85.39272) | > loss_disc: 2.48531 (2.55576) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48531 (2.55576) | > grad_norm_1: 15.93058 (19.12332) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.65470 (10.39669) | > loader_time: 0.01450 (0.01225)  --> STEP: 89/120 -- GLOBAL_STEP: 5050 | > loss_gen: 1.71375 (2.04653) | > loss_kl: 2.14686 (2.21352) | > loss_feat: 3.01787 (2.98786) | > loss_mel: 20.65455 (20.91490) | > loss_duration: 0.93513 (0.94652) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.46816 (29.10933) | > grad_norm_0: 150.87215 (92.81554) | > loss_disc: 2.59496 (2.55712) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59496 (2.55712) | > grad_norm_1: 22.31337 (19.12883) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.57780 (10.59456) | > loader_time: 0.02030 (0.01369)  --> STEP: 114/120 -- GLOBAL_STEP: 5075 | > loss_gen: 2.10335 (2.05360) | > loss_kl: 2.02850 (2.19196) | > loss_feat: 2.80789 (2.99234) | > loss_mel: 20.92542 (20.86255) | > loss_duration: 0.99335 (0.94697) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.85851 (29.04741) | > grad_norm_0: 36.91899 (89.51205) | > loss_disc: 2.49172 (2.55229) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.49172 (2.55229) | > grad_norm_1: 32.09967 (19.58777) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.88130 (10.87734) | > loader_time: 0.02100 (0.01506) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.33907 (2.33907) | > loss_kl: 1.93615 (1.93615) | > loss_feat: 3.02876 (3.02876) | > loss_mel: 21.35789 (21.35789) | > loss_duration: 0.85863 (0.85863) | > loss_0: 29.52051 (29.52051) | > loss_disc: 2.47186 (2.47186) | > loss_1: 2.47186 (2.47186)  --> STEP: 1 | > loss_gen: 2.30387 (2.30387) | > loss_kl: 2.31530 (2.31530) | > loss_feat: 2.86028 (2.86028) | > loss_mel: 20.58508 (20.58508) | > loss_duration: 0.88609 (0.88609) | > loss_0: 28.95062 (28.95062) | > loss_disc: 2.59889 (2.59889) | > loss_1: 2.59889 (2.59889)  --> STEP: 2 | > loss_gen: 2.20556 (2.25471) | > loss_kl: 2.01038 (2.16284) | > loss_feat: 2.94806 (2.90417) | > loss_mel: 20.95314 (20.76911) | > loss_duration: 0.90449 (0.89529) | > loss_0: 29.02163 (28.98613) | > loss_disc: 2.62852 (2.61370) | > loss_1: 2.62852 (2.61370)  --> STEP: 3 | > loss_gen: 2.15475 (2.22139) | > loss_kl: 2.35139 (2.22569) | > loss_feat: 3.29283 (3.03372) | > loss_mel: 21.77157 (21.10326) | > loss_duration: 0.95533 (0.91530) | > loss_0: 30.52587 (29.49937) | > loss_disc: 2.50995 (2.57912) | > loss_1: 2.50995 (2.57912) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00740 (+0.00016) | > avg_loss_gen: 2.22139 (+0.11297) | > avg_loss_kl: 2.22569 (+0.11021) | > avg_loss_feat: 3.03372 (+0.09424) | > avg_loss_mel: 21.10326 (+0.94684) | > avg_loss_duration: 0.91530 (-0.00573) | > avg_loss_0: 29.49937 (+1.25854) | > avg_loss_disc: 2.57912 (-0.01850) | > avg_loss_1: 2.57912 (-0.01850)  > EPOCH: 42/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 05:36:49)   --> STEP: 18/120 -- GLOBAL_STEP: 5100 | > loss_gen: 1.94190 (2.02100) | > loss_kl: 2.26796 (2.37011) | > loss_feat: 2.71648 (3.01324) | > loss_mel: 20.93548 (21.42177) | > loss_duration: 0.91555 (1.05014) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.77738 (29.87626) | > grad_norm_0: 125.59540 (93.63011) | > loss_disc: 2.69059 (2.65010) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.69059 (2.65010) | > grad_norm_1: 10.75314 (22.66121) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.09650 (10.04667) | > loader_time: 0.01020 (0.00872)  --> STEP: 43/120 -- GLOBAL_STEP: 5125 | > loss_gen: 2.13215 (2.02926) | > loss_kl: 2.18890 (2.28674) | > loss_feat: 3.14109 (2.96319) | > loss_mel: 20.46551 (21.22987) | > loss_duration: 0.87713 (0.96809) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.80479 (29.47716) | > grad_norm_0: 25.07912 (110.21626) | > loss_disc: 2.50418 (2.60076) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50418 (2.60076) | > grad_norm_1: 29.10999 (22.22150) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.66290 (10.26981) | > loader_time: 0.01290 (0.01051)  --> STEP: 68/120 -- GLOBAL_STEP: 5150 | > loss_gen: 1.87851 (2.02782) | > loss_kl: 2.02750 (2.23109) | > loss_feat: 2.89254 (2.92971) | > loss_mel: 21.12136 (21.08840) | > loss_duration: 0.93589 (0.95118) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.85579 (29.22820) | > grad_norm_0: 60.82270 (106.11436) | > loss_disc: 2.62542 (2.60641) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.62542 (2.60641) | > grad_norm_1: 39.76780 (21.86456) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.48600 (10.41950) | > loader_time: 0.01510 (0.01218)  --> STEP: 93/120 -- GLOBAL_STEP: 5175 | > loss_gen: 2.21719 (2.02920) | > loss_kl: 2.03120 (2.19639) | > loss_feat: 3.26796 (2.91750) | > loss_mel: 20.32442 (20.89049) | > loss_duration: 0.93637 (0.94641) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.77714 (28.97999) | > grad_norm_0: 112.58363 (97.94071) | > loss_disc: 2.43070 (2.59204) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.43070 (2.59204) | > grad_norm_1: 7.44571 (20.75038) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.58310 (10.64101) | > loader_time: 0.01820 (0.01339)  --> STEP: 118/120 -- GLOBAL_STEP: 5200 | > loss_gen: 2.28639 (2.03474) | > loss_kl: 2.05852 (2.17724) | > loss_feat: 2.90339 (2.92975) | > loss_mel: 20.55179 (20.82101) | > loss_duration: 1.03565 (0.95041) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.83574 (28.91316) | > grad_norm_0: 123.50945 (91.02852) | > loss_disc: 2.54480 (2.58857) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54480 (2.58857) | > grad_norm_1: 16.63628 (22.27996) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.78070 (10.94889) | > loader_time: 0.02390 (0.01491) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.21519 (2.21519) | > loss_kl: 2.09036 (2.09036) | > loss_feat: 3.26753 (3.26753) | > loss_mel: 21.58841 (21.58841) | > loss_duration: 0.86739 (0.86739) | > loss_0: 30.02888 (30.02888) | > loss_disc: 2.41104 (2.41104) | > loss_1: 2.41104 (2.41104)  --> STEP: 1 | > loss_gen: 2.02704 (2.02704) | > loss_kl: 2.58452 (2.58452) | > loss_feat: 2.88553 (2.88553) | > loss_mel: 20.58415 (20.58415) | > loss_duration: 0.88391 (0.88391) | > loss_0: 28.96515 (28.96515) | > loss_disc: 2.57448 (2.57448) | > loss_1: 2.57448 (2.57448)  --> STEP: 2 | > loss_gen: 2.09840 (2.06272) | > loss_kl: 2.25499 (2.41976) | > loss_feat: 3.21264 (3.04908) | > loss_mel: 20.57254 (20.57834) | > loss_duration: 0.90663 (0.89527) | > loss_0: 29.04520 (29.00518) | > loss_disc: 2.40454 (2.48951) | > loss_1: 2.40454 (2.48951)  --> STEP: 3 | > loss_gen: 2.12856 (2.08467) | > loss_kl: 2.10434 (2.31462) | > loss_feat: 2.97604 (3.02474) | > loss_mel: 20.23171 (20.46280) | > loss_duration: 0.94737 (0.91264) | > loss_0: 28.38803 (28.79946) | > loss_disc: 2.49543 (2.49148) | > loss_1: 2.49543 (2.49148) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00732 (-0.00008) | > avg_loss_gen: 2.08467 (-0.13672) | > avg_loss_kl: 2.31462 (+0.08893) | > avg_loss_feat: 3.02474 (-0.00899) | > avg_loss_mel: 20.46280 (-0.64046) | > avg_loss_duration: 0.91264 (-0.00267) | > avg_loss_0: 28.79946 (-0.69991) | > avg_loss_disc: 2.49148 (-0.08764) | > avg_loss_1: 2.49148 (-0.08764)  > EPOCH: 43/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 05:59:03)   --> STEP: 22/120 -- GLOBAL_STEP: 5225 | > loss_gen: 2.31964 (2.06861) | > loss_kl: 2.16637 (2.31708) | > loss_feat: 3.55054 (3.13338) | > loss_mel: 21.19456 (21.33331) | > loss_duration: 0.89942 (1.01888) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.13052 (29.87127) | > grad_norm_0: 47.76867 (73.48834) | > loss_disc: 2.46910 (2.58488) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.46910 (2.58488) | > grad_norm_1: 26.21136 (11.33615) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.39460 (10.14261) | > loader_time: 0.01110 (0.00879)  --> STEP: 47/120 -- GLOBAL_STEP: 5250 | > loss_gen: 2.07922 (2.05430) | > loss_kl: 2.33872 (2.26785) | > loss_feat: 2.97795 (3.05451) | > loss_mel: 20.54167 (21.02520) | > loss_duration: 0.92733 (0.95921) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.86488 (29.36107) | > grad_norm_0: 52.86697 (68.29431) | > loss_disc: 2.62732 (2.57931) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.62732 (2.57931) | > grad_norm_1: 11.07434 (18.24682) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.69870 (10.34973) | > loader_time: 0.01340 (0.01060)  --> STEP: 72/120 -- GLOBAL_STEP: 5275 | > loss_gen: 1.92137 (2.04966) | > loss_kl: 1.88580 (2.22923) | > loss_feat: 2.77735 (3.02452) | > loss_mel: 20.66794 (20.90404) | > loss_duration: 0.93062 (0.94830) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.18307 (29.15575) | > grad_norm_0: 120.93546 (75.50336) | > loss_disc: 2.58025 (2.57113) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58025 (2.57113) | > grad_norm_1: 21.38129 (17.41249) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.82860 (10.47562) | > loader_time: 0.01520 (0.01288)  --> STEP: 97/120 -- GLOBAL_STEP: 5300 | > loss_gen: 1.56373 (2.04416) | > loss_kl: 2.16757 (2.20460) | > loss_feat: 3.21952 (3.00332) | > loss_mel: 20.15838 (20.84050) | > loss_duration: 0.94961 (0.94443) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.05882 (29.03701) | > grad_norm_0: 69.46932 (81.44997) | > loss_disc: 2.70567 (2.57554) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.70567 (2.57554) | > grad_norm_1: 38.08490 (20.23454) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.45510 (10.69319) | > loader_time: 0.01880 (0.01397) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.16474 (2.16474) | > loss_kl: 2.24303 (2.24303) | > loss_feat: 3.04055 (3.04055) | > loss_mel: 20.71375 (20.71375) | > loss_duration: 0.86263 (0.86263) | > loss_0: 29.02470 (29.02470) | > loss_disc: 2.45160 (2.45160) | > loss_1: 2.45160 (2.45160)  --> STEP: 1 | > loss_gen: 2.12169 (2.12169) | > loss_kl: 2.29240 (2.29240) | > loss_feat: 3.16029 (3.16029) | > loss_mel: 21.94158 (21.94158) | > loss_duration: 0.88706 (0.88706) | > loss_0: 30.40302 (30.40302) | > loss_disc: 2.47692 (2.47692) | > loss_1: 2.47692 (2.47692)  --> STEP: 2 | > loss_gen: 2.12418 (2.12293) | > loss_kl: 2.20858 (2.25049) | > loss_feat: 2.87997 (3.02013) | > loss_mel: 21.28674 (21.61416) | > loss_duration: 0.90300 (0.89503) | > loss_0: 29.40247 (29.90274) | > loss_disc: 2.53947 (2.50820) | > loss_1: 2.53947 (2.50820)  --> STEP: 3 | > loss_gen: 1.98011 (2.07533) | > loss_kl: 2.20333 (2.23477) | > loss_feat: 2.85962 (2.96662) | > loss_mel: 19.21090 (20.81307) | > loss_duration: 0.96335 (0.91780) | > loss_0: 27.21730 (29.00760) | > loss_disc: 2.61986 (2.54542) | > loss_1: 2.61986 (2.54542) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00742 (+0.00010) | > avg_loss_gen: 2.07533 (-0.00934) | > avg_loss_kl: 2.23477 (-0.07985) | > avg_loss_feat: 2.96662 (-0.05811) | > avg_loss_mel: 20.81307 (+0.35027) | > avg_loss_duration: 0.91780 (+0.00517) | > avg_loss_0: 29.00760 (+0.20813) | > avg_loss_disc: 2.54542 (+0.05393) | > avg_loss_1: 2.54542 (+0.05393)  > EPOCH: 44/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 06:21:13)   --> STEP: 1/120 -- GLOBAL_STEP: 5325 | > loss_gen: 2.21773 (2.21773) | > loss_kl: 1.88590 (1.88590) | > loss_feat: 3.37910 (3.37910) | > loss_mel: 22.37125 (22.37125) | > loss_duration: 0.94724 (0.94724) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.80121 (30.80121) | > grad_norm_0: 187.95576 (187.95576) | > loss_disc: 2.40602 (2.40602) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.40602 (2.40602) | > grad_norm_1: 7.03651 (7.03651) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.31010 (9.31014) | > loader_time: 0.00550 (0.00552)  --> STEP: 26/120 -- GLOBAL_STEP: 5350 | > loss_gen: 1.89513 (2.08228) | > loss_kl: 2.40447 (2.33025) | > loss_feat: 2.79505 (3.16402) | > loss_mel: 20.87568 (21.22873) | > loss_duration: 0.90569 (1.00187) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.87602 (29.80715) | > grad_norm_0: 32.14994 (95.68089) | > loss_disc: 2.58639 (2.57072) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58639 (2.57072) | > grad_norm_1: 10.47587 (21.96480) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.35600 (10.04100) | > loader_time: 0.01130 (0.00918)  --> STEP: 51/120 -- GLOBAL_STEP: 5375 | > loss_gen: 1.92058 (2.06026) | > loss_kl: 1.94041 (2.25262) | > loss_feat: 2.94612 (3.05248) | > loss_mel: 20.56470 (20.96644) | > loss_duration: 0.91123 (0.95638) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.28304 (29.28818) | > grad_norm_0: 32.16292 (107.07943) | > loss_disc: 2.50373 (2.57578) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50373 (2.57578) | > grad_norm_1: 20.98163 (22.28160) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.43890 (10.22499) | > loader_time: 0.01360 (0.01080)  --> STEP: 76/120 -- GLOBAL_STEP: 5400 | > loss_gen: 2.04119 (2.05316) | > loss_kl: 2.16926 (2.22547) | > loss_feat: 2.86707 (3.01968) | > loss_mel: 20.32148 (20.82007) | > loss_duration: 0.95452 (0.94852) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.35351 (29.06690) | > grad_norm_0: 67.34004 (109.61143) | > loss_disc: 2.58566 (2.56678) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58566 (2.56678) | > grad_norm_1: 8.06126 (20.89036) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.18800 (10.39257) | > loader_time: 0.01620 (0.01265)  --> STEP: 101/120 -- GLOBAL_STEP: 5425 | > loss_gen: 2.27100 (2.05191) | > loss_kl: 2.16060 (2.19532) | > loss_feat: 2.60494 (2.99882) | > loss_mel: 20.55115 (20.74045) | > loss_duration: 0.93286 (0.94400) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.52055 (28.93050) | > grad_norm_0: 107.23608 (113.97573) | > loss_disc: 2.68701 (2.57419) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.68701 (2.57419) | > grad_norm_1: 13.82751 (21.26688) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.00090 (10.67675) | > loader_time: 0.01970 (0.01389) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.43325 (2.43325) | > loss_kl: 2.15550 (2.15550) | > loss_feat: 3.25164 (3.25164) | > loss_mel: 21.47021 (21.47021) | > loss_duration: 0.87549 (0.87549) | > loss_0: 30.18608 (30.18608) | > loss_disc: 2.56446 (2.56446) | > loss_1: 2.56446 (2.56446)  --> STEP: 1 | > loss_gen: 2.40737 (2.40737) | > loss_kl: 2.26115 (2.26115) | > loss_feat: 2.77480 (2.77480) | > loss_mel: 20.54231 (20.54231) | > loss_duration: 0.88319 (0.88319) | > loss_0: 28.86882 (28.86882) | > loss_disc: 2.62495 (2.62495) | > loss_1: 2.62495 (2.62495)  --> STEP: 2 | > loss_gen: 2.36845 (2.38791) | > loss_kl: 2.19229 (2.22672) | > loss_feat: 2.91735 (2.84607) | > loss_mel: 20.67932 (20.61082) | > loss_duration: 0.90416 (0.89368) | > loss_0: 29.06158 (28.96520) | > loss_disc: 2.52991 (2.57743) | > loss_1: 2.52991 (2.57743)  --> STEP: 3 | > loss_gen: 2.44896 (2.40826) | > loss_kl: 2.07933 (2.17759) | > loss_feat: 3.31335 (3.00183) | > loss_mel: 19.83821 (20.35328) | > loss_duration: 0.95494 (0.91410) | > loss_0: 28.63480 (28.85507) | > loss_disc: 2.47479 (2.54322) | > loss_1: 2.47479 (2.54322) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00745 (+0.00003) | > avg_loss_gen: 2.40826 (+0.33294) | > avg_loss_kl: 2.17759 (-0.05718) | > avg_loss_feat: 3.00183 (+0.03521) | > avg_loss_mel: 20.35328 (-0.45979) | > avg_loss_duration: 0.91410 (-0.00371) | > avg_loss_0: 28.85507 (-0.15252) | > avg_loss_disc: 2.54322 (-0.00220) | > avg_loss_1: 2.54322 (-0.00220)  > EPOCH: 45/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 06:43:19)   --> STEP: 5/120 -- GLOBAL_STEP: 5450 | > loss_gen: 2.23093 (2.13829) | > loss_kl: 2.68180 (2.53814) | > loss_feat: 3.08579 (3.38721) | > loss_mel: 21.33878 (21.41411) | > loss_duration: 0.94668 (0.90987) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.28397 (30.38763) | > grad_norm_0: 80.52044 (76.33366) | > loss_disc: 2.75881 (2.58026) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.75881 (2.58026) | > grad_norm_1: 24.36528 (27.61233) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.27610 (9.87820) | > loader_time: 0.00930 (0.00752)  --> STEP: 30/120 -- GLOBAL_STEP: 5475 | > loss_gen: 2.05809 (2.05708) | > loss_kl: 2.12927 (2.37041) | > loss_feat: 2.98680 (3.08399) | > loss_mel: 21.34023 (20.90245) | > loss_duration: 0.96391 (0.99198) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.47829 (29.40591) | > grad_norm_0: 125.95527 (68.50254) | > loss_disc: 2.50263 (2.59393) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50263 (2.59393) | > grad_norm_1: 25.50142 (16.09405) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34920 (10.24200) | > loader_time: 0.01180 (0.00966)  --> STEP: 55/120 -- GLOBAL_STEP: 5500 | > loss_gen: 2.03269 (2.03334) | > loss_kl: 2.25910 (2.28770) | > loss_feat: 2.82977 (2.99456) | > loss_mel: 20.88898 (20.76715) | > loss_duration: 0.94219 (0.95382) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.95273 (29.03658) | > grad_norm_0: 62.14025 (84.15934) | > loss_disc: 2.55833 (2.59482) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55833 (2.59482) | > grad_norm_1: 6.57735 (19.68308) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.56370 (10.42969) | > loader_time: 0.03770 (0.01162)  --> STEP: 80/120 -- GLOBAL_STEP: 5525 | > loss_gen: 2.04911 (2.03672) | > loss_kl: 2.11113 (2.24826) | > loss_feat: 3.09936 (2.98336) | > loss_mel: 20.44054 (20.65385) | > loss_duration: 0.93161 (0.94693) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.63173 (28.86913) | > grad_norm_0: 154.22838 (85.08099) | > loss_disc: 2.57128 (2.58582) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57128 (2.58582) | > grad_norm_1: 11.50480 (20.18410) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.58090 (10.59220) | > loader_time: 0.01660 (0.01352)  --> STEP: 105/120 -- GLOBAL_STEP: 5550 | > loss_gen: 1.98820 (2.04052) | > loss_kl: 2.09641 (2.21850) | > loss_feat: 3.32923 (2.98809) | > loss_mel: 20.20509 (20.58879) | > loss_duration: 0.98378 (0.94408) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.60271 (28.77998) | > grad_norm_0: 62.85714 (83.75728) | > loss_disc: 2.39401 (2.57797) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.39401 (2.57797) | > grad_norm_1: 19.40450 (21.02654) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.21890 (10.88872) | > loader_time: 0.02000 (0.01465) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.21473 (2.21473) | > loss_kl: 1.97512 (1.97512) | > loss_feat: 3.21548 (3.21548) | > loss_mel: 21.05054 (21.05054) | > loss_duration: 0.87323 (0.87323) | > loss_0: 29.32909 (29.32909) | > loss_disc: 2.44011 (2.44011) | > loss_1: 2.44011 (2.44011)  --> STEP: 1 | > loss_gen: 2.12472 (2.12472) | > loss_kl: 2.08395 (2.08395) | > loss_feat: 3.17005 (3.17005) | > loss_mel: 20.63851 (20.63851) | > loss_duration: 0.88751 (0.88751) | > loss_0: 28.90474 (28.90474) | > loss_disc: 2.53126 (2.53126) | > loss_1: 2.53126 (2.53126)  --> STEP: 2 | > loss_gen: 2.10283 (2.11377) | > loss_kl: 2.15521 (2.11958) | > loss_feat: 3.16422 (3.16714) | > loss_mel: 20.92728 (20.78289) | > loss_duration: 0.89669 (0.89210) | > loss_0: 29.24623 (29.07548) | > loss_disc: 2.51501 (2.52313) | > loss_1: 2.51501 (2.52313)  --> STEP: 3 | > loss_gen: 1.96230 (2.06328) | > loss_kl: 2.14656 (2.12857) | > loss_feat: 2.72202 (3.01876) | > loss_mel: 19.55897 (20.37492) | > loss_duration: 0.95257 (0.91226) | > loss_0: 27.34242 (28.49780) | > loss_disc: 2.64224 (2.56283) | > loss_1: 2.64224 (2.56283) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00730 (-0.00016) | > avg_loss_gen: 2.06328 (-0.34498) | > avg_loss_kl: 2.12857 (-0.04902) | > avg_loss_feat: 3.01876 (+0.01693) | > avg_loss_mel: 20.37492 (+0.02164) | > avg_loss_duration: 0.91226 (-0.00184) | > avg_loss_0: 28.49780 (-0.35727) | > avg_loss_disc: 2.56283 (+0.01962) | > avg_loss_1: 2.56283 (+0.01962)  > EPOCH: 46/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 07:05:44)   --> STEP: 9/120 -- GLOBAL_STEP: 5575 | > loss_gen: 2.05507 (2.15373) | > loss_kl: 2.45719 (2.42879) | > loss_feat: 3.34411 (3.33273) | > loss_mel: 21.17522 (21.39062) | > loss_duration: 0.93344 (0.91328) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.96505 (30.21915) | > grad_norm_0: 43.83808 (84.90942) | > loss_disc: 2.44377 (2.55077) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.44377 (2.55077) | > grad_norm_1: 17.53319 (21.49320) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.98000 (10.07538) | > loader_time: 0.00850 (0.00785)  --> STEP: 34/120 -- GLOBAL_STEP: 5600 | > loss_gen: 1.85336 (2.06038) | > loss_kl: 2.10778 (2.29593) | > loss_feat: 2.96415 (3.08209) | > loss_mel: 21.08567 (21.00876) | > loss_duration: 0.91919 (0.97828) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.93014 (29.42544) | > grad_norm_0: 50.30187 (80.47618) | > loss_disc: 2.58633 (2.58856) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58633 (2.58856) | > grad_norm_1: 28.25836 (19.31990) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.54410 (10.40815) | > loader_time: 0.01200 (0.00993)  --> STEP: 59/120 -- GLOBAL_STEP: 5625 | > loss_gen: 3.19089 (2.06565) | > loss_kl: 2.03112 (2.24060) | > loss_feat: 2.82048 (3.04595) | > loss_mel: 20.99469 (20.89677) | > loss_duration: 0.92467 (0.95097) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.96186 (29.19994) | > grad_norm_0: 62.28474 (102.70009) | > loss_disc: 3.91342 (2.61409) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 3.91342 (2.61409) | > grad_norm_1: 93.71780 (23.30385) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.22530 (10.59396) | > loader_time: 0.03280 (0.01271)  --> STEP: 84/120 -- GLOBAL_STEP: 5650 | > loss_gen: 3.02805 (2.12152) | > loss_kl: 2.12194 (2.21021) | > loss_feat: 5.72375 (3.25526) | > loss_mel: 22.12748 (21.09825) | > loss_duration: 0.91065 (0.94427) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 33.91186 (29.62951) | > grad_norm_0: 137.02927 (102.03506) | > loss_disc: 1.46430 (2.76841) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 1.46430 (2.76841) | > grad_norm_1: 10.71682 (27.70333) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.71050 (10.77411) | > loader_time: 0.01660 (0.01358)  --> STEP: 109/120 -- GLOBAL_STEP: 5675 | > loss_gen: 2.63164 (2.15858) | > loss_kl: 2.09831 (2.17645) | > loss_feat: 5.12878 (3.44785) | > loss_mel: 20.72969 (21.20639) | > loss_duration: 0.97896 (0.94352) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 31.56738 (29.93280) | > grad_norm_0: 185.21571 (103.22562) | > loss_disc: 1.78067 (2.83187) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 1.78067 (2.83187) | > grad_norm_1: 34.81441 (28.96023) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.97390 (11.05229) | > loader_time: 0.02250 (0.01493) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.91849 (1.91849) | > loss_kl: 2.17192 (2.17192) | > loss_feat: 1.86540 (1.86540) | > loss_mel: 21.95356 (21.95356) | > loss_duration: 0.85877 (0.85877) | > loss_0: 28.76814 (28.76814) | > loss_disc: 3.53804 (3.53804) | > loss_1: 3.53804 (3.53804)  --> STEP: 1 | > loss_gen: 1.96913 (1.96913) | > loss_kl: 2.59869 (2.59869) | > loss_feat: 1.77456 (1.77456) | > loss_mel: 22.57312 (22.57312) | > loss_duration: 0.88400 (0.88400) | > loss_0: 29.79951 (29.79951) | > loss_disc: 3.56906 (3.56906) | > loss_1: 3.56906 (3.56906)  --> STEP: 2 | > loss_gen: 1.85397 (1.91155) | > loss_kl: 2.07004 (2.33437) | > loss_feat: 2.00282 (1.88869) | > loss_mel: 21.22316 (21.89814) | > loss_duration: 0.90184 (0.89292) | > loss_0: 28.05182 (28.92566) | > loss_disc: 3.50822 (3.53864) | > loss_1: 3.50822 (3.53864)  --> STEP: 3 | > loss_gen: 1.94086 (1.92132) | > loss_kl: 2.21197 (2.29357) | > loss_feat: 1.80707 (1.86148) | > loss_mel: 21.37288 (21.72305) | > loss_duration: 0.95360 (0.91315) | > loss_0: 28.28639 (28.71257) | > loss_disc: 3.52469 (3.53399) | > loss_1: 3.52469 (3.53399) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00686 (-0.00044) | > avg_loss_gen: 1.92132 (-0.14196) | > avg_loss_kl: 2.29357 (+0.16500) | > avg_loss_feat: 1.86148 (-1.15728) | > avg_loss_mel: 21.72305 (+1.34813) | > avg_loss_duration: 0.91315 (+0.00089) | > avg_loss_0: 28.71257 (+0.21478) | > avg_loss_disc: 3.53399 (+0.97116) | > avg_loss_1: 3.53399 (+0.97116)  > EPOCH: 47/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 07:28:23)   --> STEP: 13/120 -- GLOBAL_STEP: 5700 | > loss_gen: 1.93476 (1.96702) | > loss_kl: 2.27987 (2.32821) | > loss_feat: 2.44722 (2.55029) | > loss_mel: 21.17637 (22.17525) | > loss_duration: 0.91090 (1.08787) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.74912 (30.10864) | > grad_norm_0: 104.91681 (74.05009) | > loss_disc: 2.48707 (2.62134) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48707 (2.62134) | > grad_norm_1: 3.71033 (7.15022) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.41690 (10.26217) | > loader_time: 0.00960 (0.00841)  --> STEP: 38/120 -- GLOBAL_STEP: 5725 | > loss_gen: 1.63420 (1.75810) | > loss_kl: 2.23361 (2.27115) | > loss_feat: 1.66906 (1.96331) | > loss_mel: 19.76640 (21.24986) | > loss_duration: 0.89388 (0.96933) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.19715 (28.21175) | > grad_norm_0: 111.73109 (72.14781) | > loss_disc: 2.78389 (2.85116) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.78389 (2.85116) | > grad_norm_1: 2.27210 (5.04577) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.92540 (10.52399) | > loader_time: 0.01240 (0.01025)  --> STEP: 63/120 -- GLOBAL_STEP: 5750 | > loss_gen: 1.94118 (1.74973) | > loss_kl: 2.20679 (2.23762) | > loss_feat: 2.03166 (1.88813) | > loss_mel: 20.47841 (20.78731) | > loss_duration: 0.91291 (0.94671) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.57095 (27.60949) | > grad_norm_0: 46.93727 (68.56284) | > loss_disc: 2.71318 (2.81064) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.71318 (2.81064) | > grad_norm_1: 16.69226 (6.41577) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.22070 (10.69207) | > loader_time: 0.01490 (0.01324)  --> STEP: 88/120 -- GLOBAL_STEP: 5775 | > loss_gen: 1.99407 (1.77431) | > loss_kl: 2.17471 (2.20825) | > loss_feat: 2.03079 (1.95997) | > loss_mel: 19.62550 (20.54454) | > loss_duration: 0.94481 (0.94215) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.76988 (27.42922) | > grad_norm_0: 49.49039 (75.58872) | > loss_disc: 2.64055 (2.76617) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.64055 (2.76617) | > grad_norm_1: 19.21723 (8.70310) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.55680 (10.87132) | > loader_time: 0.01740 (0.01406)  --> STEP: 113/120 -- GLOBAL_STEP: 5800 | > loss_gen: 2.00527 (1.80136) | > loss_kl: 2.05196 (2.18474) | > loss_feat: 2.66814 (2.04337) | > loss_mel: 20.69873 (20.44253) | > loss_duration: 0.96177 (0.94370) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.38587 (27.41569) | > grad_norm_0: 62.09514 (78.20623) | > loss_disc: 2.53134 (2.73604) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53134 (2.73604) | > grad_norm_1: 13.61217 (11.24532) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.29230 (11.12687) | > loader_time: 0.02100 (0.01527) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.89800 (1.89800) | > loss_kl: 2.21516 (2.21516) | > loss_feat: 2.43535 (2.43535) | > loss_mel: 20.69648 (20.69648) | > loss_duration: 0.86594 (0.86594) | > loss_0: 28.11093 (28.11093) | > loss_disc: 2.61874 (2.61874) | > loss_1: 2.61874 (2.61874)  --> STEP: 1 | > loss_gen: 1.87252 (1.87252) | > loss_kl: 2.27716 (2.27716) | > loss_feat: 2.42697 (2.42697) | > loss_mel: 20.89118 (20.89118) | > loss_duration: 0.87776 (0.87776) | > loss_0: 28.34560 (28.34560) | > loss_disc: 2.64628 (2.64628) | > loss_1: 2.64628 (2.64628)  --> STEP: 2 | > loss_gen: 1.96227 (1.91739) | > loss_kl: 2.24643 (2.26180) | > loss_feat: 2.33648 (2.38173) | > loss_mel: 20.22647 (20.55882) | > loss_duration: 0.89575 (0.88675) | > loss_0: 27.66740 (28.00650) | > loss_disc: 2.62774 (2.63701) | > loss_1: 2.62774 (2.63701)  --> STEP: 3 | > loss_gen: 1.88111 (1.90530) | > loss_kl: 1.94997 (2.15786) | > loss_feat: 2.43333 (2.39893) | > loss_mel: 19.17903 (20.09889) | > loss_duration: 0.95229 (0.90860) | > loss_0: 26.39573 (27.46957) | > loss_disc: 2.65418 (2.64274) | > loss_1: 2.65418 (2.64274) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00759 (+0.00073) | > avg_loss_gen: 1.90530 (-0.01602) | > avg_loss_kl: 2.15786 (-0.13571) | > avg_loss_feat: 2.39893 (+0.53744) | > avg_loss_mel: 20.09889 (-1.62416) | > avg_loss_duration: 0.90860 (-0.00455) | > avg_loss_0: 27.46957 (-1.24300) | > avg_loss_disc: 2.64274 (-0.89126) | > avg_loss_1: 2.64274 (-0.89126)  > EPOCH: 48/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 07:51:05)   --> STEP: 17/120 -- GLOBAL_STEP: 5825 | > loss_gen: 1.81008 (1.97463) | > loss_kl: 2.21038 (2.37576) | > loss_feat: 2.62830 (2.66265) | > loss_mel: 20.78350 (20.77710) | > loss_duration: 0.93045 (1.04487) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.36271 (28.83501) | > grad_norm_0: 170.67220 (127.03540) | > loss_disc: 2.60094 (2.61596) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60094 (2.61596) | > grad_norm_1: 5.66532 (24.38433) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.21440 (10.24076) | > loader_time: 0.00990 (0.00817)  --> STEP: 42/120 -- GLOBAL_STEP: 5850 | > loss_gen: 1.85624 (1.96026) | > loss_kl: 2.22850 (2.27138) | > loss_feat: 2.63867 (2.63612) | > loss_mel: 20.45103 (20.41727) | > loss_duration: 0.91250 (0.96271) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.08694 (28.24774) | > grad_norm_0: 95.49750 (119.85305) | > loss_disc: 2.54932 (2.59808) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54932 (2.59808) | > grad_norm_1: 17.25261 (17.50796) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.81940 (10.46611) | > loader_time: 0.01300 (0.01023)  --> STEP: 67/120 -- GLOBAL_STEP: 5875 | > loss_gen: 1.89615 (1.96193) | > loss_kl: 2.14143 (2.25070) | > loss_feat: 2.79350 (2.62590) | > loss_mel: 20.23797 (20.32587) | > loss_duration: 0.92039 (0.94501) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.98944 (28.10941) | > grad_norm_0: 141.10785 (96.12573) | > loss_disc: 2.61500 (2.61521) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61500 (2.61521) | > grad_norm_1: 5.86966 (19.13914) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.08740 (10.61984) | > loader_time: 0.01680 (0.01202)  --> STEP: 92/120 -- GLOBAL_STEP: 5900 | > loss_gen: 1.98838 (1.96345) | > loss_kl: 2.34980 (2.22708) | > loss_feat: 2.50193 (2.63167) | > loss_mel: 20.24687 (20.29527) | > loss_duration: 0.92466 (0.94232) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.01164 (28.05978) | > grad_norm_0: 170.55823 (91.35696) | > loss_disc: 2.67810 (2.61524) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.67810 (2.61524) | > grad_norm_1: 8.99969 (19.63811) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.35890 (10.82423) | > loader_time: 0.01780 (0.01352)  --> STEP: 117/120 -- GLOBAL_STEP: 5925 | > loss_gen: 2.00845 (1.96827) | > loss_kl: 2.22392 (2.20950) | > loss_feat: 2.44269 (2.64670) | > loss_mel: 20.26697 (20.28110) | > loss_duration: 1.05335 (0.94616) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.99538 (28.05174) | > grad_norm_0: 129.89082 (98.31581) | > loss_disc: 2.68903 (2.61725) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.68903 (2.61725) | > grad_norm_1: 9.29261 (20.23833) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.39850 (11.10836) | > loader_time: 0.02560 (0.01499) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.03683 (2.03683) | > loss_kl: 2.01998 (2.01998) | > loss_feat: 2.63024 (2.63024) | > loss_mel: 20.95753 (20.95753) | > loss_duration: 0.86334 (0.86334) | > loss_0: 28.50793 (28.50793) | > loss_disc: 2.59688 (2.59688) | > loss_1: 2.59688 (2.59688)  --> STEP: 1 | > loss_gen: 2.04295 (2.04295) | > loss_kl: 2.29004 (2.29004) | > loss_feat: 2.89944 (2.89944) | > loss_mel: 20.46487 (20.46487) | > loss_duration: 0.87902 (0.87902) | > loss_0: 28.57631 (28.57631) | > loss_disc: 2.63329 (2.63329) | > loss_1: 2.63329 (2.63329)  --> STEP: 2 | > loss_gen: 2.04968 (2.04631) | > loss_kl: 2.22668 (2.25836) | > loss_feat: 3.19004 (3.04474) | > loss_mel: 20.58193 (20.52340) | > loss_duration: 0.89570 (0.88736) | > loss_0: 28.94403 (28.76017) | > loss_disc: 2.49512 (2.56420) | > loss_1: 2.49512 (2.56420)  --> STEP: 3 | > loss_gen: 2.11487 (2.06916) | > loss_kl: 2.01651 (2.17774) | > loss_feat: 2.59225 (2.89391) | > loss_mel: 19.48215 (20.17632) | > loss_duration: 0.95448 (0.90973) | > loss_0: 27.16027 (28.22687) | > loss_disc: 2.57146 (2.56662) | > loss_1: 2.57146 (2.56662) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00707 (-0.00052) | > avg_loss_gen: 2.06916 (+0.16387) | > avg_loss_kl: 2.17774 (+0.01989) | > avg_loss_feat: 2.89391 (+0.49498) | > avg_loss_mel: 20.17632 (+0.07743) | > avg_loss_duration: 0.90973 (+0.00114) | > avg_loss_0: 28.22687 (+0.75730) | > avg_loss_disc: 2.56662 (-0.07611) | > avg_loss_1: 2.56662 (-0.07611)  > EPOCH: 49/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 08:13:39)   --> STEP: 21/120 -- GLOBAL_STEP: 5950 | > loss_gen: 1.98426 (2.03237) | > loss_kl: 2.21967 (2.41762) | > loss_feat: 3.29294 (2.94292) | > loss_mel: 20.96151 (20.75009) | > loss_duration: 0.89947 (1.02167) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.35784 (29.16466) | > grad_norm_0: 32.80252 (58.03059) | > loss_disc: 2.45802 (2.59404) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.45802 (2.59404) | > grad_norm_1: 10.73459 (20.75083) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.60910 (10.34300) | > loader_time: 0.01080 (0.00893)  --> STEP: 46/120 -- GLOBAL_STEP: 5975 | > loss_gen: 1.97826 (2.03991) | > loss_kl: 2.17589 (2.31079) | > loss_feat: 2.87396 (2.94298) | > loss_mel: 20.30848 (20.49030) | > loss_duration: 0.93321 (0.95891) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.26981 (28.74288) | > grad_norm_0: 90.77121 (59.56470) | > loss_disc: 2.60075 (2.60375) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60075 (2.60375) | > grad_norm_1: 32.03830 (22.05896) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.75520 (10.55659) | > loader_time: 0.01350 (0.01067)  --> STEP: 71/120 -- GLOBAL_STEP: 6000 | > loss_gen: 1.89178 (2.03839) | > loss_kl: 2.09353 (2.26375) | > loss_feat: 2.88535 (2.96054) | > loss_mel: 20.37187 (20.39336) | > loss_duration: 0.92270 (0.94576) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.16523 (28.60180) | > grad_norm_0: 70.21152 (61.48089) | > loss_disc: 2.76295 (2.58818) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.76295 (2.58818) | > grad_norm_1: 12.77311 (19.27692) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.90860 (10.71343) | > loader_time: 0.01550 (0.01261)  --> STEP: 96/120 -- GLOBAL_STEP: 6025 | > loss_gen: 1.99052 (2.03103) | > loss_kl: 2.11678 (2.22998) | > loss_feat: 2.79306 (2.93186) | > loss_mel: 19.49151 (20.33676) | > loss_duration: 0.91288 (0.94228) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.30476 (28.47193) | > grad_norm_0: 25.29624 (65.12117) | > loss_disc: 2.62072 (2.59483) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.62072 (2.59483) | > grad_norm_1: 43.41383 (19.02149) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.97010 (10.94487) | > loader_time: 0.01880 (0.01382) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.59425 (2.59425) | > loss_kl: 1.92174 (1.92174) | > loss_feat: 2.99736 (2.99736) | > loss_mel: 21.10620 (21.10620) | > loss_duration: 0.85541 (0.85541) | > loss_0: 29.47496 (29.47496) | > loss_disc: 2.68992 (2.68992) | > loss_1: 2.68992 (2.68992)  --> STEP: 1 | > loss_gen: 2.64338 (2.64338) | > loss_kl: 2.41969 (2.41969) | > loss_feat: 3.34432 (3.34432) | > loss_mel: 20.77214 (20.77214) | > loss_duration: 0.87880 (0.87880) | > loss_0: 30.05833 (30.05833) | > loss_disc: 2.66683 (2.66683) | > loss_1: 2.66683 (2.66683)  --> STEP: 2 | > loss_gen: 2.61617 (2.62977) | > loss_kl: 2.11651 (2.26810) | > loss_feat: 2.91547 (3.12989) | > loss_mel: 19.91092 (20.34153) | > loss_duration: 0.89966 (0.88923) | > loss_0: 28.45874 (29.25853) | > loss_disc: 2.76668 (2.71675) | > loss_1: 2.76668 (2.71675)  --> STEP: 3 | > loss_gen: 2.49662 (2.58539) | > loss_kl: 2.25030 (2.26217) | > loss_feat: 3.07372 (3.11117) | > loss_mel: 19.98043 (20.22116) | > loss_duration: 0.95105 (0.90984) | > loss_0: 28.75213 (29.08973) | > loss_disc: 2.93612 (2.78988) | > loss_1: 2.93612 (2.78988) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00740 (+0.00033) | > avg_loss_gen: 2.58539 (+0.51622) | > avg_loss_kl: 2.26217 (+0.08443) | > avg_loss_feat: 3.11117 (+0.21726) | > avg_loss_mel: 20.22116 (+0.04484) | > avg_loss_duration: 0.90984 (+0.00010) | > avg_loss_0: 29.08973 (+0.86286) | > avg_loss_disc: 2.78988 (+0.22326) | > avg_loss_1: 2.78988 (+0.22326)  > EPOCH: 50/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 08:36:22)   --> STEP: 0/120 -- GLOBAL_STEP: 6050 | > loss_gen: 2.58019 (2.58019) | > loss_kl: 2.16887 (2.16887) | > loss_feat: 3.18095 (3.18095) | > loss_mel: 21.32848 (21.32848) | > loss_duration: 0.94626 (0.94626) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 30.20475 (30.20475) | > grad_norm_0: 142.10390 (142.10390) | > loss_disc: 2.66831 (2.66831) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.66831 (2.66831) | > grad_norm_1: 46.76810 (46.76810) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.98890 (10.98895) | > loader_time: 0.66510 (0.66508)  --> STEP: 25/120 -- GLOBAL_STEP: 6075 | > loss_gen: 1.78911 (2.00211) | > loss_kl: 2.35652 (2.35222) | > loss_feat: 3.03414 (2.99077) | > loss_mel: 20.96294 (20.72997) | > loss_duration: 0.94792 (1.00117) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.09063 (29.07625) | > grad_norm_0: 45.00413 (56.37832) | > loss_disc: 2.58078 (2.62002) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.58078 (2.62002) | > grad_norm_1: 20.94589 (22.00776) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34950 (10.41559) | > loader_time: 0.01100 (0.00929)  --> STEP: 50/120 -- GLOBAL_STEP: 6100 | > loss_gen: 1.79166 (2.01674) | > loss_kl: 2.14043 (2.27080) | > loss_feat: 2.92220 (2.98316) | > loss_mel: 20.11810 (20.42192) | > loss_duration: 0.90684 (0.95154) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.87923 (28.64416) | > grad_norm_0: 163.86739 (96.11914) | > loss_disc: 2.56948 (2.58704) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.56948 (2.58704) | > grad_norm_1: 8.21288 (21.93045) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.72070 (10.61852) | > loader_time: 0.01370 (0.01095)  --> STEP: 75/120 -- GLOBAL_STEP: 6125 | > loss_gen: 1.98896 (2.02949) | > loss_kl: 2.27275 (2.23778) | > loss_feat: 3.13563 (3.01706) | > loss_mel: 20.61513 (20.37305) | > loss_duration: 0.93976 (0.94439) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.95222 (28.60178) | > grad_norm_0: 158.67184 (104.58746) | > loss_disc: 2.48225 (2.57552) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.48225 (2.57552) | > grad_norm_1: 12.46307 (20.56519) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.69520 (10.76168) | > loader_time: 0.01600 (0.01357)  --> STEP: 100/120 -- GLOBAL_STEP: 6150 | > loss_gen: 2.15594 (2.03260) | > loss_kl: 2.28788 (2.21720) | > loss_feat: 2.81643 (3.01408) | > loss_mel: 19.59532 (20.31467) | > loss_duration: 0.91644 (0.94106) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.77201 (28.51962) | > grad_norm_0: 62.10205 (102.37877) | > loss_disc: 2.57987 (2.57616) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.57987 (2.57616) | > grad_norm_1: 29.33695 (20.27089) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.93520 (11.00490) | > loader_time: 0.01980 (0.01474) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.36502 (2.36502) | > loss_kl: 2.00180 (2.00180) | > loss_feat: 2.94265 (2.94265) | > loss_mel: 20.66209 (20.66209) | > loss_duration: 0.86798 (0.86798) | > loss_0: 28.83953 (28.83953) | > loss_disc: 2.58308 (2.58308) | > loss_1: 2.58308 (2.58308)  --> STEP: 1 | > loss_gen: 2.30899 (2.30899) | > loss_kl: 2.09909 (2.09909) | > loss_feat: 2.99496 (2.99496) | > loss_mel: 21.76817 (21.76817) | > loss_duration: 0.88249 (0.88249) | > loss_0: 30.05371 (30.05371) | > loss_disc: 2.50887 (2.50887) | > loss_1: 2.50887 (2.50887)  --> STEP: 2 | > loss_gen: 2.30623 (2.30761) | > loss_kl: 2.20136 (2.15022) | > loss_feat: 3.05434 (3.02465) | > loss_mel: 20.57004 (21.16911) | > loss_duration: 0.89785 (0.89017) | > loss_0: 29.02981 (29.54176) | > loss_disc: 2.63955 (2.57421) | > loss_1: 2.63955 (2.57421)  --> STEP: 3 | > loss_gen: 2.29700 (2.30408) | > loss_kl: 2.29499 (2.19848) | > loss_feat: 2.70680 (2.91870) | > loss_mel: 19.14784 (20.49535) | > loss_duration: 0.96230 (0.91421) | > loss_0: 27.40894 (28.83082) | > loss_disc: 2.79463 (2.64768) | > loss_1: 2.79463 (2.64768) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00726 (-0.00014) | > avg_loss_gen: 2.30408 (-0.28131) | > avg_loss_kl: 2.19848 (-0.06369) | > avg_loss_feat: 2.91870 (-0.19247) | > avg_loss_mel: 20.49535 (+0.27419) | > avg_loss_duration: 0.91421 (+0.00438) | > avg_loss_0: 28.83082 (-0.25891) | > avg_loss_disc: 2.64768 (-0.14220) | > avg_loss_1: 2.64768 (-0.14220)  > EPOCH: 51/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 08:59:07)   --> STEP: 4/120 -- GLOBAL_STEP: 6175 | > loss_gen: 2.42264 (2.14168) | > loss_kl: 2.72892 (2.40279) | > loss_feat: 3.53551 (3.43429) | > loss_mel: 21.53937 (21.06685) | > loss_duration: 0.90478 (0.90730) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 31.13122 (29.95292) | > grad_norm_0: 137.18976 (110.17264) | > loss_disc: 2.47837 (2.47489) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.47837 (2.47489) | > grad_norm_1: 40.86840 (22.39586) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.14150 (9.99311) | > loader_time: 0.01140 (0.00707)  --> STEP: 29/120 -- GLOBAL_STEP: 6200 | > loss_gen: 1.88474 (2.02251) | > loss_kl: 2.16817 (2.33188) | > loss_feat: 2.90601 (3.04791) | > loss_mel: 20.07162 (20.65681) | > loss_duration: 0.91966 (0.98468) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.95021 (29.04379) | > grad_norm_0: 127.88791 (109.88725) | > loss_disc: 2.53066 (2.59116) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.53066 (2.59116) | > grad_norm_1: 16.22680 (21.78595) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.65510 (10.47171) | > loader_time: 0.01150 (0.00941)  --> STEP: 54/120 -- GLOBAL_STEP: 6225 | > loss_gen: 1.87473 (2.02728) | > loss_kl: 2.25117 (2.27727) | > loss_feat: 3.14573 (3.03709) | > loss_mel: 20.08147 (20.39953) | > loss_duration: 0.93840 (0.94921) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.29150 (28.69038) | > grad_norm_0: 121.69086 (105.60651) | > loss_disc: 2.55684 (2.58710) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.55684 (2.58710) | > grad_norm_1: 22.57614 (21.03142) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.88650 (10.63410) | > loader_time: 0.01340 (0.01106)  --> STEP: 79/120 -- GLOBAL_STEP: 6250 | > loss_gen: 1.87558 (2.04009) | > loss_kl: 2.13066 (2.24542) | > loss_feat: 3.28234 (3.05709) | > loss_mel: 20.39569 (20.33486) | > loss_duration: 0.93201 (0.94278) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.61628 (28.62024) | > grad_norm_0: 53.54686 (101.04710) | > loss_disc: 2.52098 (2.56999) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.52098 (2.56999) | > grad_norm_1: 44.72263 (21.32152) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.97150 (10.75185) | > loader_time: 0.01650 (0.01322)  --> STEP: 104/120 -- GLOBAL_STEP: 6275 | > loss_gen: 1.87505 (2.04109) | > loss_kl: 2.11995 (2.22758) | > loss_feat: 3.11928 (3.03638) | > loss_mel: 19.45981 (20.28597) | > loss_duration: 0.92837 (0.93978) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.50245 (28.53080) | > grad_norm_0: 23.90299 (97.32449) | > loss_disc: 2.66316 (2.57278) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.66316 (2.57278) | > grad_norm_1: 8.01490 (20.13411) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.90110 (10.99962) | > loader_time: 0.02000 (0.01447) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.80094 (1.80094) | > loss_kl: 2.20798 (2.20798) | > loss_feat: 2.76510 (2.76510) | > loss_mel: 20.21762 (20.21762) | > loss_duration: 0.86775 (0.86775) | > loss_0: 27.85938 (27.85938) | > loss_disc: 2.57022 (2.57022) | > loss_1: 2.57022 (2.57022)  --> STEP: 1 | > loss_gen: 1.73529 (1.73529) | > loss_kl: 2.14536 (2.14536) | > loss_feat: 2.76128 (2.76128) | > loss_mel: 20.30469 (20.30469) | > loss_duration: 0.89937 (0.89937) | > loss_0: 27.84599 (27.84599) | > loss_disc: 2.65850 (2.65850) | > loss_1: 2.65850 (2.65850)  --> STEP: 2 | > loss_gen: 1.83460 (1.78494) | > loss_kl: 2.59040 (2.36788) | > loss_feat: 2.52675 (2.64402) | > loss_mel: 19.78349 (20.04409) | > loss_duration: 0.91059 (0.90498) | > loss_0: 27.64583 (27.74591) | > loss_disc: 2.59277 (2.62564) | > loss_1: 2.59277 (2.62564)  --> STEP: 3 | > loss_gen: 1.72069 (1.76353) | > loss_kl: 2.37108 (2.36895) | > loss_feat: 2.57909 (2.62238) | > loss_mel: 19.80025 (19.96281) | > loss_duration: 0.95423 (0.92140) | > loss_0: 27.42535 (27.63906) | > loss_disc: 2.68334 (2.64487) | > loss_1: 2.68334 (2.64487) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00733 (+0.00008) | > avg_loss_gen: 1.76353 (-0.54055) | > avg_loss_kl: 2.36895 (+0.17047) | > avg_loss_feat: 2.62238 (-0.29633) | > avg_loss_mel: 19.96281 (-0.53254) | > avg_loss_duration: 0.92140 (+0.00718) | > avg_loss_0: 27.63906 (-1.19176) | > avg_loss_disc: 2.64487 (-0.00281) | > avg_loss_1: 2.64487 (-0.00281)  > EPOCH: 52/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 09:21:44)   --> STEP: 8/120 -- GLOBAL_STEP: 6300 | > loss_gen: 2.34960 (2.09415) | > loss_kl: 2.37503 (2.31508) | > loss_feat: 3.08079 (3.18496) | > loss_mel: 20.78739 (20.71985) | > loss_duration: 0.88717 (0.90720) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.47999 (29.22124) | > grad_norm_0: 49.85612 (74.27252) | > loss_disc: 2.56538 (2.53664) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.56538 (2.53664) | > grad_norm_1: 16.91527 (16.63032) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.37570 (10.05411) | > loader_time: 0.00840 (0.00730)  --> STEP: 33/120 -- GLOBAL_STEP: 6325 | > loss_gen: 1.72397 (2.04858) | > loss_kl: 2.23191 (2.31420) | > loss_feat: 2.83828 (3.12089) | > loss_mel: 19.67764 (20.45298) | > loss_duration: 0.91286 (0.97746) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.38467 (28.91411) | > grad_norm_0: 90.70211 (62.71064) | > loss_disc: 2.72925 (2.56231) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.72925 (2.56231) | > grad_norm_1: 53.51641 (19.49788) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.55230 (10.36225) | > loader_time: 0.01180 (0.00973)  --> STEP: 58/120 -- GLOBAL_STEP: 6350 | > loss_gen: 2.39827 (2.04961) | > loss_kl: 2.37931 (2.27377) | > loss_feat: 3.07511 (3.07182) | > loss_mel: 20.04039 (20.31192) | > loss_duration: 0.89633 (0.94875) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.78942 (28.65587) | > grad_norm_0: 111.87057 (83.51984) | > loss_disc: 2.56009 (2.56492) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.56009 (2.56492) | > grad_norm_1: 19.76066 (19.66464) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.80370 (10.54295) | > loader_time: 0.01410 (0.01234)  --> STEP: 83/120 -- GLOBAL_STEP: 6375 | > loss_gen: 2.01844 (2.04940) | > loss_kl: 2.22200 (2.24317) | > loss_feat: 2.75645 (3.07144) | > loss_mel: 19.76868 (20.21252) | > loss_duration: 0.90902 (0.94298) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.67458 (28.51951) | > grad_norm_0: 101.57902 (84.91004) | > loss_disc: 2.60213 (2.55846) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60213 (2.55846) | > grad_norm_1: 19.72031 (18.43946) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.49620 (10.72295) | > loader_time: 0.01660 (0.01354)  --> STEP: 108/120 -- GLOBAL_STEP: 6400 | > loss_gen: 1.95196 (2.04783) | > loss_kl: 1.97716 (2.22642) | > loss_feat: 3.14093 (3.06358) | > loss_mel: 20.00607 (20.16791) | > loss_duration: 0.98412 (0.94224) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.06026 (28.44798) | > grad_norm_0: 46.95544 (86.54449) | > loss_disc: 2.44137 (2.55937) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.44137 (2.55937) | > grad_norm_1: 7.04355 (19.83792) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.24800 (10.98857) | > loader_time: 0.02020 (0.01479) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.86800 (1.86800) | > loss_kl: 2.65136 (2.65136) | > loss_feat: 3.03910 (3.03910) | > loss_mel: 21.68022 (21.68022) | > loss_duration: 0.86348 (0.86348) | > loss_0: 30.10216 (30.10216) | > loss_disc: 2.71023 (2.71023) | > loss_1: 2.71023 (2.71023)  --> STEP: 1 | > loss_gen: 1.96463 (1.96463) | > loss_kl: 2.42651 (2.42651) | > loss_feat: 3.21238 (3.21238) | > loss_mel: 21.13787 (21.13787) | > loss_duration: 0.89211 (0.89211) | > loss_0: 29.63350 (29.63350) | > loss_disc: 2.57052 (2.57052) | > loss_1: 2.57052 (2.57052)  --> STEP: 2 | > loss_gen: 1.91383 (1.93923) | > loss_kl: 2.25529 (2.34090) | > loss_feat: 2.87313 (3.04275) | > loss_mel: 20.37177 (20.75482) | > loss_duration: 0.90002 (0.89606) | > loss_0: 28.31404 (28.97377) | > loss_disc: 2.64447 (2.60750) | > loss_1: 2.64447 (2.60750)  --> STEP: 3 | > loss_gen: 1.71829 (1.86558) | > loss_kl: 2.27120 (2.31767) | > loss_feat: 2.71630 (2.93393) | > loss_mel: 19.36539 (20.29168) | > loss_duration: 0.95617 (0.91610) | > loss_0: 27.02735 (28.32496) | > loss_disc: 2.83601 (2.68367) | > loss_1: 2.83601 (2.68367) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00667 (-0.00067) | > avg_loss_gen: 1.86558 (+0.10206) | > avg_loss_kl: 2.31767 (-0.05128) | > avg_loss_feat: 2.93393 (+0.31156) | > avg_loss_mel: 20.29168 (+0.32887) | > avg_loss_duration: 0.91610 (-0.00530) | > avg_loss_0: 28.32496 (+0.68591) | > avg_loss_disc: 2.68367 (+0.03880) | > avg_loss_1: 2.68367 (+0.03880)  > EPOCH: 53/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 09:44:16)   --> STEP: 12/120 -- GLOBAL_STEP: 6425 | > loss_gen: 2.12484 (2.05249) | > loss_kl: 2.42680 (2.40371) | > loss_feat: 2.89526 (3.14260) | > loss_mel: 20.47445 (20.89396) | > loss_duration: 0.92895 (0.90888) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.85031 (29.40163) | > grad_norm_0: 200.41772 (131.48929) | > loss_disc: 2.59876 (2.64219) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.59876 (2.64219) | > grad_norm_1: 10.73329 (25.24079) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.26260 (10.13272) | > loader_time: 0.00940 (0.00799)  --> STEP: 37/120 -- GLOBAL_STEP: 6450 | > loss_gen: 1.83156 (2.03841) | > loss_kl: 2.19271 (2.35220) | > loss_feat: 3.07237 (3.07350) | > loss_mel: 20.06409 (20.51960) | > loss_duration: 0.88465 (0.96833) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.04538 (28.95204) | > grad_norm_0: 94.21903 (97.70997) | > loss_disc: 2.67589 (2.60424) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.67589 (2.60424) | > grad_norm_1: 11.02849 (20.87062) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.48510 (10.41040) | > loader_time: 0.01240 (0.01006)  --> STEP: 62/120 -- GLOBAL_STEP: 6475 | > loss_gen: 1.99729 (2.02933) | > loss_kl: 2.12908 (2.28308) | > loss_feat: 3.04205 (3.02832) | > loss_mel: 19.89434 (20.32292) | > loss_duration: 0.93321 (0.94630) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.99597 (28.60995) | > grad_norm_0: 28.40576 (86.21778) | > loss_disc: 2.60469 (2.59528) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60469 (2.59528) | > grad_norm_1: 24.82503 (18.69652) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.02420 (10.58024) | > loader_time: 0.01500 (0.01332)  --> STEP: 87/120 -- GLOBAL_STEP: 6500 | > loss_gen: 1.98647 (2.02588) | > loss_kl: 2.21922 (2.24892) | > loss_feat: 3.03019 (2.99853) | > loss_mel: 20.20904 (20.26161) | > loss_duration: 0.93880 (0.94113) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.38372 (28.47608) | > grad_norm_0: 98.92358 (82.94826) | > loss_disc: 2.53891 (2.59865) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.53891 (2.59865) | > grad_norm_1: 9.88275 (20.08733) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.51500 (10.77172) | > loader_time: 0.01930 (0.01419)  --> STEP: 112/120 -- GLOBAL_STEP: 6525 | > loss_gen: 1.86804 (2.02513) | > loss_kl: 2.09038 (2.22782) | > loss_feat: 2.65784 (2.99971) | > loss_mel: 20.15202 (20.19312) | > loss_duration: 0.96884 (0.94199) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.73712 (28.38777) | > grad_norm_0: 88.92262 (85.60939) | > loss_disc: 2.70056 (2.59340) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.70056 (2.59340) | > grad_norm_1: 33.90837 (18.63575) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.45490 (11.04313) | > loader_time: 0.02350 (0.01542) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.34199 (2.34199) | > loss_kl: 2.27151 (2.27151) | > loss_feat: 3.04709 (3.04709) | > loss_mel: 19.60365 (19.60365) | > loss_duration: 0.87629 (0.87629) | > loss_0: 28.14052 (28.14052) | > loss_disc: 2.56830 (2.56830) | > loss_1: 2.56830 (2.56830)  --> STEP: 1 | > loss_gen: 2.40182 (2.40182) | > loss_kl: 2.24228 (2.24228) | > loss_feat: 2.66654 (2.66654) | > loss_mel: 19.74883 (19.74883) | > loss_duration: 0.87972 (0.87972) | > loss_0: 27.93919 (27.93919) | > loss_disc: 2.75225 (2.75225) | > loss_1: 2.75225 (2.75225)  --> STEP: 2 | > loss_gen: 2.22987 (2.31585) | > loss_kl: 2.17783 (2.21005) | > loss_feat: 2.88276 (2.77465) | > loss_mel: 20.01178 (19.88030) | > loss_duration: 0.89948 (0.88960) | > loss_0: 28.20171 (28.07045) | > loss_disc: 2.59051 (2.67138) | > loss_1: 2.59051 (2.67138)  --> STEP: 3 | > loss_gen: 2.11917 (2.25029) | > loss_kl: 2.14328 (2.18780) | > loss_feat: 2.67448 (2.74126) | > loss_mel: 19.46792 (19.74284) | > loss_duration: 0.94578 (0.90832) | > loss_0: 27.35063 (27.83051) | > loss_disc: 2.81460 (2.71912) | > loss_1: 2.81460 (2.71912) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00664 (-0.00003) | > avg_loss_gen: 2.25029 (+0.38470) | > avg_loss_kl: 2.18780 (-0.12987) | > avg_loss_feat: 2.74126 (-0.19267) | > avg_loss_mel: 19.74284 (-0.54883) | > avg_loss_duration: 0.90832 (-0.00778) | > avg_loss_0: 27.83051 (-0.49445) | > avg_loss_disc: 2.71912 (+0.03545) | > avg_loss_1: 2.71912 (+0.03545)  > EPOCH: 54/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 10:06:50)   --> STEP: 16/120 -- GLOBAL_STEP: 6550 | > loss_gen: 1.76441 (2.02130) | > loss_kl: 2.37413 (2.36356) | > loss_feat: 2.90556 (3.11651) | > loss_mel: 20.36392 (20.76962) | > loss_duration: 0.88421 (1.04574) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.29224 (29.31673) | > grad_norm_0: 33.82477 (81.61288) | > loss_disc: 2.68467 (2.63772) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.68467 (2.63772) | > grad_norm_1: 14.09547 (20.33532) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.28310 (10.22205) | > loader_time: 0.01020 (0.00855)  --> STEP: 41/120 -- GLOBAL_STEP: 6575 | > loss_gen: 2.18686 (2.02290) | > loss_kl: 2.11285 (2.28413) | > loss_feat: 2.97967 (3.01607) | > loss_mel: 19.74449 (20.36491) | > loss_duration: 0.92417 (0.96123) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.94805 (28.64923) | > grad_norm_0: 147.92670 (76.71490) | > loss_disc: 2.65284 (2.61808) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.65284 (2.61808) | > grad_norm_1: 26.17992 (20.69540) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.81610 (10.41903) | > loader_time: 0.01320 (0.01042)  --> STEP: 66/120 -- GLOBAL_STEP: 6600 | > loss_gen: 2.27093 (2.03566) | > loss_kl: 2.28475 (2.25487) | > loss_feat: 3.15277 (3.03784) | > loss_mel: 20.51427 (20.25901) | > loss_duration: 0.94918 (0.94403) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 29.17189 (28.53140) | > grad_norm_0: 47.27859 (84.40674) | > loss_disc: 2.46925 (2.60417) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.46925 (2.60417) | > grad_norm_1: 17.49412 (21.03170) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93570 (10.58380) | > loader_time: 0.01520 (0.01220)  --> STEP: 91/120 -- GLOBAL_STEP: 6625 | > loss_gen: 1.96928 (2.03259) | > loss_kl: 2.06154 (2.22487) | > loss_feat: 3.09470 (3.03508) | > loss_mel: 20.14841 (20.16602) | > loss_duration: 0.92621 (0.93984) | > amp_scaler: 512.00000 (883.34066) | > loss_0: 28.20014 (28.39840) | > grad_norm_0: 140.89528 (90.00290) | > loss_disc: 2.56192 (2.58959) | > amp_scaler-1: 512.00000 (883.34066) | > loss_1: 2.56192 (2.58959) | > grad_norm_1: 66.35757 (19.73392) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.43950 (10.80018) | > loader_time: 0.01810 (0.01337)  --> STEP: 116/120 -- GLOBAL_STEP: 6650 | > loss_gen: 2.24395 (2.03520) | > loss_kl: 1.99613 (2.20685) | > loss_feat: 2.66203 (3.03233) | > loss_mel: 19.56051 (20.09681) | > loss_duration: 1.05473 (0.94313) | > amp_scaler: 512.00000 (803.31034) | > loss_0: 27.51735 (28.31432) | > grad_norm_0: 49.23236 (86.53467) | > loss_disc: 2.68086 (2.58602) | > amp_scaler-1: 512.00000 (803.31034) | > loss_1: 2.68086 (2.58602) | > grad_norm_1: 14.47461 (20.45356) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.20070 (11.07677) | > loader_time: 0.02260 (0.01482) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.06022 (2.06022) | > loss_kl: 2.11399 (2.11399) | > loss_feat: 3.59931 (3.59931) | > loss_mel: 20.95951 (20.95951) | > loss_duration: 0.87204 (0.87204) | > loss_0: 29.60507 (29.60507) | > loss_disc: 2.46466 (2.46466) | > loss_1: 2.46466 (2.46466)  --> STEP: 1 | > loss_gen: 2.10232 (2.10232) | > loss_kl: 2.29433 (2.29433) | > loss_feat: 3.06241 (3.06241) | > loss_mel: 20.34381 (20.34381) | > loss_duration: 0.88569 (0.88569) | > loss_0: 28.68856 (28.68856) | > loss_disc: 2.52952 (2.52952) | > loss_1: 2.52952 (2.52952)  --> STEP: 2 | > loss_gen: 2.08590 (2.09411) | > loss_kl: 1.95793 (2.12613) | > loss_feat: 3.62722 (3.34481) | > loss_mel: 20.38368 (20.36375) | > loss_duration: 0.91141 (0.89855) | > loss_0: 28.96614 (28.82735) | > loss_disc: 2.48885 (2.50918) | > loss_1: 2.48885 (2.50918)  --> STEP: 3 | > loss_gen: 1.84015 (2.00946) | > loss_kl: 2.18464 (2.14563) | > loss_feat: 3.30243 (3.33069) | > loss_mel: 19.64481 (20.12410) | > loss_duration: 0.96166 (0.91959) | > loss_0: 27.93370 (28.52947) | > loss_disc: 2.73347 (2.58395) | > loss_1: 2.73347 (2.58395) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00661 (-0.00003) | > avg_loss_gen: 2.00946 (-0.24083) | > avg_loss_kl: 2.14563 (-0.04216) | > avg_loss_feat: 3.33069 (+0.58943) | > avg_loss_mel: 20.12410 (+0.38126) | > avg_loss_duration: 0.91959 (+0.01126) | > avg_loss_0: 28.52947 (+0.69896) | > avg_loss_disc: 2.58395 (-0.13517) | > avg_loss_1: 2.58395 (-0.13517)  > EPOCH: 55/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 10:29:24)   --> STEP: 20/120 -- GLOBAL_STEP: 6675 | > loss_gen: 2.09298 (2.06176) | > loss_kl: 2.34664 (2.41305) | > loss_feat: 3.18568 (3.19805) | > loss_mel: 19.69467 (20.60839) | > loss_duration: 0.91624 (1.02335) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.23621 (29.30461) | > grad_norm_0: 126.53609 (99.15872) | > loss_disc: 2.47606 (2.56584) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.47606 (2.56584) | > grad_norm_1: 19.69524 (19.92287) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.27840 (10.26978) | > loader_time: 0.01040 (0.00860)  --> STEP: 45/120 -- GLOBAL_STEP: 6700 | > loss_gen: 1.68743 (2.05340) | > loss_kl: 2.35191 (2.31051) | > loss_feat: 2.90696 (3.14935) | > loss_mel: 20.34553 (20.35970) | > loss_duration: 0.92554 (0.95930) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.21737 (28.83225) | > grad_norm_0: 34.16745 (91.06710) | > loss_disc: 2.72615 (2.58630) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.72615 (2.58630) | > grad_norm_1: 30.31624 (20.70221) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.53730 (10.43649) | > loader_time: 0.01290 (0.01040)  --> STEP: 70/120 -- GLOBAL_STEP: 6725 | > loss_gen: 2.05022 (2.04102) | > loss_kl: 2.15785 (2.26752) | > loss_feat: 2.76638 (3.08220) | > loss_mel: 19.66408 (20.21239) | > loss_duration: 0.95196 (0.94553) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.59049 (28.54865) | > grad_norm_0: 61.94176 (83.53418) | > loss_disc: 2.68606 (2.59036) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.68606 (2.59036) | > grad_norm_1: 12.48325 (19.61600) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.00880 (10.56082) | > loader_time: 0.01580 (0.01273)  --> STEP: 95/120 -- GLOBAL_STEP: 6750 | > loss_gen: 2.20977 (2.04444) | > loss_kl: 2.17858 (2.23783) | > loss_feat: 3.15611 (3.07927) | > loss_mel: 19.25420 (20.14925) | > loss_duration: 0.93923 (0.94197) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.73790 (28.45275) | > grad_norm_0: 33.94248 (79.98415) | > loss_disc: 2.46001 (2.57791) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.46001 (2.57791) | > grad_norm_1: 21.68911 (19.87497) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.20920 (10.77832) | > loader_time: 0.01870 (0.01382)  --> STEP: 120/120 -- GLOBAL_STEP: 6775 | > loss_gen: 2.05821 (2.04280) | > loss_kl: 2.18590 (2.22407) | > loss_feat: 3.14076 (3.06580) | > loss_mel: 20.71259 (20.11196) | > loss_duration: 1.14698 (0.94934) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.24444 (28.39397) | > grad_norm_0: 53.29023 (82.52594) | > loss_disc: 2.50542 (2.57910) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50542 (2.57910) | > grad_norm_1: 13.11626 (20.75381) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.35900 (11.05823) | > loader_time: 0.02460 (0.01533) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.07934 (2.07934) | > loss_kl: 2.32397 (2.32397) | > loss_feat: 3.37319 (3.37319) | > loss_mel: 20.59955 (20.59955) | > loss_duration: 0.86209 (0.86209) | > loss_0: 29.23814 (29.23814) | > loss_disc: 2.46577 (2.46577) | > loss_1: 2.46577 (2.46577)  --> STEP: 1 | > loss_gen: 1.99361 (1.99361) | > loss_kl: 2.57307 (2.57307) | > loss_feat: 3.48558 (3.48558) | > loss_mel: 20.97704 (20.97704) | > loss_duration: 0.88458 (0.88458) | > loss_0: 29.91387 (29.91387) | > loss_disc: 2.52452 (2.52452) | > loss_1: 2.52452 (2.52452)  --> STEP: 2 | > loss_gen: 2.10170 (2.04765) | > loss_kl: 2.18069 (2.37688) | > loss_feat: 2.98008 (3.23283) | > loss_mel: 19.55287 (20.26495) | > loss_duration: 0.89899 (0.89179) | > loss_0: 27.71433 (28.81410) | > loss_disc: 2.50422 (2.51437) | > loss_1: 2.50422 (2.51437)  --> STEP: 3 | > loss_gen: 2.02568 (2.04033) | > loss_kl: 2.34886 (2.36754) | > loss_feat: 2.98123 (3.14896) | > loss_mel: 19.57544 (20.03512) | > loss_duration: 0.94764 (0.91041) | > loss_0: 27.87885 (28.50235) | > loss_disc: 2.62223 (2.55032) | > loss_1: 2.62223 (2.55032) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00735 (+0.00074) | > avg_loss_gen: 2.04033 (+0.03087) | > avg_loss_kl: 2.36754 (+0.22191) | > avg_loss_feat: 3.14896 (-0.18173) | > avg_loss_mel: 20.03512 (-0.08898) | > avg_loss_duration: 0.91041 (-0.00918) | > avg_loss_0: 28.50235 (-0.02712) | > avg_loss_disc: 2.55032 (-0.03362) | > avg_loss_1: 2.55032 (-0.03362)  > EPOCH: 56/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 10:51:48)   --> STEP: 24/120 -- GLOBAL_STEP: 6800 | > loss_gen: 2.02975 (2.10126) | > loss_kl: 2.17053 (2.36967) | > loss_feat: 2.95000 (3.32410) | > loss_mel: 19.76535 (20.35580) | > loss_duration: 0.93255 (0.99982) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.84817 (29.15065) | > grad_norm_0: 30.21329 (86.76546) | > loss_disc: 2.61529 (2.55341) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61529 (2.55341) | > grad_norm_1: 8.84058 (16.54723) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.34980 (10.17905) | > loader_time: 0.01080 (0.00906)  --> STEP: 49/120 -- GLOBAL_STEP: 6825 | > loss_gen: 2.32057 (2.08754) | > loss_kl: 2.15277 (2.29995) | > loss_feat: 2.75758 (3.22630) | > loss_mel: 19.73967 (20.22885) | > loss_duration: 0.92679 (0.95136) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.89738 (28.79400) | > grad_norm_0: 71.76318 (81.23251) | > loss_disc: 2.66027 (2.55692) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.66027 (2.55692) | > grad_norm_1: 46.93244 (20.01105) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.91390 (10.37982) | > loader_time: 0.01350 (0.01077)  --> STEP: 74/120 -- GLOBAL_STEP: 6850 | > loss_gen: 1.94463 (2.06488) | > loss_kl: 2.30955 (2.26304) | > loss_feat: 3.12094 (3.18180) | > loss_mel: 20.07502 (20.14198) | > loss_duration: 0.90248 (0.94280) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.35261 (28.59450) | > grad_norm_0: 25.47296 (88.35401) | > loss_disc: 2.65945 (2.55908) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65945 (2.55908) | > grad_norm_1: 39.57358 (19.97301) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.95400 (10.52428) | > loader_time: 0.01590 (0.01240)  --> STEP: 99/120 -- GLOBAL_STEP: 6875 | > loss_gen: 2.12605 (2.05978) | > loss_kl: 2.16318 (2.23902) | > loss_feat: 3.20353 (3.15084) | > loss_mel: 19.33868 (20.05660) | > loss_duration: 0.91075 (0.93984) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.74219 (28.44607) | > grad_norm_0: 29.25509 (82.12009) | > loss_disc: 2.44091 (2.56175) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.44091 (2.56175) | > grad_norm_1: 7.18091 (20.47180) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.18280 (10.78683) | > loader_time: 0.01910 (0.01380) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.95908 (1.95908) | > loss_kl: 2.15282 (2.15282) | > loss_feat: 3.05637 (3.05637) | > loss_mel: 20.05957 (20.05957) | > loss_duration: 0.87195 (0.87195) | > loss_0: 28.09978 (28.09978) | > loss_disc: 2.55026 (2.55026) | > loss_1: 2.55026 (2.55026)  --> STEP: 1 | > loss_gen: 2.02839 (2.02839) | > loss_kl: 2.31088 (2.31088) | > loss_feat: 3.10453 (3.10453) | > loss_mel: 20.33830 (20.33830) | > loss_duration: 0.88912 (0.88912) | > loss_0: 28.67121 (28.67121) | > loss_disc: 2.51293 (2.51293) | > loss_1: 2.51293 (2.51293)  --> STEP: 2 | > loss_gen: 1.94017 (1.98428) | > loss_kl: 2.17904 (2.24496) | > loss_feat: 3.39344 (3.24898) | > loss_mel: 20.23270 (20.28550) | > loss_duration: 0.89804 (0.89358) | > loss_0: 28.64338 (28.65730) | > loss_disc: 2.43309 (2.47301) | > loss_1: 2.43309 (2.47301)  --> STEP: 3 | > loss_gen: 1.92524 (1.96460) | > loss_kl: 2.07183 (2.18725) | > loss_feat: 3.22145 (3.23981) | > loss_mel: 19.65738 (20.07613) | > loss_duration: 0.95155 (0.91290) | > loss_0: 27.82746 (28.38069) | > loss_disc: 2.52961 (2.49188) | > loss_1: 2.52961 (2.49188) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00749 (+0.00015) | > avg_loss_gen: 1.96460 (-0.07573) | > avg_loss_kl: 2.18725 (-0.18029) | > avg_loss_feat: 3.23981 (+0.09085) | > avg_loss_mel: 20.07613 (+0.04101) | > avg_loss_duration: 0.91290 (+0.00250) | > avg_loss_0: 28.38069 (-0.12167) | > avg_loss_disc: 2.49188 (-0.05845) | > avg_loss_1: 2.49188 (-0.05845)  > EPOCH: 57/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 11:14:08)   --> STEP: 3/120 -- GLOBAL_STEP: 6900 | > loss_gen: 1.87686 (2.17558) | > loss_kl: 2.36056 (2.37946) | > loss_feat: 3.56543 (3.49958) | > loss_mel: 20.84512 (20.63154) | > loss_duration: 0.86886 (0.88324) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.51683 (29.56940) | > grad_norm_0: 54.60588 (97.88779) | > loss_disc: 2.55290 (2.50384) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55290 (2.50384) | > grad_norm_1: 40.66891 (30.17815) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.84830 (9.73747) | > loader_time: 0.00750 (0.00606)  --> STEP: 28/120 -- GLOBAL_STEP: 6925 | > loss_gen: 2.14338 (2.08854) | > loss_kl: 2.41955 (2.35462) | > loss_feat: 3.15846 (3.30319) | > loss_mel: 20.63266 (20.33599) | > loss_duration: 0.87609 (0.98662) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.23013 (29.06897) | > grad_norm_0: 64.58879 (83.02921) | > loss_disc: 2.50589 (2.55209) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50589 (2.55209) | > grad_norm_1: 17.61524 (18.55411) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.40420 (10.27286) | > loader_time: 0.01140 (0.00943)  --> STEP: 53/120 -- GLOBAL_STEP: 6950 | > loss_gen: 1.89681 (2.06767) | > loss_kl: 2.40338 (2.30765) | > loss_feat: 3.24426 (3.22191) | > loss_mel: 20.17330 (20.15056) | > loss_duration: 0.93935 (0.94871) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.65710 (28.69650) | > grad_norm_0: 29.03798 (78.83429) | > loss_disc: 2.53157 (2.56119) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53157 (2.56119) | > grad_norm_1: 19.10064 (18.18732) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.66760 (10.44953) | > loader_time: 0.01370 (0.01102)  --> STEP: 78/120 -- GLOBAL_STEP: 6975 | > loss_gen: 2.21691 (2.06244) | > loss_kl: 2.15747 (2.26755) | > loss_feat: 3.27422 (3.17651) | > loss_mel: 20.05462 (20.02970) | > loss_duration: 0.91740 (0.94203) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.62062 (28.47823) | > grad_norm_0: 100.08451 (86.88880) | > loss_disc: 2.51701 (2.56045) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51701 (2.56045) | > grad_norm_1: 7.86830 (18.90620) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.04530 (10.60391) | > loader_time: 0.01620 (0.01305)  --> STEP: 103/120 -- GLOBAL_STEP: 7000 | > loss_gen: 1.82676 (2.05160) | > loss_kl: 2.32399 (2.25441) | > loss_feat: 2.89653 (3.13371) | > loss_mel: 19.47689 (19.98133) | > loss_duration: 0.92538 (0.93890) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.44956 (28.35994) | > grad_norm_0: 85.22129 (87.75293) | > loss_disc: 2.65398 (2.57973) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65398 (2.57973) | > grad_norm_1: 15.74738 (19.16198) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.88830 (10.87710) | > loader_time: 0.01970 (0.01425) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.77213 (1.77213) | > loss_kl: 1.94302 (1.94302) | > loss_feat: 3.24824 (3.24824) | > loss_mel: 19.97829 (19.97829) | > loss_duration: 0.87029 (0.87029) | > loss_0: 27.81196 (27.81196) | > loss_disc: 2.66560 (2.66560) | > loss_1: 2.66560 (2.66560)  --> STEP: 1 | > loss_gen: 1.83972 (1.83972) | > loss_kl: 2.34024 (2.34024) | > loss_feat: 2.96457 (2.96457) | > loss_mel: 20.44721 (20.44721) | > loss_duration: 0.89091 (0.89091) | > loss_0: 28.48265 (28.48265) | > loss_disc: 2.73718 (2.73718) | > loss_1: 2.73718 (2.73718)  --> STEP: 2 | > loss_gen: 1.78913 (1.81442) | > loss_kl: 2.30207 (2.32115) | > loss_feat: 2.80009 (2.88233) | > loss_mel: 19.27607 (19.86164) | > loss_duration: 0.90188 (0.89640) | > loss_0: 27.06924 (27.77594) | > loss_disc: 2.75406 (2.74562) | > loss_1: 2.75406 (2.74562)  --> STEP: 3 | > loss_gen: 1.69021 (1.77302) | > loss_kl: 2.29776 (2.31336) | > loss_feat: 2.88134 (2.88200) | > loss_mel: 19.02414 (19.58247) | > loss_duration: 0.95611 (0.91630) | > loss_0: 26.84956 (27.46715) | > loss_disc: 2.85841 (2.78322) | > loss_1: 2.85841 (2.78322) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00732 (-0.00017) | > avg_loss_gen: 1.77302 (-0.19158) | > avg_loss_kl: 2.31336 (+0.12611) | > avg_loss_feat: 2.88200 (-0.35780) | > avg_loss_mel: 19.58247 (-0.49366) | > avg_loss_duration: 0.91630 (+0.00340) | > avg_loss_0: 27.46715 (-0.91354) | > avg_loss_disc: 2.78322 (+0.29134) | > avg_loss_1: 2.78322 (+0.29134)  > EPOCH: 58/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 11:36:35)   --> STEP: 7/120 -- GLOBAL_STEP: 7025 | > loss_gen: 2.12731 (2.07999) | > loss_kl: 2.31754 (2.33571) | > loss_feat: 2.95030 (3.18415) | > loss_mel: 20.11668 (20.76810) | > loss_duration: 0.90859 (0.89803) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.42042 (29.26598) | > grad_norm_0: 52.46907 (52.25737) | > loss_disc: 2.67668 (2.64740) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.67668 (2.64740) | > grad_norm_1: 27.15558 (24.08679) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.02630 (10.07180) | > loader_time: 0.00840 (0.00785)  --> STEP: 32/120 -- GLOBAL_STEP: 7050 | > loss_gen: 2.16311 (2.06122) | > loss_kl: 2.28403 (2.36960) | > loss_feat: 3.26211 (3.21035) | > loss_mel: 19.60853 (20.26575) | > loss_duration: 0.91675 (0.97573) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.23453 (28.88265) | > grad_norm_0: 100.94675 (104.11702) | > loss_disc: 2.42027 (2.56891) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.42027 (2.56891) | > grad_norm_1: 9.46550 (16.17175) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.54050 (10.34179) | > loader_time: 0.01200 (0.00988)  --> STEP: 57/120 -- GLOBAL_STEP: 7075 | > loss_gen: 1.66501 (2.05554) | > loss_kl: 2.27794 (2.28689) | > loss_feat: 3.17204 (3.17692) | > loss_mel: 19.59065 (20.11465) | > loss_duration: 0.93825 (0.94675) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.64390 (28.58075) | > grad_norm_0: 51.95790 (98.89390) | > loss_disc: 2.57432 (2.57277) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57432 (2.57277) | > grad_norm_1: 20.82692 (18.20039) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.79600 (10.50757) | > loader_time: 0.01410 (0.01171)  --> STEP: 82/120 -- GLOBAL_STEP: 7100 | > loss_gen: 1.97569 (2.04898) | > loss_kl: 2.28630 (2.25197) | > loss_feat: 3.11474 (3.14013) | > loss_mel: 19.16705 (20.00669) | > loss_duration: 0.90922 (0.94165) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.45300 (28.38941) | > grad_norm_0: 73.24243 (91.45502) | > loss_disc: 2.52991 (2.56945) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.52991 (2.56945) | > grad_norm_1: 7.45051 (17.73100) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.09660 (10.65079) | > loader_time: 0.01660 (0.01311)  --> STEP: 107/120 -- GLOBAL_STEP: 7125 | > loss_gen: 2.05193 (2.04767) | > loss_kl: 2.05748 (2.22905) | > loss_feat: 2.84234 (3.12098) | > loss_mel: 19.66989 (19.93614) | > loss_duration: 0.97412 (0.94060) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.59577 (28.27443) | > grad_norm_0: 76.67051 (92.11256) | > loss_disc: 2.61887 (2.57758) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61887 (2.57758) | > grad_norm_1: 13.49831 (19.69723) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.18350 (10.93345) | > loader_time: 0.02040 (0.01439) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.87490 (1.87490) | > loss_kl: 2.15885 (2.15885) | > loss_feat: 3.36710 (3.36710) | > loss_mel: 20.83557 (20.83557) | > loss_duration: 0.86513 (0.86513) | > loss_0: 29.10154 (29.10154) | > loss_disc: 2.56875 (2.56875) | > loss_1: 2.56875 (2.56875)  --> STEP: 1 | > loss_gen: 1.83602 (1.83602) | > loss_kl: 2.30534 (2.30534) | > loss_feat: 2.82617 (2.82617) | > loss_mel: 20.10774 (20.10774) | > loss_duration: 0.88992 (0.88992) | > loss_0: 27.96519 (27.96519) | > loss_disc: 2.77321 (2.77321) | > loss_1: 2.77321 (2.77321)  --> STEP: 2 | > loss_gen: 1.94796 (1.89199) | > loss_kl: 2.10754 (2.20644) | > loss_feat: 2.78225 (2.80421) | > loss_mel: 18.91685 (19.51229) | > loss_duration: 0.89927 (0.89460) | > loss_0: 26.65387 (27.30953) | > loss_disc: 2.64376 (2.70848) | > loss_1: 2.64376 (2.70848)  --> STEP: 3 | > loss_gen: 1.79146 (1.85848) | > loss_kl: 2.43485 (2.28258) | > loss_feat: 3.36628 (2.99157) | > loss_mel: 19.71528 (19.57995) | > loss_duration: 0.94925 (0.91281) | > loss_0: 28.25713 (27.62539) | > loss_disc: 2.66340 (2.69345) | > loss_1: 2.66340 (2.69345) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00726 (-0.00006) | > avg_loss_gen: 1.85848 (+0.08546) | > avg_loss_kl: 2.28258 (-0.03078) | > avg_loss_feat: 2.99157 (+0.10956) | > avg_loss_mel: 19.57995 (-0.00252) | > avg_loss_duration: 0.91281 (-0.00348) | > avg_loss_0: 27.62539 (+0.15825) | > avg_loss_disc: 2.69345 (-0.08976) | > avg_loss_1: 2.69345 (-0.08976)  > EPOCH: 59/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 11:59:00)   --> STEP: 11/120 -- GLOBAL_STEP: 7150 | > loss_gen: 1.96342 (2.09309) | > loss_kl: 2.34849 (2.37063) | > loss_feat: 3.34623 (3.24542) | > loss_mel: 20.65388 (20.58783) | > loss_duration: 0.91819 (1.09636) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.23021 (29.39333) | > grad_norm_0: 33.14827 (94.98134) | > loss_disc: 2.53277 (2.62733) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53277 (2.62733) | > grad_norm_1: 10.60442 (13.57656) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.49510 (10.14515) | > loader_time: 0.00930 (0.00775)  --> STEP: 36/120 -- GLOBAL_STEP: 7175 | > loss_gen: 1.96978 (2.06257) | > loss_kl: 2.25603 (2.33584) | > loss_feat: 3.28277 (3.17767) | > loss_mel: 19.92365 (20.23627) | > loss_duration: 0.89903 (0.96440) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.33126 (28.77674) | > grad_norm_0: 189.54184 (106.81150) | > loss_disc: 2.60294 (2.59578) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60294 (2.59578) | > grad_norm_1: 22.78250 (20.09564) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.57520 (10.33872) | > loader_time: 0.01190 (0.00989)  --> STEP: 61/120 -- GLOBAL_STEP: 7200 | > loss_gen: 2.04042 (2.05129) | > loss_kl: 2.21480 (2.28334) | > loss_feat: 3.15522 (3.13297) | > loss_mel: 19.50280 (20.10741) | > loss_duration: 0.95364 (0.94330) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.86687 (28.51832) | > grad_norm_0: 154.09906 (105.17485) | > loss_disc: 2.53574 (2.58246) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53574 (2.58246) | > grad_norm_1: 12.01996 (19.27293) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.70370 (10.47368) | > loader_time: 0.01470 (0.01262)  --> STEP: 86/120 -- GLOBAL_STEP: 7225 | > loss_gen: 2.19510 (2.04526) | > loss_kl: 2.13441 (2.24973) | > loss_feat: 3.08862 (3.10369) | > loss_mel: 19.89357 (20.02141) | > loss_duration: 0.93771 (0.93880) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.24941 (28.35890) | > grad_norm_0: 57.42654 (109.70415) | > loss_disc: 2.54505 (2.58300) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54505 (2.58300) | > grad_norm_1: 10.97742 (18.85128) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.49070 (10.65148) | > loader_time: 0.01710 (0.01351)  --> STEP: 111/120 -- GLOBAL_STEP: 7250 | > loss_gen: 1.97508 (2.04162) | > loss_kl: 2.09708 (2.23020) | > loss_feat: 2.76711 (3.07973) | > loss_mel: 19.18155 (19.92879) | > loss_duration: 0.98118 (0.93972) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.00200 (28.22006) | > grad_norm_0: 90.11685 (105.17553) | > loss_disc: 2.61281 (2.58892) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61281 (2.58892) | > grad_norm_1: 31.69799 (20.80353) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.83250 (10.92077) | > loader_time: 0.02180 (0.01483) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.17969 (2.17969) | > loss_kl: 2.03408 (2.03408) | > loss_feat: 3.21747 (3.21747) | > loss_mel: 20.81294 (20.81294) | > loss_duration: 0.86659 (0.86659) | > loss_0: 29.11077 (29.11077) | > loss_disc: 2.60634 (2.60634) | > loss_1: 2.60634 (2.60634)  --> STEP: 1 | > loss_gen: 2.24715 (2.24715) | > loss_kl: 2.20178 (2.20178) | > loss_feat: 2.97814 (2.97814) | > loss_mel: 20.43399 (20.43399) | > loss_duration: 0.88718 (0.88718) | > loss_0: 28.74825 (28.74825) | > loss_disc: 2.53735 (2.53735) | > loss_1: 2.53735 (2.53735)  --> STEP: 2 | > loss_gen: 2.16637 (2.20676) | > loss_kl: 2.02119 (2.11148) | > loss_feat: 3.15908 (3.06861) | > loss_mel: 19.82462 (20.12931) | > loss_duration: 0.90167 (0.89443) | > loss_0: 28.07294 (28.41059) | > loss_disc: 2.56271 (2.55003) | > loss_1: 2.56271 (2.55003)  --> STEP: 3 | > loss_gen: 2.10567 (2.17306) | > loss_kl: 2.14937 (2.12411) | > loss_feat: 3.19900 (3.11208) | > loss_mel: 19.33689 (19.86517) | > loss_duration: 0.94855 (0.91247) | > loss_0: 27.73948 (28.18689) | > loss_disc: 2.40986 (2.50331) | > loss_1: 2.40986 (2.50331) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00776 (+0.00050) | > avg_loss_gen: 2.17306 (+0.31458) | > avg_loss_kl: 2.12411 (-0.15847) | > avg_loss_feat: 3.11208 (+0.12051) | > avg_loss_mel: 19.86517 (+0.28522) | > avg_loss_duration: 0.91247 (-0.00035) | > avg_loss_0: 28.18689 (+0.56149) | > avg_loss_disc: 2.50331 (-0.19015) | > avg_loss_1: 2.50331 (-0.19015)  > EPOCH: 60/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 12:21:21)   --> STEP: 15/120 -- GLOBAL_STEP: 7275 | > loss_gen: 2.14514 (2.06920) | > loss_kl: 2.23026 (2.37849) | > loss_feat: 3.15477 (3.25708) | > loss_mel: 20.19878 (20.30463) | > loss_duration: 0.88081 (1.05473) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.60976 (29.06413) | > grad_norm_0: 84.11644 (81.87167) | > loss_disc: 2.65563 (2.55859) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65563 (2.55859) | > grad_norm_1: 8.99525 (14.92493) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.22990 (10.07078) | > loader_time: 0.01000 (0.00819)  --> STEP: 40/120 -- GLOBAL_STEP: 7300 | > loss_gen: 1.91241 (2.06236) | > loss_kl: 2.27949 (2.31691) | > loss_feat: 3.05431 (3.22501) | > loss_mel: 20.07209 (20.09810) | > loss_duration: 0.91337 (0.96279) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.23167 (28.66517) | > grad_norm_0: 193.74080 (83.74781) | > loss_disc: 2.65197 (2.55967) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65197 (2.55967) | > grad_norm_1: 79.19829 (15.73883) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.59120 (10.30532) | > loader_time: 0.01270 (0.01024)  --> STEP: 65/120 -- GLOBAL_STEP: 7325 | > loss_gen: 2.30850 (2.05339) | > loss_kl: 2.14248 (2.26800) | > loss_feat: 3.29833 (3.16760) | > loss_mel: 19.67635 (19.91144) | > loss_duration: 0.94088 (0.94410) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.36655 (28.34454) | > grad_norm_0: 35.20560 (78.67752) | > loss_disc: 2.41360 (2.56697) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.41360 (2.56697) | > grad_norm_1: 11.97977 (18.80631) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.73260 (10.46631) | > loader_time: 0.01530 (0.01342)  --> STEP: 90/120 -- GLOBAL_STEP: 7350 | > loss_gen: 2.08864 (2.04839) | > loss_kl: 2.13837 (2.23871) | > loss_feat: 2.98589 (3.14760) | > loss_mel: 19.21502 (19.84642) | > loss_duration: 0.92058 (0.93953) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.34851 (28.22064) | > grad_norm_0: 56.23258 (84.48105) | > loss_disc: 2.51948 (2.57188) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51948 (2.57188) | > grad_norm_1: 18.27751 (18.35810) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 13.40240 (10.69675) | > loader_time: 0.01820 (0.01422)  --> STEP: 115/120 -- GLOBAL_STEP: 7375 | > loss_gen: 2.11005 (2.04431) | > loss_kl: 2.19502 (2.21722) | > loss_feat: 2.95239 (3.12735) | > loss_mel: 19.64587 (19.78032) | > loss_duration: 1.03165 (0.94191) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.93497 (28.11110) | > grad_norm_0: 161.74161 (80.65556) | > loss_disc: 2.54955 (2.57346) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54955 (2.57346) | > grad_norm_1: 7.73712 (18.61072) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.57510 (10.97794) | > loader_time: 0.02390 (0.01554) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.00959 (2.00959) | > loss_kl: 2.23684 (2.23684) | > loss_feat: 2.98031 (2.98031) | > loss_mel: 20.03806 (20.03806) | > loss_duration: 0.87665 (0.87665) | > loss_0: 28.14145 (28.14145) | > loss_disc: 2.60724 (2.60724) | > loss_1: 2.60724 (2.60724)  --> STEP: 1 | > loss_gen: 1.96744 (1.96744) | > loss_kl: 2.23113 (2.23113) | > loss_feat: 3.11509 (3.11509) | > loss_mel: 20.68769 (20.68769) | > loss_duration: 0.88075 (0.88075) | > loss_0: 28.88209 (28.88209) | > loss_disc: 2.70098 (2.70098) | > loss_1: 2.70098 (2.70098)  --> STEP: 2 | > loss_gen: 1.98482 (1.97613) | > loss_kl: 2.17472 (2.20293) | > loss_feat: 3.03318 (3.07414) | > loss_mel: 19.75267 (20.22018) | > loss_duration: 0.90373 (0.89224) | > loss_0: 27.84912 (28.36561) | > loss_disc: 2.62164 (2.66131) | > loss_1: 2.62164 (2.66131)  --> STEP: 3 | > loss_gen: 2.03022 (1.99416) | > loss_kl: 2.31670 (2.24085) | > loss_feat: 3.38263 (3.17697) | > loss_mel: 19.73447 (20.05828) | > loss_duration: 0.95717 (0.91388) | > loss_0: 28.42119 (28.38413) | > loss_disc: 2.56548 (2.62937) | > loss_1: 2.56548 (2.62937) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00737 (-0.00039) | > avg_loss_gen: 1.99416 (-0.17891) | > avg_loss_kl: 2.24085 (+0.11674) | > avg_loss_feat: 3.17697 (+0.06489) | > avg_loss_mel: 20.05828 (+0.19311) | > avg_loss_duration: 0.91388 (+0.00141) | > avg_loss_0: 28.38413 (+0.19725) | > avg_loss_disc: 2.62937 (+0.12606) | > avg_loss_1: 2.62937 (+0.12606)  > EPOCH: 61/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 12:43:43)   --> STEP: 19/120 -- GLOBAL_STEP: 7400 | > loss_gen: 2.24367 (2.09712) | > loss_kl: 2.39865 (2.36296) | > loss_feat: 3.17916 (3.32752) | > loss_mel: 20.16001 (20.35294) | > loss_duration: 0.86812 (1.02274) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.84960 (29.16328) | > grad_norm_0: 66.14008 (91.76148) | > loss_disc: 2.61678 (2.55250) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61678 (2.55250) | > grad_norm_1: 14.71101 (14.31022) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.42940 (10.16456) | > loader_time: 0.01010 (0.00879)  --> STEP: 44/120 -- GLOBAL_STEP: 7425 | > loss_gen: 2.10771 (2.05724) | > loss_kl: 2.16794 (2.30377) | > loss_feat: 3.03000 (3.19021) | > loss_mel: 19.76460 (20.02527) | > loss_duration: 0.89654 (0.95599) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.96679 (28.53249) | > grad_norm_0: 24.22799 (85.41744) | > loss_disc: 2.64730 (2.57558) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.64730 (2.57558) | > grad_norm_1: 49.49896 (17.39286) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.43330 (10.37441) | > loader_time: 0.01310 (0.01058)  --> STEP: 69/120 -- GLOBAL_STEP: 7450 | > loss_gen: 1.89939 (2.05432) | > loss_kl: 2.13027 (2.27907) | > loss_feat: 2.88244 (3.16153) | > loss_mel: 19.80887 (19.92707) | > loss_duration: 0.96920 (0.94143) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.69017 (28.36341) | > grad_norm_0: 50.90362 (80.18891) | > loss_disc: 2.58984 (2.57174) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58984 (2.57174) | > grad_norm_1: 9.53329 (17.45859) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.88640 (10.53282) | > loader_time: 0.01530 (0.01266)  --> STEP: 94/120 -- GLOBAL_STEP: 7475 | > loss_gen: 2.02222 (2.04713) | > loss_kl: 2.08431 (2.25826) | > loss_feat: 3.23290 (3.13091) | > loss_mel: 19.72701 (19.86864) | > loss_duration: 0.93993 (0.93850) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.00637 (28.24343) | > grad_norm_0: 60.67556 (87.36433) | > loss_disc: 2.57138 (2.57839) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57138 (2.57839) | > grad_norm_1: 31.72515 (19.44137) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.80390 (10.74533) | > loader_time: 0.01840 (0.01375)  --> STEP: 119/120 -- GLOBAL_STEP: 7500 | > loss_gen: 2.02278 (2.04071) | > loss_kl: 2.06308 (2.23818) | > loss_feat: 2.59868 (3.09748) | > loss_mel: 19.67082 (19.82149) | > loss_duration: 1.03824 (0.94466) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.39361 (28.14252) | > grad_norm_0: 85.26974 (87.60104) | > loss_disc: 2.67759 (2.58917) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.67759 (2.58917) | > grad_norm_1: 20.08768 (20.13992) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.36320 (11.04763) | > loader_time: 0.02350 (0.01525) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.06463 (2.06463) | > loss_kl: 2.34708 (2.34708) | > loss_feat: 3.22454 (3.22454) | > loss_mel: 20.82752 (20.82752) | > loss_duration: 0.86077 (0.86077) | > loss_0: 29.32454 (29.32454) | > loss_disc: 2.55705 (2.55705) | > loss_1: 2.55705 (2.55705)  --> STEP: 1 | > loss_gen: 2.05433 (2.05433) | > loss_kl: 2.28419 (2.28419) | > loss_feat: 2.86325 (2.86325) | > loss_mel: 20.10190 (20.10190) | > loss_duration: 0.88374 (0.88374) | > loss_0: 28.18741 (28.18741) | > loss_disc: 2.67981 (2.67981) | > loss_1: 2.67981 (2.67981)  --> STEP: 2 | > loss_gen: 1.97309 (2.01371) | > loss_kl: 2.10012 (2.19215) | > loss_feat: 3.13740 (3.00033) | > loss_mel: 19.76765 (19.93477) | > loss_duration: 0.90268 (0.89321) | > loss_0: 27.88094 (28.03417) | > loss_disc: 2.70821 (2.69401) | > loss_1: 2.70821 (2.69401)  --> STEP: 3 | > loss_gen: 2.04902 (2.02548) | > loss_kl: 2.41940 (2.26790) | > loss_feat: 2.99817 (2.99961) | > loss_mel: 19.33886 (19.73613) | > loss_duration: 0.94525 (0.91056) | > loss_0: 27.75070 (27.93968) | > loss_disc: 2.62151 (2.66985) | > loss_1: 2.62151 (2.66985) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00717 (-0.00020) | > avg_loss_gen: 2.02548 (+0.03132) | > avg_loss_kl: 2.26790 (+0.02705) | > avg_loss_feat: 2.99961 (-0.17736) | > avg_loss_mel: 19.73613 (-0.32214) | > avg_loss_duration: 0.91056 (-0.00332) | > avg_loss_0: 27.93968 (-0.44445) | > avg_loss_disc: 2.66985 (+0.04048) | > avg_loss_1: 2.66985 (+0.04048)  > EPOCH: 62/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 13:06:07)   --> STEP: 23/120 -- GLOBAL_STEP: 7525 | > loss_gen: 2.00996 (2.07229) | > loss_kl: 2.20179 (2.36470) | > loss_feat: 3.24272 (3.28295) | > loss_mel: 19.79523 (20.28193) | > loss_duration: 0.92011 (1.00103) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.16981 (29.00290) | > grad_norm_0: 41.75062 (94.00300) | > loss_disc: 2.56134 (2.55207) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56134 (2.55207) | > grad_norm_1: 25.86481 (13.97608) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.40760 (10.24064) | > loader_time: 0.01050 (0.00907)  --> STEP: 48/120 -- GLOBAL_STEP: 7550 | > loss_gen: 2.24245 (2.06823) | > loss_kl: 2.12698 (2.29903) | > loss_feat: 3.03079 (3.23297) | > loss_mel: 19.33966 (20.01693) | > loss_duration: 0.90715 (0.94898) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.64702 (28.56614) | > grad_norm_0: 136.29170 (96.94650) | > loss_disc: 2.52468 (2.54836) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.52468 (2.54836) | > grad_norm_1: 10.12978 (14.46470) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.85930 (10.44182) | > loader_time: 0.01340 (0.01072)  --> STEP: 73/120 -- GLOBAL_STEP: 7575 | > loss_gen: 2.37659 (2.05876) | > loss_kl: 1.98607 (2.26880) | > loss_feat: 2.86529 (3.17636) | > loss_mel: 19.49137 (19.86279) | > loss_duration: 0.96373 (0.94104) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.68306 (28.30775) | > grad_norm_0: 97.96059 (86.29139) | > loss_disc: 2.76410 (2.56342) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.76410 (2.56342) | > grad_norm_1: 70.80351 (17.78226) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.90430 (10.58882) | > loader_time: 0.01580 (0.01292)  --> STEP: 98/120 -- GLOBAL_STEP: 7600 | > loss_gen: 2.00339 (2.05008) | > loss_kl: 2.04197 (2.23746) | > loss_feat: 2.90834 (3.14220) | > loss_mel: 19.58329 (19.79046) | > loss_duration: 0.93506 (0.93792) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.47203 (28.15811) | > grad_norm_0: 88.53051 (84.93713) | > loss_disc: 2.55588 (2.57071) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55588 (2.57071) | > grad_norm_1: 9.46111 (18.15086) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.74610 (10.82574) | > loader_time: 0.01890 (0.01401) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.95351 (1.95351) | > loss_kl: 2.18427 (2.18427) | > loss_feat: 2.99317 (2.99317) | > loss_mel: 20.80023 (20.80023) | > loss_duration: 0.88531 (0.88531) | > loss_0: 28.81649 (28.81649) | > loss_disc: 2.61953 (2.61953) | > loss_1: 2.61953 (2.61953)  --> STEP: 1 | > loss_gen: 2.04447 (2.04447) | > loss_kl: 2.34923 (2.34923) | > loss_feat: 2.95184 (2.95184) | > loss_mel: 19.77112 (19.77112) | > loss_duration: 0.88840 (0.88840) | > loss_0: 28.00506 (28.00506) | > loss_disc: 2.57697 (2.57697) | > loss_1: 2.57697 (2.57697)  --> STEP: 2 | > loss_gen: 2.01910 (2.03178) | > loss_kl: 2.22405 (2.28664) | > loss_feat: 3.26596 (3.10890) | > loss_mel: 19.02590 (19.39851) | > loss_duration: 0.89364 (0.89102) | > loss_0: 27.42865 (27.71686) | > loss_disc: 2.52972 (2.55334) | > loss_1: 2.52972 (2.55334)  --> STEP: 3 | > loss_gen: 2.00359 (2.02239) | > loss_kl: 2.51379 (2.36236) | > loss_feat: 3.36615 (3.19465) | > loss_mel: 20.29656 (19.69786) | > loss_duration: 0.95490 (0.91231) | > loss_0: 29.13499 (28.18957) | > loss_disc: 2.58846 (2.56505) | > loss_1: 2.58846 (2.56505) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00737 (+0.00020) | > avg_loss_gen: 2.02239 (-0.00309) | > avg_loss_kl: 2.36236 (+0.09445) | > avg_loss_feat: 3.19465 (+0.19504) | > avg_loss_mel: 19.69786 (-0.03827) | > avg_loss_duration: 0.91231 (+0.00176) | > avg_loss_0: 28.18957 (+0.24988) | > avg_loss_disc: 2.56505 (-0.10480) | > avg_loss_1: 2.56505 (-0.10480)  > EPOCH: 63/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 13:28:36)   --> STEP: 2/120 -- GLOBAL_STEP: 7625 | > loss_gen: 2.18972 (2.20889) | > loss_kl: 2.09555 (2.22030) | > loss_feat: 3.09059 (3.20583) | > loss_mel: 19.31790 (19.76684) | > loss_duration: 0.89556 (0.91095) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.58932 (28.31281) | > grad_norm_0: 71.32873 (111.23141) | > loss_disc: 2.60155 (2.56834) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60155 (2.56834) | > grad_norm_1: 24.53301 (22.46862) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.82230 (9.67725) | > loader_time: 0.00510 (0.00634)  --> STEP: 27/120 -- GLOBAL_STEP: 7650 | > loss_gen: 1.70815 (2.04940) | > loss_kl: 2.37393 (2.35220) | > loss_feat: 3.09727 (3.22384) | > loss_mel: 19.86360 (20.06968) | > loss_duration: 0.91020 (0.99091) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.95315 (28.68603) | > grad_norm_0: 109.92156 (112.13412) | > loss_disc: 2.63294 (2.57509) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.63294 (2.57509) | > grad_norm_1: 50.99286 (18.97555) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.42530 (10.22899) | > loader_time: 0.01130 (0.00936)  --> STEP: 52/120 -- GLOBAL_STEP: 7675 | > loss_gen: 2.16252 (2.06382) | > loss_kl: 2.32356 (2.29703) | > loss_feat: 3.35653 (3.22121) | > loss_mel: 19.36875 (19.91293) | > loss_duration: 0.94096 (0.94913) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 28.15233 (28.44413) | > grad_norm_0: 35.41684 (107.17219) | > loss_disc: 2.47559 (2.55454) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.47559 (2.55454) | > grad_norm_1: 21.63180 (19.45779) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.69940 (10.39943) | > loader_time: 0.01380 (0.01096)  --> STEP: 77/120 -- GLOBAL_STEP: 7700 | > loss_gen: 1.98965 (2.06571) | > loss_kl: 2.07764 (2.26772) | > loss_feat: 2.80517 (3.20920) | > loss_mel: 20.11066 (19.81998) | > loss_duration: 0.96911 (0.94156) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.95222 (28.30417) | > grad_norm_0: 88.45143 (97.82262) | > loss_disc: 2.60556 (2.56734) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.60556 (2.56734) | > grad_norm_1: 23.74390 (20.52335) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.87040 (10.53492) | > loader_time: 0.01610 (0.01317)  --> STEP: 102/120 -- GLOBAL_STEP: 7725 | > loss_gen: 2.07666 (2.05499) | > loss_kl: 2.10775 (2.24990) | > loss_feat: 3.15026 (3.17469) | > loss_mel: 19.13179 (19.73546) | > loss_duration: 0.91068 (0.93886) | > amp_scaler: 1024.00000 (1024.00000) | > loss_0: 27.37714 (28.15391) | > grad_norm_0: 34.48217 (93.35261) | > loss_disc: 2.46612 (2.56479) | > amp_scaler-1: 1024.00000 (1024.00000) | > loss_1: 2.46612 (2.56479) | > grad_norm_1: 14.43228 (19.05395) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.90890 (10.79521) | > loader_time: 0.01850 (0.01454) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.89692 (1.89692) | > loss_kl: 2.32216 (2.32216) | > loss_feat: 3.22221 (3.22221) | > loss_mel: 19.97321 (19.97321) | > loss_duration: 0.88028 (0.88028) | > loss_0: 28.29477 (28.29477) | > loss_disc: 2.60792 (2.60792) | > loss_1: 2.60792 (2.60792)  --> STEP: 1 | > loss_gen: 2.00166 (2.00166) | > loss_kl: 2.23225 (2.23225) | > loss_feat: 3.25192 (3.25192) | > loss_mel: 19.80547 (19.80547) | > loss_duration: 0.88577 (0.88577) | > loss_0: 28.17706 (28.17706) | > loss_disc: 2.65952 (2.65952) | > loss_1: 2.65952 (2.65952)  --> STEP: 2 | > loss_gen: 2.02331 (2.01248) | > loss_kl: 2.23797 (2.23511) | > loss_feat: 3.16465 (3.20828) | > loss_mel: 19.36187 (19.58367) | > loss_duration: 0.90887 (0.89732) | > loss_0: 27.69667 (27.93686) | > loss_disc: 2.48816 (2.57384) | > loss_1: 2.48816 (2.57384)  --> STEP: 3 | > loss_gen: 1.92978 (1.98491) | > loss_kl: 2.28723 (2.25248) | > loss_feat: 3.06559 (3.16072) | > loss_mel: 19.26157 (19.47630) | > loss_duration: 0.94984 (0.91482) | > loss_0: 27.49400 (27.78924) | > loss_disc: 2.66357 (2.60375) | > loss_1: 2.66357 (2.60375) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00719 (-0.00018) | > avg_loss_gen: 1.98491 (-0.03747) | > avg_loss_kl: 2.25248 (-0.10987) | > avg_loss_feat: 3.16072 (-0.03393) | > avg_loss_mel: 19.47630 (-0.22156) | > avg_loss_duration: 0.91482 (+0.00251) | > avg_loss_0: 27.78924 (-0.40032) | > avg_loss_disc: 2.60375 (+0.03870) | > avg_loss_1: 2.60375 (+0.03870)  > EPOCH: 64/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 13:50:54)   --> STEP: 6/120 -- GLOBAL_STEP: 7750 | > loss_gen: 1.83620 (2.13949) | > loss_kl: 2.43318 (2.39482) | > loss_feat: 3.37499 (3.56828) | > loss_mel: 19.96329 (20.26132) | > loss_duration: 0.95316 (0.92485) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.56082 (29.28875) | > grad_norm_0: 96.69296 (72.79628) | > loss_disc: 2.64406 (2.56320) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.64406 (2.56320) | > grad_norm_1: 13.06307 (26.28142) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.36820 (9.85449) | > loader_time: 0.00850 (0.00785)  --> STEP: 31/120 -- GLOBAL_STEP: 7775 | > loss_gen: 2.02149 (2.08622) | > loss_kl: 2.25873 (2.35022) | > loss_feat: 3.33003 (3.33444) | > loss_mel: 20.16823 (19.96198) | > loss_duration: 0.90319 (0.97975) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.68166 (28.71261) | > grad_norm_0: 104.74156 (66.60847) | > loss_disc: 2.51185 (2.57625) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51185 (2.57625) | > grad_norm_1: 22.33179 (18.43998) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.24180 (10.24146) | > loader_time: 0.01210 (0.00981)  --> STEP: 56/120 -- GLOBAL_STEP: 7800 | > loss_gen: 2.21760 (2.07428) | > loss_kl: 2.23081 (2.28943) | > loss_feat: 3.13957 (3.26272) | > loss_mel: 19.09965 (19.83677) | > loss_duration: 0.90514 (0.94559) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.59276 (28.40879) | > grad_norm_0: 86.29110 (87.06149) | > loss_disc: 2.49649 (2.56051) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.49649 (2.56051) | > grad_norm_1: 9.78941 (19.37690) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.89050 (10.41458) | > loader_time: 0.01430 (0.01166)  --> STEP: 81/120 -- GLOBAL_STEP: 7825 | > loss_gen: 1.84920 (2.06872) | > loss_kl: 2.12379 (2.26401) | > loss_feat: 3.48427 (3.24993) | > loss_mel: 19.90114 (19.74888) | > loss_duration: 0.91809 (0.93991) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.27649 (28.27146) | > grad_norm_0: 104.98607 (92.49677) | > loss_disc: 2.58664 (2.55920) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58664 (2.55920) | > grad_norm_1: 41.32860 (20.21970) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.05400 (10.56630) | > loader_time: 0.01670 (0.01337)  --> STEP: 106/120 -- GLOBAL_STEP: 7850 | > loss_gen: 2.09321 (2.06957) | > loss_kl: 2.07460 (2.23925) | > loss_feat: 3.34037 (3.23713) | > loss_mel: 19.85522 (19.70035) | > loss_duration: 0.95749 (0.93812) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.32089 (28.18442) | > grad_norm_0: 38.83794 (91.33960) | > loss_disc: 2.53989 (2.55595) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53989 (2.55595) | > grad_norm_1: 11.89943 (20.12357) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.93770 (10.85748) | > loader_time: 0.02190 (0.01460) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.10814 (2.10814) | > loss_kl: 2.37600 (2.37600) | > loss_feat: 3.06235 (3.06235) | > loss_mel: 19.34264 (19.34264) | > loss_duration: 0.87450 (0.87450) | > loss_0: 27.76363 (27.76363) | > loss_disc: 2.44453 (2.44453) | > loss_1: 2.44453 (2.44453)  --> STEP: 1 | > loss_gen: 2.02609 (2.02609) | > loss_kl: 2.23015 (2.23015) | > loss_feat: 2.92327 (2.92327) | > loss_mel: 20.35059 (20.35059) | > loss_duration: 0.87931 (0.87931) | > loss_0: 28.40940 (28.40940) | > loss_disc: 2.63890 (2.63890) | > loss_1: 2.63890 (2.63890)  --> STEP: 2 | > loss_gen: 1.99685 (2.01147) | > loss_kl: 2.55429 (2.39222) | > loss_feat: 3.03748 (2.98038) | > loss_mel: 19.69293 (20.02176) | > loss_duration: 0.90673 (0.89302) | > loss_0: 28.18828 (28.29884) | > loss_disc: 2.59438 (2.61664) | > loss_1: 2.59438 (2.61664)  --> STEP: 3 | > loss_gen: 2.05453 (2.02582) | > loss_kl: 2.30450 (2.36298) | > loss_feat: 2.87234 (2.94436) | > loss_mel: 19.36865 (19.80406) | > loss_duration: 0.96544 (0.91716) | > loss_0: 27.56546 (28.05438) | > loss_disc: 2.68585 (2.63971) | > loss_1: 2.68585 (2.63971) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00730 (+0.00011) | > avg_loss_gen: 2.02582 (+0.04091) | > avg_loss_kl: 2.36298 (+0.11050) | > avg_loss_feat: 2.94436 (-0.21635) | > avg_loss_mel: 19.80406 (+0.32775) | > avg_loss_duration: 0.91716 (+0.00233) | > avg_loss_0: 28.05438 (+0.26514) | > avg_loss_disc: 2.63971 (+0.03596) | > avg_loss_1: 2.63971 (+0.03596)  > EPOCH: 65/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 14:13:13)   --> STEP: 10/120 -- GLOBAL_STEP: 7875 | > loss_gen: 2.20973 (2.07196) | > loss_kl: 2.58283 (2.41826) | > loss_feat: 3.69140 (3.37572) | > loss_mel: 20.77580 (20.42956) | > loss_duration: 0.89887 (0.90947) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.15863 (29.20497) | > grad_norm_0: 195.37115 (91.14139) | > loss_disc: 2.51835 (2.58478) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51835 (2.58478) | > grad_norm_1: 30.96264 (25.72701) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.16530 (10.04945) | > loader_time: 0.00860 (0.00771)  --> STEP: 35/120 -- GLOBAL_STEP: 7900 | > loss_gen: 1.91838 (2.08199) | > loss_kl: 2.26236 (2.31073) | > loss_feat: 3.49827 (3.32230) | > loss_mel: 19.84685 (20.02116) | > loss_duration: 0.88417 (0.96549) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.41002 (28.70166) | > grad_norm_0: 41.32654 (112.79308) | > loss_disc: 2.53087 (2.56127) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53087 (2.56127) | > grad_norm_1: 6.31833 (21.49747) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.45720 (10.32223) | > loader_time: 0.01240 (0.00998)  --> STEP: 60/120 -- GLOBAL_STEP: 7925 | > loss_gen: 1.99440 (2.06719) | > loss_kl: 2.18988 (2.26863) | > loss_feat: 3.12244 (3.23541) | > loss_mel: 19.47517 (19.85359) | > loss_duration: 0.94645 (0.94258) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.72834 (28.36741) | > grad_norm_0: 243.01015 (106.63850) | > loss_disc: 2.54792 (2.56039) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.54792 (2.56039) | > grad_norm_1: 32.48194 (21.44268) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.14130 (10.49248) | > loader_time: 0.01470 (0.01345)  --> STEP: 85/120 -- GLOBAL_STEP: 7950 | > loss_gen: 1.89412 (2.07044) | > loss_kl: 2.15481 (2.24965) | > loss_feat: 2.97227 (3.22586) | > loss_mel: 19.35299 (19.77537) | > loss_duration: 0.92958 (0.93737) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.30377 (28.25870) | > grad_norm_0: 126.33882 (103.55995) | > loss_disc: 2.63330 (2.56085) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63330 (2.56085) | > grad_norm_1: 10.54891 (20.71940) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.72290 (10.66036) | > loader_time: 0.01650 (0.01414)  --> STEP: 110/120 -- GLOBAL_STEP: 7975 | > loss_gen: 2.33679 (2.06691) | > loss_kl: 2.17749 (2.23564) | > loss_feat: 3.18010 (3.21013) | > loss_mel: 19.44680 (19.70537) | > loss_duration: 0.94533 (0.93742) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.08650 (28.15547) | > grad_norm_0: 39.65416 (101.72397) | > loss_disc: 2.61610 (2.56197) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61610 (2.56197) | > grad_norm_1: 37.77835 (19.73339) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.07900 (10.94555) | > loader_time: 0.02100 (0.01523) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.37078 (2.37078) | > loss_kl: 2.38575 (2.38575) | > loss_feat: 3.07165 (3.07165) | > loss_mel: 20.72994 (20.72994) | > loss_duration: 0.88069 (0.88069) | > loss_0: 29.43881 (29.43881) | > loss_disc: 2.70580 (2.70580) | > loss_1: 2.70580 (2.70580)  --> STEP: 1 | > loss_gen: 2.32340 (2.32340) | > loss_kl: 2.39185 (2.39185) | > loss_feat: 2.86601 (2.86601) | > loss_mel: 19.64978 (19.64978) | > loss_duration: 0.90716 (0.90716) | > loss_0: 28.13821 (28.13821) | > loss_disc: 2.69457 (2.69457) | > loss_1: 2.69457 (2.69457)  --> STEP: 2 | > loss_gen: 2.45482 (2.38911) | > loss_kl: 1.95081 (2.17133) | > loss_feat: 3.15018 (3.00810) | > loss_mel: 19.27954 (19.46466) | > loss_duration: 0.91140 (0.90928) | > loss_0: 27.74675 (27.94248) | > loss_disc: 2.62594 (2.66026) | > loss_1: 2.62594 (2.66026)  --> STEP: 3 | > loss_gen: 2.41081 (2.39634) | > loss_kl: 2.02729 (2.12332) | > loss_feat: 3.30040 (3.10553) | > loss_mel: 19.37623 (19.43518) | > loss_duration: 0.96644 (0.92834) | > loss_0: 28.08117 (27.98871) | > loss_disc: 2.62501 (2.64851) | > loss_1: 2.62501 (2.64851) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00733 (+0.00003) | > avg_loss_gen: 2.39634 (+0.37052) | > avg_loss_kl: 2.12332 (-0.23966) | > avg_loss_feat: 3.10553 (+0.16117) | > avg_loss_mel: 19.43518 (-0.36888) | > avg_loss_duration: 0.92834 (+0.01118) | > avg_loss_0: 27.98871 (-0.06567) | > avg_loss_disc: 2.64851 (+0.00880) | > avg_loss_1: 2.64851 (+0.00880)  > EPOCH: 66/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 14:35:38)   --> STEP: 14/120 -- GLOBAL_STEP: 8000 | > loss_gen: 2.12496 (2.07122) | > loss_kl: 2.54529 (2.47705) | > loss_feat: 3.41001 (3.44584) | > loss_mel: 19.66202 (20.13095) | > loss_duration: 3.05249 (1.05746) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 30.79477 (29.18253) | > grad_norm_0: 28.54446 (84.40405) | > loss_disc: 2.59830 (2.50882) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59830 (2.50882) | > grad_norm_1: 7.48203 (11.91768) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.76540 (10.15703) | > loader_time: 0.00940 (0.00818)  --> STEP: 39/120 -- GLOBAL_STEP: 8025 | > loss_gen: 1.95949 (2.07321) | > loss_kl: 2.23818 (2.36344) | > loss_feat: 3.06854 (3.33959) | > loss_mel: 20.00803 (19.98907) | > loss_duration: 0.88856 (0.95906) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.16279 (28.72438) | > grad_norm_0: 118.95966 (88.55050) | > loss_disc: 2.58439 (2.57112) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58439 (2.57112) | > grad_norm_1: 4.87342 (17.68685) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.80460 (10.38564) | > loader_time: 0.01240 (0.01008)  --> STEP: 64/120 -- GLOBAL_STEP: 8050 | > loss_gen: 2.26447 (2.06153) | > loss_kl: 2.16456 (2.32393) | > loss_feat: 3.51273 (3.25767) | > loss_mel: 19.68966 (19.82356) | > loss_duration: 0.94738 (0.94011) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.57880 (28.40680) | > grad_norm_0: 46.93285 (88.48358) | > loss_disc: 2.41258 (2.57505) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.41258 (2.57505) | > grad_norm_1: 18.20573 (19.55763) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.79490 (10.55187) | > loader_time: 0.01460 (0.01356)  --> STEP: 89/120 -- GLOBAL_STEP: 8075 | > loss_gen: 2.38776 (2.05510) | > loss_kl: 2.21603 (2.28575) | > loss_feat: 3.21846 (3.20746) | > loss_mel: 19.61639 (19.72842) | > loss_duration: 0.92623 (0.93685) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.36488 (28.21359) | > grad_norm_0: 162.54172 (101.33804) | > loss_disc: 2.66568 (2.57545) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.66568 (2.57545) | > grad_norm_1: 66.75716 (19.68871) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.68530 (10.75677) | > loader_time: 0.01730 (0.01427)  --> STEP: 114/120 -- GLOBAL_STEP: 8100 | > loss_gen: 2.02914 (2.05412) | > loss_kl: 2.09571 (2.26277) | > loss_feat: 2.85384 (3.19416) | > loss_mel: 18.79588 (19.65085) | > loss_duration: 0.96228 (0.93886) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.73686 (28.10076) | > grad_norm_0: 63.08590 (98.39704) | > loss_disc: 2.69416 (2.57713) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.69416 (2.57713) | > grad_norm_1: 11.95376 (19.71067) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.02590 (11.03419) | > loader_time: 0.02100 (0.01546) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.90427 (1.90427) | > loss_kl: 2.29785 (2.29785) | > loss_feat: 3.30613 (3.30613) | > loss_mel: 19.66795 (19.66795) | > loss_duration: 0.87164 (0.87164) | > loss_0: 28.04784 (28.04784) | > loss_disc: 2.43554 (2.43554) | > loss_1: 2.43554 (2.43554)  --> STEP: 1 | > loss_gen: 1.94379 (1.94379) | > loss_kl: 2.02151 (2.02151) | > loss_feat: 3.47143 (3.47143) | > loss_mel: 19.86694 (19.86694) | > loss_duration: 0.88113 (0.88113) | > loss_0: 28.18481 (28.18481) | > loss_disc: 2.39871 (2.39871) | > loss_1: 2.39871 (2.39871)  --> STEP: 2 | > loss_gen: 1.95360 (1.94870) | > loss_kl: 2.34601 (2.18376) | > loss_feat: 3.22779 (3.34961) | > loss_mel: 19.00028 (19.43361) | > loss_duration: 0.89937 (0.89025) | > loss_0: 27.42704 (27.80593) | > loss_disc: 2.37322 (2.38597) | > loss_1: 2.37322 (2.38597)  --> STEP: 3 | > loss_gen: 1.89177 (1.92972) | > loss_kl: 2.12365 (2.16372) | > loss_feat: 3.05525 (3.25149) | > loss_mel: 18.89889 (19.25537) | > loss_duration: 0.95851 (0.91301) | > loss_0: 26.92808 (27.51331) | > loss_disc: 2.57408 (2.44867) | > loss_1: 2.57408 (2.44867) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00740 (+0.00007) | > avg_loss_gen: 1.92972 (-0.46662) | > avg_loss_kl: 2.16372 (+0.04040) | > avg_loss_feat: 3.25149 (+0.14596) | > avg_loss_mel: 19.25537 (-0.17981) | > avg_loss_duration: 0.91301 (-0.01533) | > avg_loss_0: 27.51331 (-0.47540) | > avg_loss_disc: 2.44867 (-0.19984) | > avg_loss_1: 2.44867 (-0.19984)  > EPOCH: 67/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 14:58:08)   --> STEP: 18/120 -- GLOBAL_STEP: 8125 | > loss_gen: 2.43454 (2.11313) | > loss_kl: 2.24256 (2.42445) | > loss_feat: 3.49404 (3.32194) | > loss_mel: 19.79354 (20.02908) | > loss_duration: 0.91799 (1.02478) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.88267 (28.91337) | > grad_norm_0: 154.22223 (110.46191) | > loss_disc: 2.46165 (2.56607) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.46165 (2.56607) | > grad_norm_1: 12.98776 (16.17447) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.55180 (10.15295) | > loader_time: 0.01050 (0.00874)  --> STEP: 43/120 -- GLOBAL_STEP: 8150 | > loss_gen: 2.16260 (2.08443) | > loss_kl: 2.20594 (2.33291) | > loss_feat: 2.95199 (3.28638) | > loss_mel: 19.36769 (19.75745) | > loss_duration: 0.90999 (0.95361) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.59822 (28.41479) | > grad_norm_0: 43.06282 (102.57701) | > loss_disc: 2.60644 (2.56225) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60644 (2.56225) | > grad_norm_1: 10.60443 (21.19404) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.87120 (10.35478) | > loader_time: 0.01300 (0.01047)  --> STEP: 68/120 -- GLOBAL_STEP: 8175 | > loss_gen: 1.91793 (2.07518) | > loss_kl: 2.22308 (2.29637) | > loss_feat: 3.39551 (3.27757) | > loss_mel: 19.84448 (19.67695) | > loss_duration: 0.92357 (0.94083) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.30458 (28.26690) | > grad_norm_0: 36.72970 (95.38201) | > loss_disc: 2.48707 (2.55589) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48707 (2.55589) | > grad_norm_1: 47.57534 (20.99480) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.98780 (10.49658) | > loader_time: 0.01520 (0.01273)  --> STEP: 93/120 -- GLOBAL_STEP: 8200 | > loss_gen: 2.30676 (2.07941) | > loss_kl: 2.32167 (2.26353) | > loss_feat: 2.92780 (3.26068) | > loss_mel: 19.56902 (19.61141) | > loss_duration: 0.94727 (0.93758) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.07252 (28.15261) | > grad_norm_0: 36.89656 (86.15941) | > loss_disc: 2.58009 (2.56392) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58009 (2.56392) | > grad_norm_1: 11.40886 (21.80264) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.42920 (10.70587) | > loader_time: 0.02010 (0.01384)  --> STEP: 118/120 -- GLOBAL_STEP: 8225 | > loss_gen: 2.26675 (2.06708) | > loss_kl: 2.06029 (2.24481) | > loss_feat: 2.98466 (3.22156) | > loss_mel: 19.02085 (19.56726) | > loss_duration: 1.03442 (0.94298) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.36696 (28.04369) | > grad_norm_0: 43.59752 (83.93261) | > loss_disc: 2.63994 (2.56399) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63994 (2.56399) | > grad_norm_1: 25.57780 (21.19754) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.17430 (10.99639) | > loader_time: 0.02590 (0.01533) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.96620 (1.96620) | > loss_kl: 2.29578 (2.29578) | > loss_feat: 3.23621 (3.23621) | > loss_mel: 20.37747 (20.37747) | > loss_duration: 0.88073 (0.88073) | > loss_0: 28.75638 (28.75638) | > loss_disc: 2.63068 (2.63068) | > loss_1: 2.63068 (2.63068)  --> STEP: 1 | > loss_gen: 2.01876 (2.01876) | > loss_kl: 2.43191 (2.43191) | > loss_feat: 3.28164 (3.28164) | > loss_mel: 19.83760 (19.83760) | > loss_duration: 0.88690 (0.88690) | > loss_0: 28.45682 (28.45682) | > loss_disc: 2.52733 (2.52733) | > loss_1: 2.52733 (2.52733)  --> STEP: 2 | > loss_gen: 2.04263 (2.03070) | > loss_kl: 1.99892 (2.21541) | > loss_feat: 2.97845 (3.13005) | > loss_mel: 19.63240 (19.73500) | > loss_duration: 0.90527 (0.89608) | > loss_0: 27.55767 (28.00724) | > loss_disc: 2.50061 (2.51397) | > loss_1: 2.50061 (2.51397)  --> STEP: 3 | > loss_gen: 1.99023 (2.01721) | > loss_kl: 2.09180 (2.17421) | > loss_feat: 3.10502 (3.12171) | > loss_mel: 19.92527 (19.79842) | > loss_duration: 0.95647 (0.91621) | > loss_0: 28.06878 (28.02776) | > loss_disc: 2.62975 (2.55256) | > loss_1: 2.62975 (2.55256) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00723 (-0.00017) | > avg_loss_gen: 2.01721 (+0.08749) | > avg_loss_kl: 2.17421 (+0.01049) | > avg_loss_feat: 3.12171 (-0.12979) | > avg_loss_mel: 19.79842 (+0.54305) | > avg_loss_duration: 0.91621 (+0.00321) | > avg_loss_0: 28.02776 (+0.51445) | > avg_loss_disc: 2.55256 (+0.10389) | > avg_loss_1: 2.55256 (+0.10389)  > EPOCH: 68/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 15:20:28)   --> STEP: 22/120 -- GLOBAL_STEP: 8250 | > loss_gen: 2.25755 (2.06862) | > loss_kl: 2.26787 (2.37157) | > loss_feat: 3.13213 (3.26943) | > loss_mel: 19.67604 (20.16580) | > loss_duration: 0.90704 (1.00558) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.24063 (28.88100) | > grad_norm_0: 148.81564 (80.75447) | > loss_disc: 2.51683 (2.62524) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51683 (2.62524) | > grad_norm_1: 16.73501 (22.88623) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.56200 (10.20542) | > loader_time: 0.01090 (0.00891)  --> STEP: 47/120 -- GLOBAL_STEP: 8275 | > loss_gen: 2.08518 (2.05346) | > loss_kl: 2.10658 (2.32837) | > loss_feat: 3.17821 (3.22026) | > loss_mel: 19.63728 (19.85435) | > loss_duration: 0.90908 (0.95148) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.91632 (28.40792) | > grad_norm_0: 32.16253 (89.25785) | > loss_disc: 2.69576 (2.59066) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.69576 (2.59066) | > grad_norm_1: 28.28272 (17.60301) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.50070 (10.39271) | > loader_time: 0.01350 (0.01072)  --> STEP: 72/120 -- GLOBAL_STEP: 8300 | > loss_gen: 1.97489 (2.04237) | > loss_kl: 2.27287 (2.29369) | > loss_feat: 3.14790 (3.19086) | > loss_mel: 19.62223 (19.70102) | > loss_duration: 0.96534 (0.94132) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.98323 (28.16926) | > grad_norm_0: 40.67175 (101.46066) | > loss_disc: 2.51910 (2.59046) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51910 (2.59046) | > grad_norm_1: 7.36662 (18.41820) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.15130 (10.57858) | > loader_time: 0.01650 (0.01325)  --> STEP: 97/120 -- GLOBAL_STEP: 8325 | > loss_gen: 2.11570 (2.04156) | > loss_kl: 2.10036 (2.27495) | > loss_feat: 3.34095 (3.17673) | > loss_mel: 19.18139 (19.63677) | > loss_duration: 0.92530 (0.93800) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.66369 (28.06802) | > grad_norm_0: 51.67953 (103.76897) | > loss_disc: 2.51780 (2.58916) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51780 (2.58916) | > grad_norm_1: 10.03026 (19.32116) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.31690 (10.83283) | > loader_time: 0.01900 (0.01427) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.02354 (2.02354) | > loss_kl: 2.53399 (2.53399) | > loss_feat: 3.21218 (3.21218) | > loss_mel: 19.00040 (19.00040) | > loss_duration: 0.87035 (0.87035) | > loss_0: 27.64047 (27.64047) | > loss_disc: 2.69731 (2.69731) | > loss_1: 2.69731 (2.69731)  --> STEP: 1 | > loss_gen: 1.98774 (1.98774) | > loss_kl: 2.26081 (2.26081) | > loss_feat: 3.23695 (3.23695) | > loss_mel: 19.94589 (19.94589) | > loss_duration: 0.89177 (0.89177) | > loss_0: 28.32315 (28.32315) | > loss_disc: 2.66689 (2.66689) | > loss_1: 2.66689 (2.66689)  --> STEP: 2 | > loss_gen: 2.09195 (2.03985) | > loss_kl: 2.16306 (2.21193) | > loss_feat: 3.45845 (3.34770) | > loss_mel: 19.33773 (19.64181) | > loss_duration: 0.90730 (0.89953) | > loss_0: 27.95848 (28.14082) | > loss_disc: 2.70194 (2.68441) | > loss_1: 2.70194 (2.68441)  --> STEP: 3 | > loss_gen: 2.02857 (2.03609) | > loss_kl: 2.18203 (2.20197) | > loss_feat: 3.43924 (3.37821) | > loss_mel: 19.86246 (19.71536) | > loss_duration: 0.96427 (0.92111) | > loss_0: 28.47656 (28.25273) | > loss_disc: 2.74679 (2.70520) | > loss_1: 2.74679 (2.70520) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00737 (+0.00015) | > avg_loss_gen: 2.03609 (+0.01888) | > avg_loss_kl: 2.20197 (+0.02776) | > avg_loss_feat: 3.37821 (+0.25651) | > avg_loss_mel: 19.71536 (-0.08307) | > avg_loss_duration: 0.92111 (+0.00490) | > avg_loss_0: 28.25273 (+0.22498) | > avg_loss_disc: 2.70520 (+0.15264) | > avg_loss_1: 2.70520 (+0.15264)  > EPOCH: 69/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 15:42:57)   --> STEP: 1/120 -- GLOBAL_STEP: 8350 | > loss_gen: 2.45970 (2.45970) | > loss_kl: 2.22764 (2.22764) | > loss_feat: 3.67309 (3.67309) | > loss_mel: 20.38396 (20.38396) | > loss_duration: 0.91812 (0.91812) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.66252 (29.66252) | > grad_norm_0: 64.14699 (64.14699) | > loss_disc: 2.66746 (2.66746) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.66746 (2.66746) | > grad_norm_1: 21.28176 (21.28176) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.79670 (9.79667) | > loader_time: 0.00730 (0.00727)  --> STEP: 26/120 -- GLOBAL_STEP: 8375 | > loss_gen: 2.22047 (2.07016) | > loss_kl: 2.08614 (2.37212) | > loss_feat: 3.03085 (3.30512) | > loss_mel: 19.31864 (19.78702) | > loss_duration: 0.92390 (0.98496) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.58001 (28.51939) | > grad_norm_0: 111.49171 (80.01086) | > loss_disc: 2.59763 (2.60177) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59763 (2.60177) | > grad_norm_1: 23.91972 (22.92690) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.49680 (10.36652) | > loader_time: 0.01190 (0.00943)  --> STEP: 51/120 -- GLOBAL_STEP: 8400 | > loss_gen: 2.23124 (2.05069) | > loss_kl: 2.32013 (2.29445) | > loss_feat: 3.09226 (3.22937) | > loss_mel: 19.46565 (19.69016) | > loss_duration: 0.93909 (0.94347) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.04836 (28.20815) | > grad_norm_0: 36.79823 (88.73267) | > loss_disc: 2.50389 (2.59003) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.50389 (2.59003) | > grad_norm_1: 21.35301 (22.52137) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.81410 (10.53237) | > loader_time: 0.01370 (0.01106)  --> STEP: 76/120 -- GLOBAL_STEP: 8425 | > loss_gen: 2.09738 (2.04404) | > loss_kl: 2.09011 (2.26249) | > loss_feat: 2.99231 (3.20496) | > loss_mel: 19.32248 (19.66029) | > loss_duration: 0.92538 (0.93782) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.42766 (28.10961) | > grad_norm_0: 69.93002 (98.46638) | > loss_disc: 2.55453 (2.59168) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55453 (2.59168) | > grad_norm_1: 8.68894 (21.01655) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.16050 (10.68929) | > loader_time: 0.01640 (0.01317)  --> STEP: 101/120 -- GLOBAL_STEP: 8450 | > loss_gen: 2.32538 (2.04243) | > loss_kl: 2.30845 (2.24124) | > loss_feat: 3.25283 (3.18912) | > loss_mel: 19.71826 (19.59364) | > loss_duration: 0.94455 (0.93580) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.54946 (28.00223) | > grad_norm_0: 67.18662 (99.65319) | > loss_disc: 2.58361 (2.58949) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.58361 (2.58949) | > grad_norm_1: 10.36525 (20.96693) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.27710 (10.94542) | > loader_time: 0.01880 (0.01432) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.04580 (2.04580) | > loss_kl: 2.03771 (2.03771) | > loss_feat: 3.32417 (3.32417) | > loss_mel: 20.23190 (20.23190) | > loss_duration: 0.87408 (0.87408) | > loss_0: 28.51367 (28.51367) | > loss_disc: 2.43479 (2.43479) | > loss_1: 2.43479 (2.43479)  --> STEP: 1 | > loss_gen: 2.04531 (2.04531) | > loss_kl: 2.37987 (2.37987) | > loss_feat: 2.95272 (2.95272) | > loss_mel: 19.66192 (19.66192) | > loss_duration: 0.90416 (0.90416) | > loss_0: 27.94398 (27.94398) | > loss_disc: 2.52970 (2.52970) | > loss_1: 2.52970 (2.52970)  --> STEP: 2 | > loss_gen: 2.04619 (2.04575) | > loss_kl: 2.18875 (2.28431) | > loss_feat: 3.08659 (3.01965) | > loss_mel: 18.97460 (19.31826) | > loss_duration: 0.91014 (0.90715) | > loss_0: 27.20627 (27.57513) | > loss_disc: 2.56338 (2.54654) | > loss_1: 2.56338 (2.54654)  --> STEP: 3 | > loss_gen: 2.00324 (2.03158) | > loss_kl: 1.91528 (2.16130) | > loss_feat: 3.33153 (3.12361) | > loss_mel: 18.70139 (19.11264) | > loss_duration: 0.96895 (0.92775) | > loss_0: 26.92038 (27.35688) | > loss_disc: 2.45645 (2.51651) | > loss_1: 2.45645 (2.51651) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00718 (-0.00019) | > avg_loss_gen: 2.03158 (-0.00451) | > avg_loss_kl: 2.16130 (-0.04066) | > avg_loss_feat: 3.12361 (-0.25460) | > avg_loss_mel: 19.11264 (-0.60272) | > avg_loss_duration: 0.92775 (+0.00664) | > avg_loss_0: 27.35688 (-0.89586) | > avg_loss_disc: 2.51651 (-0.18869) | > avg_loss_1: 2.51651 (-0.18869) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481/best_model_8470.pth.tar  > EPOCH: 70/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 16:05:53)   --> STEP: 5/120 -- GLOBAL_STEP: 8475 | > loss_gen: 2.03725 (2.13021) | > loss_kl: 2.11010 (2.29560) | > loss_feat: 3.55619 (3.62880) | > loss_mel: 20.52237 (20.46649) | > loss_duration: 0.89556 (0.90956) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.12148 (29.43067) | > grad_norm_0: 44.59156 (105.34214) | > loss_disc: 2.53225 (2.51866) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.53225 (2.51866) | > grad_norm_1: 12.26614 (14.66054) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.70770 (8.72624) | > loader_time: 0.00930 (0.00719)  --> STEP: 30/120 -- GLOBAL_STEP: 8500 | > loss_gen: 2.10114 (2.08387) | > loss_kl: 2.17427 (2.30556) | > loss_feat: 3.37779 (3.37875) | > loss_mel: 20.04238 (19.91240) | > loss_duration: 0.90796 (0.97856) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.60353 (28.65914) | > grad_norm_0: 28.98596 (94.31116) | > loss_disc: 2.45731 (2.56220) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.45731 (2.56220) | > grad_norm_1: 8.09076 (19.12784) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.95900 (10.17258) | > loader_time: 0.01180 (0.00956)  --> STEP: 55/120 -- GLOBAL_STEP: 8525 | > loss_gen: 2.11523 (2.05712) | > loss_kl: 2.35971 (2.27775) | > loss_feat: 3.12881 (3.27496) | > loss_mel: 19.33477 (19.71270) | > loss_duration: 0.90892 (0.94406) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.84743 (28.26659) | > grad_norm_0: 45.34800 (87.27721) | > loss_disc: 2.57549 (2.58859) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57549 (2.58859) | > grad_norm_1: 13.75596 (18.97540) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.06150 (10.50449) | > loader_time: 0.04100 (0.01166)  --> STEP: 80/120 -- GLOBAL_STEP: 8550 | > loss_gen: 2.19436 (2.04342) | > loss_kl: 2.21811 (2.26425) | > loss_feat: 2.94980 (3.20954) | > loss_mel: 18.98848 (19.59999) | > loss_duration: 0.88354 (0.93817) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.23429 (28.05536) | > grad_norm_0: 136.09729 (90.19343) | > loss_disc: 2.65292 (2.58874) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65292 (2.58874) | > grad_norm_1: 18.82790 (18.33710) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.28690 (10.70583) | > loader_time: 0.01620 (0.01308)  --> STEP: 105/120 -- GLOBAL_STEP: 8575 | > loss_gen: 2.04001 (2.04334) | > loss_kl: 2.18804 (2.24503) | > loss_feat: 2.93966 (3.19007) | > loss_mel: 20.17611 (19.53374) | > loss_duration: 0.97248 (0.93654) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.31629 (27.94871) | > grad_norm_0: 65.71875 (91.47680) | > loss_disc: 2.60953 (2.58741) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60953 (2.58741) | > grad_norm_1: 27.16665 (18.69488) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.21530 (11.00160) | > loader_time: 0.02000 (0.01434) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.22353 (2.22353) | > loss_kl: 2.26621 (2.26621) | > loss_feat: 3.03637 (3.03637) | > loss_mel: 19.81356 (19.81356) | > loss_duration: 0.86793 (0.86793) | > loss_0: 28.20760 (28.20760) | > loss_disc: 2.63059 (2.63059) | > loss_1: 2.63059 (2.63059)  --> STEP: 1 | > loss_gen: 2.26240 (2.26240) | > loss_kl: 2.40796 (2.40796) | > loss_feat: 3.21877 (3.21877) | > loss_mel: 20.27582 (20.27582) | > loss_duration: 0.88288 (0.88288) | > loss_0: 29.04782 (29.04782) | > loss_disc: 2.57870 (2.57870) | > loss_1: 2.57870 (2.57870)  --> STEP: 2 | > loss_gen: 2.33642 (2.29941) | > loss_kl: 2.28385 (2.34590) | > loss_feat: 3.05098 (3.13487) | > loss_mel: 18.63808 (19.45695) | > loss_duration: 0.89752 (0.89020) | > loss_0: 27.20685 (28.12733) | > loss_disc: 2.63610 (2.60740) | > loss_1: 2.63610 (2.60740)  --> STEP: 3 | > loss_gen: 2.24102 (2.27995) | > loss_kl: 2.33720 (2.34300) | > loss_feat: 3.09845 (3.12273) | > loss_mel: 19.19792 (19.37061) | > loss_duration: 0.95225 (0.91088) | > loss_0: 27.82684 (28.02717) | > loss_disc: 2.79538 (2.67006) | > loss_1: 2.79538 (2.67006) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00718 (-0.00000) | > avg_loss_gen: 2.27995 (+0.24837) | > avg_loss_kl: 2.34300 (+0.18170) | > avg_loss_feat: 3.12273 (-0.00088) | > avg_loss_mel: 19.37061 (+0.25797) | > avg_loss_duration: 0.91088 (-0.01687) | > avg_loss_0: 28.02717 (+0.67029) | > avg_loss_disc: 2.67006 (+0.15355) | > avg_loss_1: 2.67006 (+0.15355)  > EPOCH: 71/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 16:28:30)   --> STEP: 9/120 -- GLOBAL_STEP: 8600 | > loss_gen: 1.78531 (2.07934) | > loss_kl: 2.62755 (2.40212) | > loss_feat: 3.24352 (3.50510) | > loss_mel: 19.89505 (20.31823) | > loss_duration: 0.87523 (0.90766) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.42666 (29.21244) | > grad_norm_0: 176.65605 (152.43636) | > loss_disc: 2.68386 (2.63121) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.68386 (2.63121) | > grad_norm_1: 9.23399 (31.76374) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.23330 (10.05137) | > loader_time: 0.00830 (0.00790)  --> STEP: 34/120 -- GLOBAL_STEP: 8625 | > loss_gen: 2.08623 (2.05576) | > loss_kl: 2.23746 (2.34605) | > loss_feat: 3.09647 (3.29999) | > loss_mel: 19.15175 (19.71403) | > loss_duration: 0.94080 (0.96239) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.51271 (28.37822) | > grad_norm_0: 100.70796 (116.07270) | > loss_disc: 2.61931 (2.58362) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.61931 (2.58362) | > grad_norm_1: 18.68735 (18.95186) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.67690 (10.34408) | > loader_time: 0.01210 (0.00997)  --> STEP: 59/120 -- GLOBAL_STEP: 8650 | > loss_gen: 2.12160 (2.05156) | > loss_kl: 2.20035 (2.29352) | > loss_feat: 3.17313 (3.24896) | > loss_mel: 19.20579 (19.60804) | > loss_duration: 0.90373 (0.93882) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.60459 (28.14089) | > grad_norm_0: 66.95567 (105.92409) | > loss_disc: 2.62918 (2.58993) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.62918 (2.58993) | > grad_norm_1: 26.97325 (20.68820) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.94130 (10.49865) | > loader_time: 0.03360 (0.01284)  --> STEP: 84/120 -- GLOBAL_STEP: 8675 | > loss_gen: 1.89531 (2.04968) | > loss_kl: 2.32002 (2.27100) | > loss_feat: 3.64173 (3.22927) | > loss_mel: 19.81268 (19.53927) | > loss_duration: 0.91415 (0.93541) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.58390 (28.02463) | > grad_norm_0: 284.65973 (98.78553) | > loss_disc: 2.56300 (2.57989) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56300 (2.57989) | > grad_norm_1: 24.13090 (20.42019) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.86000 (10.66381) | > loader_time: 0.01660 (0.01367)  --> STEP: 109/120 -- GLOBAL_STEP: 8700 | > loss_gen: 1.76638 (2.05008) | > loss_kl: 2.12797 (2.24719) | > loss_feat: 3.16395 (3.21948) | > loss_mel: 19.66119 (19.48750) | > loss_duration: 0.97075 (0.93617) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.69025 (27.94042) | > grad_norm_0: 180.95822 (107.00158) | > loss_disc: 2.65249 (2.57612) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65249 (2.57612) | > grad_norm_1: 49.29680 (20.84146) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.96990 (10.95457) | > loader_time: 0.02100 (0.01499) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.06358 (2.06358) | > loss_kl: 2.41686 (2.41686) | > loss_feat: 3.33723 (3.33723) | > loss_mel: 19.74964 (19.74964) | > loss_duration: 0.87992 (0.87992) | > loss_0: 28.44724 (28.44724) | > loss_disc: 2.51579 (2.51579) | > loss_1: 2.51579 (2.51579)  --> STEP: 1 | > loss_gen: 2.29526 (2.29526) | > loss_kl: 2.29531 (2.29531) | > loss_feat: 3.31604 (3.31604) | > loss_mel: 19.83203 (19.83203) | > loss_duration: 0.87899 (0.87899) | > loss_0: 28.61763 (28.61763) | > loss_disc: 2.45014 (2.45014) | > loss_1: 2.45014 (2.45014)  --> STEP: 2 | > loss_gen: 2.11690 (2.20608) | > loss_kl: 2.17955 (2.23743) | > loss_feat: 3.18722 (3.25163) | > loss_mel: 18.61797 (19.22500) | > loss_duration: 0.89198 (0.88548) | > loss_0: 26.99363 (27.80563) | > loss_disc: 2.52346 (2.48680) | > loss_1: 2.52346 (2.48680)  --> STEP: 3 | > loss_gen: 2.05388 (2.15535) | > loss_kl: 2.26733 (2.24740) | > loss_feat: 3.16595 (3.22307) | > loss_mel: 18.66849 (19.03950) | > loss_duration: 0.95478 (0.90858) | > loss_0: 27.11043 (27.57390) | > loss_disc: 2.70951 (2.56104) | > loss_1: 2.70951 (2.56104) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00725 (+0.00007) | > avg_loss_gen: 2.15535 (-0.12460) | > avg_loss_kl: 2.24740 (-0.09560) | > avg_loss_feat: 3.22307 (+0.10034) | > avg_loss_mel: 19.03950 (-0.33111) | > avg_loss_duration: 0.90858 (-0.00230) | > avg_loss_0: 27.57390 (-0.45327) | > avg_loss_disc: 2.56104 (-0.10902) | > avg_loss_1: 2.56104 (-0.10902)  > EPOCH: 72/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 16:50:58)   --> STEP: 13/120 -- GLOBAL_STEP: 8725 | > loss_gen: 1.82618 (2.08454) | > loss_kl: 2.27241 (2.44603) | > loss_feat: 3.26108 (3.45043) | > loss_mel: 19.43413 (19.99411) | > loss_duration: 0.92259 (1.06085) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.71639 (29.03596) | > grad_norm_0: 66.56290 (123.59540) | > loss_disc: 2.51560 (2.58350) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.51560 (2.58350) | > grad_norm_1: 21.15201 (23.46136) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.27940 (10.16133) | > loader_time: 0.00940 (0.00825)  --> STEP: 38/120 -- GLOBAL_STEP: 8750 | > loss_gen: 2.18881 (2.08032) | > loss_kl: 2.29736 (2.34683) | > loss_feat: 3.22092 (3.38886) | > loss_mel: 19.99791 (19.75200) | > loss_duration: 0.86880 (0.95528) | > amp_scaler: 1024.00000 (727.57895) | > loss_0: 28.57379 (28.52329) | > grad_norm_0: 167.41647 (107.36154) | > loss_disc: 2.58938 (2.55623) | > amp_scaler-1: 1024.00000 (727.57895) | > loss_1: 2.58938 (2.55623) | > grad_norm_1: 18.56664 (17.59222) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.74340 (10.41303) | > loader_time: 0.01250 (0.01022)  --> STEP: 63/120 -- GLOBAL_STEP: 8775 | > loss_gen: 2.18227 (2.06475) | > loss_kl: 2.05112 (2.28357) | > loss_feat: 3.15890 (3.31642) | > loss_mel: 19.50794 (19.62129) | > loss_duration: 0.92376 (0.93836) | > amp_scaler: 1024.00000 (845.20635) | > loss_0: 27.82399 (28.22440) | > grad_norm_0: 219.58705 (106.40374) | > loss_disc: 2.70416 (2.57341) | > amp_scaler-1: 1024.00000 (845.20635) | > loss_1: 2.70416 (2.57341) | > grad_norm_1: 9.86295 (18.09651) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.30490 (10.58191) | > loader_time: 0.01450 (0.01339)  --> STEP: 88/120 -- GLOBAL_STEP: 8800 | > loss_gen: 2.12248 (2.05511) | > loss_kl: 2.14784 (2.25439) | > loss_feat: 3.09008 (3.27525) | > loss_mel: 19.59715 (19.57268) | > loss_duration: 0.91807 (0.93529) | > amp_scaler: 1024.00000 (896.00000) | > loss_0: 27.87563 (28.09272) | > grad_norm_0: 61.61725 (101.36851) | > loss_disc: 2.53695 (2.57416) | > amp_scaler-1: 1024.00000 (896.00000) | > loss_1: 2.53695 (2.57416) | > grad_norm_1: 7.93414 (17.82380) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.75020 (10.76521) | > loader_time: 0.01780 (0.01422)  --> STEP: 113/120 -- GLOBAL_STEP: 8825 | > loss_gen: 2.22362 (2.05558) | > loss_kl: 2.17337 (2.24088) | > loss_feat: 2.89405 (3.23454) | > loss_mel: 19.76476 (19.48946) | > loss_duration: 0.96819 (0.93731) | > amp_scaler: 1024.00000 (924.31858) | > loss_0: 28.02399 (27.95777) | > grad_norm_0: 43.74953 (102.82735) | > loss_disc: 2.63891 (2.58506) | > amp_scaler-1: 1024.00000 (924.31858) | > loss_1: 2.63891 (2.58506) | > grad_norm_1: 16.20737 (20.39695) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.14290 (11.05580) | > loader_time: 0.02130 (0.01544) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.14091 (2.14091) | > loss_kl: 2.41061 (2.41061) | > loss_feat: 3.54649 (3.54649) | > loss_mel: 20.46880 (20.46880) | > loss_duration: 0.87432 (0.87432) | > loss_0: 29.44113 (29.44113) | > loss_disc: 2.42857 (2.42857) | > loss_1: 2.42857 (2.42857)  --> STEP: 1 | > loss_gen: 2.09767 (2.09767) | > loss_kl: 2.49208 (2.49208) | > loss_feat: 3.15583 (3.15583) | > loss_mel: 18.90294 (18.90294) | > loss_duration: 0.89680 (0.89680) | > loss_0: 27.54532 (27.54532) | > loss_disc: 2.50033 (2.50033) | > loss_1: 2.50033 (2.50033)  --> STEP: 2 | > loss_gen: 2.14682 (2.12225) | > loss_kl: 2.13120 (2.31164) | > loss_feat: 3.22810 (3.19196) | > loss_mel: 19.18787 (19.04541) | > loss_duration: 0.90459 (0.90069) | > loss_0: 27.59859 (27.57195) | > loss_disc: 2.47201 (2.48617) | > loss_1: 2.47201 (2.48617)  --> STEP: 3 | > loss_gen: 2.19143 (2.14531) | > loss_kl: 2.35178 (2.32502) | > loss_feat: 2.86351 (3.08248) | > loss_mel: 19.70619 (19.26567) | > loss_duration: 0.95499 (0.91879) | > loss_0: 28.06791 (27.73727) | > loss_disc: 2.55051 (2.50762) | > loss_1: 2.55051 (2.50762) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00711 (-0.00014) | > avg_loss_gen: 2.14531 (-0.01004) | > avg_loss_kl: 2.32502 (+0.07763) | > avg_loss_feat: 3.08248 (-0.14059) | > avg_loss_mel: 19.26567 (+0.22617) | > avg_loss_duration: 0.91879 (+0.01021) | > avg_loss_0: 27.73727 (+0.16338) | > avg_loss_disc: 2.50762 (-0.05342) | > avg_loss_1: 2.50762 (-0.05342)  > EPOCH: 73/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 17:13:31)   --> STEP: 17/120 -- GLOBAL_STEP: 8850 | > loss_gen: 2.07993 (2.10709) | > loss_kl: 2.57679 (2.35053) | > loss_feat: 3.17247 (3.55559) | > loss_mel: 19.34229 (20.07242) | > loss_duration: 0.88764 (1.02819) | > amp_scaler: 512.00000 (632.47059) | > loss_0: 28.05913 (29.11382) | > grad_norm_0: 63.92445 (96.88579) | > loss_disc: 2.60908 (2.52971) | > amp_scaler-1: 512.00000 (632.47059) | > loss_1: 2.60908 (2.52971) | > grad_norm_1: 5.70996 (11.32432) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.35190 (10.25780) | > loader_time: 0.01010 (0.00887)  --> STEP: 42/120 -- GLOBAL_STEP: 8875 | > loss_gen: 2.31515 (2.07224) | > loss_kl: 2.10585 (2.28774) | > loss_feat: 3.15715 (3.37981) | > loss_mel: 19.28999 (19.72638) | > loss_duration: 0.88527 (0.95492) | > amp_scaler: 512.00000 (560.76190) | > loss_0: 27.75340 (28.42109) | > grad_norm_0: 21.60191 (74.80014) | > loss_disc: 2.53186 (2.55977) | > amp_scaler-1: 512.00000 (560.76190) | > loss_1: 2.53186 (2.55977) | > grad_norm_1: 23.46547 (16.47222) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.22930 (10.46933) | > loader_time: 0.01310 (0.01073)  --> STEP: 67/120 -- GLOBAL_STEP: 8900 | > loss_gen: 1.78052 (2.05813) | > loss_kl: 2.22528 (2.24382) | > loss_feat: 3.23484 (3.30927) | > loss_mel: 20.12458 (19.54904) | > loss_duration: 0.91628 (0.93979) | > amp_scaler: 512.00000 (542.56716) | > loss_0: 28.28150 (28.10004) | > grad_norm_0: 69.11629 (67.40836) | > loss_disc: 2.73433 (2.57078) | > amp_scaler-1: 512.00000 (542.56716) | > loss_1: 2.73433 (2.57078) | > grad_norm_1: 18.36678 (17.12828) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.93690 (10.62204) | > loader_time: 0.01550 (0.01404)  --> STEP: 92/120 -- GLOBAL_STEP: 8925 | > loss_gen: 2.32228 (2.05449) | > loss_kl: 2.13397 (2.23301) | > loss_feat: 3.17604 (3.26272) | > loss_mel: 19.41134 (19.46914) | > loss_duration: 0.92961 (0.93731) | > amp_scaler: 512.00000 (534.26087) | > loss_0: 27.97324 (27.95667) | > grad_norm_0: 167.83128 (75.68670) | > loss_disc: 2.54461 (2.57479) | > amp_scaler-1: 512.00000 (534.26087) | > loss_1: 2.54461 (2.57479) | > grad_norm_1: 19.80987 (18.00550) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.69290 (10.82510) | > loader_time: 0.01850 (0.01477)  --> STEP: 117/120 -- GLOBAL_STEP: 8950 | > loss_gen: 1.75919 (2.04540) | > loss_kl: 2.03055 (2.21375) | > loss_feat: 2.85053 (3.22761) | > loss_mel: 19.25566 (19.42074) | > loss_duration: 1.03699 (0.94111) | > amp_scaler: 512.00000 (529.50427) | > loss_0: 26.93292 (27.84861) | > grad_norm_0: 103.34637 (76.40990) | > loss_disc: 2.68730 (2.57660) | > amp_scaler-1: 512.00000 (529.50427) | > loss_1: 2.68730 (2.57660) | > grad_norm_1: 50.84156 (18.35293) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.26980 (11.12178) | > loader_time: 0.02330 (0.01603) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.40899 (2.40899) | > loss_kl: 2.01671 (2.01671) | > loss_feat: 3.14222 (3.14222) | > loss_mel: 20.15679 (20.15679) | > loss_duration: 0.88161 (0.88161) | > loss_0: 28.60632 (28.60632) | > loss_disc: 2.71109 (2.71109) | > loss_1: 2.71109 (2.71109)  --> STEP: 1 | > loss_gen: 2.25464 (2.25464) | > loss_kl: 2.36403 (2.36403) | > loss_feat: 3.01865 (3.01865) | > loss_mel: 19.56020 (19.56020) | > loss_duration: 0.87190 (0.87190) | > loss_0: 28.06942 (28.06942) | > loss_disc: 2.71123 (2.71123) | > loss_1: 2.71123 (2.71123)  --> STEP: 2 | > loss_gen: 2.21976 (2.23720) | > loss_kl: 2.18267 (2.27335) | > loss_feat: 3.13249 (3.07557) | > loss_mel: 18.60901 (19.08460) | > loss_duration: 0.90169 (0.88680) | > loss_0: 27.04562 (27.55752) | > loss_disc: 2.61128 (2.66125) | > loss_1: 2.61128 (2.66125)  --> STEP: 3 | > loss_gen: 2.18814 (2.22085) | > loss_kl: 2.46630 (2.33767) | > loss_feat: 3.51978 (3.22364) | > loss_mel: 19.84824 (19.33915) | > loss_duration: 0.96130 (0.91163) | > loss_0: 28.98377 (28.03294) | > loss_disc: 2.65207 (2.65819) | > loss_1: 2.65207 (2.65819) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00698 (-0.00013) | > avg_loss_gen: 2.22085 (+0.07554) | > avg_loss_kl: 2.33767 (+0.01264) | > avg_loss_feat: 3.22364 (+0.14116) | > avg_loss_mel: 19.33915 (+0.07348) | > avg_loss_duration: 0.91163 (-0.00716) | > avg_loss_0: 28.03294 (+0.29566) | > avg_loss_disc: 2.65819 (+0.15058) | > avg_loss_1: 2.65819 (+0.15058)  > EPOCH: 74/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 17:36:07)   --> STEP: 21/120 -- GLOBAL_STEP: 8975 | > loss_gen: 1.68885 (2.03808) | > loss_kl: 2.52535 (2.34235) | > loss_feat: 2.90317 (3.34058) | > loss_mel: 19.60501 (19.99205) | > loss_duration: 0.87579 (0.99949) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.59817 (28.71255) | > grad_norm_0: 54.28123 (98.42839) | > loss_disc: 2.67723 (2.62761) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.67723 (2.62761) | > grad_norm_1: 10.28093 (14.42428) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.44250 (10.26984) | > loader_time: 0.01080 (0.00888)  --> STEP: 46/120 -- GLOBAL_STEP: 9000 | > loss_gen: 1.92275 (2.02480) | > loss_kl: 2.24906 (2.29025) | > loss_feat: 2.93214 (3.20742) | > loss_mel: 18.91464 (19.59874) | > loss_duration: 0.87750 (0.94811) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.89609 (28.06932) | > grad_norm_0: 152.59918 (85.38102) | > loss_disc: 2.66762 (2.60993) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.66762 (2.60993) | > grad_norm_1: 27.54343 (16.99079) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.91520 (10.47248) | > loader_time: 0.01360 (0.01067)  --> STEP: 71/120 -- GLOBAL_STEP: 9025 | > loss_gen: 1.92796 (2.03516) | > loss_kl: 2.21443 (2.27202) | > loss_feat: 3.42137 (3.23137) | > loss_mel: 19.10521 (19.49230) | > loss_duration: 0.95372 (0.93894) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.62269 (27.96978) | > grad_norm_0: 129.75327 (100.69865) | > loss_disc: 2.49560 (2.58631) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.49560 (2.58631) | > grad_norm_1: 12.13792 (17.34525) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.23610 (10.61259) | > loader_time: 0.01590 (0.01264)  --> STEP: 96/120 -- GLOBAL_STEP: 9050 | > loss_gen: 2.24698 (2.03830) | > loss_kl: 2.13143 (2.25574) | > loss_feat: 3.55735 (3.21827) | > loss_mel: 19.58450 (19.45499) | > loss_duration: 0.94682 (0.93615) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.46708 (27.90345) | > grad_norm_0: 80.71474 (95.24448) | > loss_disc: 2.39017 (2.58686) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.39017 (2.58686) | > grad_norm_1: 10.34940 (17.78435) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.86590 (10.83093) | > loader_time: 0.01930 (0.01389) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.97877 (1.97877) | > loss_kl: 2.25973 (2.25973) | > loss_feat: 3.08112 (3.08112) | > loss_mel: 19.90284 (19.90284) | > loss_duration: 0.87998 (0.87998) | > loss_0: 28.10245 (28.10245) | > loss_disc: 2.52747 (2.52747) | > loss_1: 2.52747 (2.52747)  --> STEP: 1 | > loss_gen: 1.99122 (1.99122) | > loss_kl: 2.34840 (2.34840) | > loss_feat: 3.09958 (3.09958) | > loss_mel: 19.16540 (19.16540) | > loss_duration: 0.88170 (0.88170) | > loss_0: 27.48631 (27.48631) | > loss_disc: 2.58326 (2.58326) | > loss_1: 2.58326 (2.58326)  --> STEP: 2 | > loss_gen: 1.97363 (1.98243) | > loss_kl: 2.21194 (2.28017) | > loss_feat: 3.23767 (3.16863) | > loss_mel: 19.56929 (19.36735) | > loss_duration: 0.90355 (0.89263) | > loss_0: 27.89608 (27.69119) | > loss_disc: 2.46269 (2.52297) | > loss_1: 2.46269 (2.52297)  --> STEP: 3 | > loss_gen: 1.88027 (1.94838) | > loss_kl: 2.12896 (2.22977) | > loss_feat: 3.08485 (3.14070) | > loss_mel: 19.68795 (19.47421) | > loss_duration: 0.95299 (0.91275) | > loss_0: 27.73502 (27.70580) | > loss_disc: 2.49564 (2.51386) | > loss_1: 2.49564 (2.51386) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00717 (+0.00019) | > avg_loss_gen: 1.94838 (-0.27247) | > avg_loss_kl: 2.22977 (-0.10790) | > avg_loss_feat: 3.14070 (-0.08294) | > avg_loss_mel: 19.47421 (+0.13506) | > avg_loss_duration: 0.91275 (+0.00111) | > avg_loss_0: 27.70580 (-0.32713) | > avg_loss_disc: 2.51386 (-0.14433) | > avg_loss_1: 2.51386 (-0.14433)  > EPOCH: 75/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 17:58:35)   --> STEP: 0/120 -- GLOBAL_STEP: 9075 | > loss_gen: 1.89012 (1.89012) | > loss_kl: 2.26374 (2.26374) | > loss_feat: 3.58956 (3.58956) | > loss_mel: 19.96317 (19.96317) | > loss_duration: 0.89682 (0.89682) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.60341 (28.60341) | > grad_norm_0: 103.14755 (103.14755) | > loss_disc: 2.49304 (2.49304) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.49304 (2.49304) | > grad_norm_1: 7.07428 (7.07428) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.95870 (10.95868) | > loader_time: 0.60610 (0.60610)  --> STEP: 25/120 -- GLOBAL_STEP: 9100 | > loss_gen: 1.80123 (2.04706) | > loss_kl: 2.17077 (2.36963) | > loss_feat: 3.04986 (3.32988) | > loss_mel: 19.57242 (19.89811) | > loss_duration: 0.91049 (0.98693) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.50477 (28.63160) | > grad_norm_0: 166.41330 (144.64784) | > loss_disc: 2.75251 (2.60139) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.75251 (2.60139) | > grad_norm_1: 25.25895 (19.94537) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.29010 (10.16654) | > loader_time: 0.01130 (0.00926)  --> STEP: 50/120 -- GLOBAL_STEP: 9125 | > loss_gen: 1.97741 (2.05070) | > loss_kl: 2.40119 (2.33127) | > loss_feat: 3.37895 (3.28189) | > loss_mel: 18.77143 (19.60696) | > loss_duration: 0.89403 (0.94289) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.42301 (28.21370) | > grad_norm_0: 64.08314 (104.14485) | > loss_disc: 2.48144 (2.59315) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.48144 (2.59315) | > grad_norm_1: 16.99170 (19.53867) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.54390 (10.35967) | > loader_time: 0.01380 (0.01094)  --> STEP: 75/120 -- GLOBAL_STEP: 9150 | > loss_gen: 1.91304 (2.05048) | > loss_kl: 2.38269 (2.30063) | > loss_feat: 3.44428 (3.27245) | > loss_mel: 19.22985 (19.48878) | > loss_duration: 0.92157 (0.93685) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.89143 (28.04918) | > grad_norm_0: 75.63818 (104.26100) | > loss_disc: 2.55268 (2.57964) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55268 (2.57964) | > grad_norm_1: 25.46993 (17.21346) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.78430 (10.50654) | > loader_time: 0.01610 (0.01326)  --> STEP: 100/120 -- GLOBAL_STEP: 9175 | > loss_gen: 1.98040 (2.04636) | > loss_kl: 2.26264 (2.27909) | > loss_feat: 3.02988 (3.24682) | > loss_mel: 18.52188 (19.39429) | > loss_duration: 0.92365 (0.93502) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.71844 (27.90158) | > grad_norm_0: 36.67223 (94.31763) | > loss_disc: 2.66682 (2.58244) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.66682 (2.58244) | > grad_norm_1: 35.64050 (18.81159) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.75660 (10.76883) | > loader_time: 0.01980 (0.01440) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.15774 (2.15774) | > loss_kl: 2.09042 (2.09042) | > loss_feat: 3.54572 (3.54572) | > loss_mel: 19.94838 (19.94838) | > loss_duration: 0.86400 (0.86400) | > loss_0: 28.60625 (28.60625) | > loss_disc: 2.59863 (2.59863) | > loss_1: 2.59863 (2.59863)  --> STEP: 1 | > loss_gen: 2.18285 (2.18285) | > loss_kl: 2.38678 (2.38678) | > loss_feat: 3.24896 (3.24896) | > loss_mel: 19.64690 (19.64690) | > loss_duration: 0.87430 (0.87430) | > loss_0: 28.33979 (28.33979) | > loss_disc: 2.59769 (2.59769) | > loss_1: 2.59769 (2.59769)  --> STEP: 2 | > loss_gen: 2.05461 (2.11873) | > loss_kl: 2.42133 (2.40405) | > loss_feat: 3.01129 (3.13013) | > loss_mel: 18.90252 (19.27471) | > loss_duration: 0.89983 (0.88706) | > loss_0: 27.28957 (27.81468) | > loss_disc: 2.65703 (2.62736) | > loss_1: 2.65703 (2.62736)  --> STEP: 3 | > loss_gen: 2.10317 (2.11354) | > loss_kl: 2.29818 (2.36876) | > loss_feat: 3.28407 (3.18144) | > loss_mel: 18.74440 (19.09794) | > loss_duration: 0.95455 (0.90956) | > loss_0: 27.38437 (27.67124) | > loss_disc: 2.51411 (2.58961) | > loss_1: 2.51411 (2.58961) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00730 (+0.00013) | > avg_loss_gen: 2.11354 (+0.16517) | > avg_loss_kl: 2.36876 (+0.13899) | > avg_loss_feat: 3.18144 (+0.04074) | > avg_loss_mel: 19.09794 (-0.37627) | > avg_loss_duration: 0.90956 (-0.00319) | > avg_loss_0: 27.67124 (-0.03456) | > avg_loss_disc: 2.58961 (+0.07575) | > avg_loss_1: 2.58961 (+0.07575)  > EPOCH: 76/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 18:20:52)   --> STEP: 4/120 -- GLOBAL_STEP: 9200 | > loss_gen: 2.23098 (2.08052) | > loss_kl: 2.70137 (2.47292) | > loss_feat: 3.59212 (3.57786) | > loss_mel: 20.04929 (19.88292) | > loss_duration: 0.90675 (0.90449) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.48051 (28.91870) | > grad_norm_0: 46.41402 (119.03735) | > loss_disc: 2.46985 (2.47658) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.46985 (2.47658) | > grad_norm_1: 16.90542 (13.82996) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.91590 (9.73047) | > loader_time: 0.00750 (0.00743)  --> STEP: 29/120 -- GLOBAL_STEP: 9225 | > loss_gen: 1.85108 (2.09318) | > loss_kl: 2.34666 (2.36210) | > loss_feat: 3.11733 (3.47530) | > loss_mel: 18.76708 (19.55530) | > loss_duration: 0.85838 (0.97444) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 26.94053 (28.46033) | > grad_norm_0: 88.56563 (117.51724) | > loss_disc: 2.65986 (2.54691) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.65986 (2.54691) | > grad_norm_1: 20.65182 (15.50011) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.38560 (10.21461) | > loader_time: 0.01150 (0.00971)  --> STEP: 54/120 -- GLOBAL_STEP: 9250 | > loss_gen: 1.98067 (2.05900) | > loss_kl: 2.27388 (2.31385) | > loss_feat: 3.16628 (3.34213) | > loss_mel: 19.90801 (19.45962) | > loss_duration: 0.94555 (0.94117) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 28.27439 (28.11577) | > grad_norm_0: 42.97716 (98.78037) | > loss_disc: 2.60196 (2.58030) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.60196 (2.58030) | > grad_norm_1: 10.30972 (19.62407) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.76770 (10.40500) | > loader_time: 0.01380 (0.01131)  --> STEP: 79/120 -- GLOBAL_STEP: 9275 | > loss_gen: 1.96856 (2.05637) | > loss_kl: 2.27596 (2.27803) | > loss_feat: 3.01828 (3.28601) | > loss_mel: 19.01806 (19.39239) | > loss_duration: 0.94456 (0.93628) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.22542 (27.94906) | > grad_norm_0: 138.11533 (98.31766) | > loss_disc: 2.56292 (2.58362) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56292 (2.58362) | > grad_norm_1: 9.04879 (19.88353) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.77650 (10.55491) | > loader_time: 0.01630 (0.01373)  --> STEP: 104/120 -- GLOBAL_STEP: 9300 | > loss_gen: 1.86246 (2.04961) | > loss_kl: 2.24518 (2.26686) | > loss_feat: 3.13644 (3.25727) | > loss_mel: 18.96600 (19.36832) | > loss_duration: 0.92061 (0.93386) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.13069 (27.87591) | > grad_norm_0: 89.09445 (101.15085) | > loss_disc: 2.69689 (2.58743) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.69689 (2.58743) | > grad_norm_1: 13.90445 (20.37231) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 11.95170 (10.83549) | > loader_time: 0.01890 (0.01484) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 2.22530 (2.22530) | > loss_kl: 2.26937 (2.26937) | > loss_feat: 3.39161 (3.39161) | > loss_mel: 19.59538 (19.59538) | > loss_duration: 0.88234 (0.88234) | > loss_0: 28.36399 (28.36399) | > loss_disc: 2.51565 (2.51565) | > loss_1: 2.51565 (2.51565)  --> STEP: 1 | > loss_gen: 2.04955 (2.04955) | > loss_kl: 2.23651 (2.23651) | > loss_feat: 3.26044 (3.26044) | > loss_mel: 19.56125 (19.56125) | > loss_duration: 0.89490 (0.89490) | > loss_0: 28.00265 (28.00265) | > loss_disc: 2.57357 (2.57357) | > loss_1: 2.57357 (2.57357)  --> STEP: 2 | > loss_gen: 2.08021 (2.06488) | > loss_kl: 2.25608 (2.24629) | > loss_feat: 2.98364 (3.12204) | > loss_mel: 19.21659 (19.38892) | > loss_duration: 0.91043 (0.90267) | > loss_0: 27.44694 (27.72480) | > loss_disc: 2.59261 (2.58309) | > loss_1: 2.59261 (2.58309)  --> STEP: 3 | > loss_gen: 2.04016 (2.05664) | > loss_kl: 2.20845 (2.23368) | > loss_feat: 2.90846 (3.05084) | > loss_mel: 19.03242 (19.27009) | > loss_duration: 0.96697 (0.92410) | > loss_0: 27.15646 (27.53535) | > loss_disc: 2.73444 (2.63354) | > loss_1: 2.73444 (2.63354) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00649 (-0.00080) | > avg_loss_gen: 2.05664 (-0.05691) | > avg_loss_kl: 2.23368 (-0.13508) | > avg_loss_feat: 3.05084 (-0.13060) | > avg_loss_mel: 19.27009 (+0.17215) | > avg_loss_duration: 0.92410 (+0.01454) | > avg_loss_0: 27.53535 (-0.13590) | > avg_loss_disc: 2.63354 (+0.04393) | > avg_loss_1: 2.63354 (+0.04393)  > EPOCH: 77/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 18:43:12)   --> STEP: 8/120 -- GLOBAL_STEP: 9325 | > loss_gen: 2.24695 (2.09826) | > loss_kl: 2.47442 (2.43060) | > loss_feat: 3.49173 (3.53686) | > loss_mel: 20.09172 (19.96956) | > loss_duration: 0.92229 (0.90559) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 29.22711 (28.94088) | > grad_norm_0: 186.44612 (90.94397) | > loss_disc: 2.55532 (2.56526) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.55532 (2.56526) | > grad_norm_1: 10.84339 (13.43078) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 9.94210 (10.00392) | > loader_time: 0.00860 (0.00766)  --> STEP: 33/120 -- GLOBAL_STEP: 9350 | > loss_gen: 1.66605 (2.04434) | > loss_kl: 2.28064 (2.36619) | > loss_feat: 3.09567 (3.33614) | > loss_mel: 19.14675 (19.53652) | > loss_duration: 0.89039 (0.96193) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.07951 (28.24512) | > grad_norm_0: 28.71897 (72.02590) | > loss_disc: 2.59445 (2.57757) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.59445 (2.57757) | > grad_norm_1: 46.25107 (14.29977) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.71130 (10.29936) | > loader_time: 0.01190 (0.01005)  --> STEP: 58/120 -- GLOBAL_STEP: 9375 | > loss_gen: 1.81290 (2.03846) | > loss_kl: 2.32308 (2.30759) | > loss_feat: 2.98582 (3.26876) | > loss_mel: 19.26315 (19.40539) | > loss_duration: 0.89519 (0.93692) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.28014 (27.95712) | > grad_norm_0: 32.95920 (72.28327) | > loss_disc: 2.63783 (2.58915) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.63783 (2.58915) | > grad_norm_1: 5.44274 (16.20371) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.68440 (10.47317) | > loader_time: 0.01500 (0.01230)  --> STEP: 83/120 -- GLOBAL_STEP: 9400 | > loss_gen: 2.27680 (2.04255) | > loss_kl: 2.06455 (2.27513) | > loss_feat: 3.32151 (3.24997) | > loss_mel: 18.96553 (19.38188) | > loss_duration: 0.91945 (0.93334) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.54784 (27.88286) | > grad_norm_0: 136.64250 (84.46626) | > loss_disc: 2.56754 (2.58302) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.56754 (2.58302) | > grad_norm_1: 12.57348 (16.69390) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.96700 (10.63361) | > loader_time: 0.01690 (0.01330)  --> STEP: 108/120 -- GLOBAL_STEP: 9425 | > loss_gen: 2.15700 (2.04395) | > loss_kl: 2.15692 (2.25949) | > loss_feat: 3.13554 (3.25303) | > loss_mel: 19.05659 (19.33250) | > loss_duration: 0.96305 (0.93408) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.46909 (27.82306) | > grad_norm_0: 152.32901 (85.55050) | > loss_disc: 2.57812 (2.57900) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.57812 (2.57900) | > grad_norm_1: 13.69328 (16.42164) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 12.00710 (10.92013) | > loader_time: 0.02030 (0.01460) > DataLoader initialization | > Use phonemes: False | > Number of instances : 58 | > Max length sequence: 139203.0 | > Min length sequence: 36179.0 | > Avg length sequence: 82500.0172413793 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 0.  > EVALUATION   --> STEP: 0 | > loss_gen: 1.89753 (1.89753) | > loss_kl: 2.13437 (2.13437) | > loss_feat: 3.54596 (3.54596) | > loss_mel: 20.15965 (20.15965) | > loss_duration: 0.87330 (0.87330) | > loss_0: 28.61081 (28.61081) | > loss_disc: 2.70988 (2.70988) | > loss_1: 2.70988 (2.70988)  --> STEP: 1 | > loss_gen: 1.89657 (1.89657) | > loss_kl: 2.58333 (2.58333) | > loss_feat: 3.17431 (3.17431) | > loss_mel: 19.50044 (19.50044) | > loss_duration: 0.87906 (0.87906) | > loss_0: 28.03372 (28.03372) | > loss_disc: 2.64069 (2.64069) | > loss_1: 2.64069 (2.64069)  --> STEP: 2 | > loss_gen: 2.08465 (1.99061) | > loss_kl: 2.04984 (2.31659) | > loss_feat: 3.15300 (3.16366) | > loss_mel: 19.63938 (19.56991) | > loss_duration: 0.90290 (0.89098) | > loss_0: 27.82976 (27.93174) | > loss_disc: 2.52589 (2.58329) | > loss_1: 2.52589 (2.58329)  --> STEP: 3 | > loss_gen: 1.90811 (1.96311) | > loss_kl: 2.30698 (2.31338) | > loss_feat: 3.11434 (3.14722) | > loss_mel: 18.89098 (19.34360) | > loss_duration: 0.95058 (0.91085) | > loss_0: 27.17100 (27.67816) | > loss_disc: 2.73960 (2.63540) | > loss_1: 2.73960 (2.63540) | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00725 (+0.00076) | > avg_loss_gen: 1.96311 (-0.09353) | > avg_loss_kl: 2.31338 (+0.07971) | > avg_loss_feat: 3.14722 (+0.09637) | > avg_loss_mel: 19.34360 (+0.07351) | > avg_loss_duration: 0.91085 (-0.01325) | > avg_loss_0: 27.67816 (+0.14281) | > avg_loss_disc: 2.63540 (+0.00185) | > avg_loss_1: 2.63540 (+0.00185)  > EPOCH: 78/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 5798 | > Max length sequence: 183618.0 | > Min length sequence: 24483.0 | > Avg length sequence: 82634.87443946188 | > Num. instances discarded by max-min (max=500000, min=0) seq limits: 0 | > Batch group size: 240.  > TRAINING (2022-01-26 19:05:36)   --> STEP: 12/120 -- GLOBAL_STEP: 9450 | > loss_gen: 1.75315 (2.06112) | > loss_kl: 2.36402 (2.40848) | > loss_feat: 3.47702 (3.51243) | > loss_mel: 19.50994 (19.92424) | > loss_duration: 0.88138 (1.06581) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.98553 (28.97207) | > grad_norm_0: 161.16425 (92.97960) | > loss_disc: 2.52962 (2.57423) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.52962 (2.57423) | > grad_norm_1: 12.71873 (17.70251) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.33860 (10.07900) | > loader_time: 0.00910 (0.00784)  --> STEP: 37/120 -- GLOBAL_STEP: 9475 | > loss_gen: 1.99507 (2.06911) | > loss_kl: 2.09598 (2.31172) | > loss_feat: 3.20806 (3.39922) | > loss_mel: 19.24780 (19.63639) | > loss_duration: 0.91469 (0.95582) | > amp_scaler: 512.00000 (512.00000) | > loss_0: 27.46160 (28.37226) | > grad_norm_0: 29.04408 (101.21238) | > loss_disc: 2.64072 (2.56433) | > amp_scaler-1: 512.00000 (512.00000) | > loss_1: 2.64072 (2.56433) | > grad_norm_1: 16.48812 (13.53958) | > current_lr_0: 0.00020 | > current_lr_1: 0.00020 | > step_time: 10.43920 (10.31539) | > loader_time: 0.01210 (0.01002) ! Run is kept in /media/popos/Barracuda/Models/TTS_new/trained/vits_ljspeech-January-25-2022_01+52PM-c63bb481