atharva253 commited on
Commit
af006fc
1 Parent(s): 66bc07b

Update model

Browse files
Files changed (16) hide show
  1. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/95epoch.pth +3 -0
  2. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/RESULTS.md +5 -5
  3. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/config.yaml +1 -1
  4. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/backward_time.png +0 -0
  5. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/clip.png +0 -0
  6. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/forward_time.png +0 -0
  7. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/gpu_max_cached_mem_GB.png +0 -0
  8. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/grad_norm.png +0 -0
  9. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/iter_time.png +0 -0
  10. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/loss.png +0 -0
  11. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/loss_scale.png +0 -0
  12. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/optim0_lr0.png +0 -0
  13. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/optim_step_time.png +0 -0
  14. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/si_snr_loss.png +0 -0
  15. exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/train_time.png +0 -0
  16. meta.yaml +2 -2
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/95epoch.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c889dab0d2819f4a38710d74f2f01544e0af04859f9942ac20d21e33c2df8af
3
+ size 28244094
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/RESULTS.md CHANGED
@@ -1,12 +1,12 @@
1
  <!-- Generated by ./scripts/utils/show_enh_score.sh -->
2
  # RESULTS
3
  ## Environments
4
- - date: `Wed Mar 20 10:56:43 EDT 2024`
5
  - python version: `3.9.18 (main, Sep 11 2023, 13:41:44) [GCC 11.2.0]`
6
  - espnet version: `espnet 202402`
7
  - pytorch version: `pytorch 2.1.0`
8
- - Git hash: `2bcd4624d49bf3cc682796470ba7fd149bc03fa4`
9
- - Commit date: `Tue Mar 19 11:00:42 2024 -0400`
10
 
11
 
12
  ## enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw
@@ -15,6 +15,6 @@ config: conf/tuning/train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_12
15
 
16
  |dataset|STOI|SAR|SDR|SIR|SI_SNR|
17
  |---|---|---|---|---|---|
18
- |enhanced_cv|83.40|9.90|9.30|20.08|8.74|
19
- |enhanced_tt|86.66|10.59|9.91|19.83|9.40|
20
 
 
1
  <!-- Generated by ./scripts/utils/show_enh_score.sh -->
2
  # RESULTS
3
  ## Environments
4
+ - date: `Mon Apr 22 17:21:05 EDT 2024`
5
  - python version: `3.9.18 (main, Sep 11 2023, 13:41:44) [GCC 11.2.0]`
6
  - espnet version: `espnet 202402`
7
  - pytorch version: `pytorch 2.1.0`
8
+ - Git hash: `37828ea9708cd2f541220fdfe180457c7f7d67f1`
9
+ - Commit date: `Thu Mar 21 22:52:57 2024 -0400`
10
 
11
 
12
  ## enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw
 
15
 
16
  |dataset|STOI|SAR|SDR|SIR|SI_SNR|
17
  |---|---|---|---|---|---|
18
+ |enhanced_cv|85.97|10.51|10.07|21.63|9.61|
19
+ |enhanced_tt|88.76|11.22|10.69|21.36|10.26|
20
 
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/config.yaml CHANGED
@@ -16,7 +16,7 @@ dist_world_size: 4
16
  dist_rank: 0
17
  local_rank: 0
18
  dist_master_addr: localhost
19
- dist_master_port: 36443
20
  dist_launcher: null
21
  multiprocessing_distributed: true
22
  unused_parameters: false
 
16
  dist_rank: 0
17
  local_rank: 0
18
  dist_master_addr: localhost
19
+ dist_master_port: 45443
20
  dist_launcher: null
21
  multiprocessing_distributed: true
22
  unused_parameters: false
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/backward_time.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/clip.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/forward_time.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/gpu_max_cached_mem_GB.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/grad_norm.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/iter_time.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/loss.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/loss_scale.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/optim0_lr0.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/optim_step_time.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/si_snr_loss.png CHANGED
exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/images/train_time.png CHANGED
meta.yaml CHANGED
@@ -1,8 +1,8 @@
1
  espnet: '202402'
2
  files:
3
- model_file: exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/39epoch.pth
4
  python: "3.9.18 (main, Sep 11 2023, 13:41:44) \n[GCC 11.2.0]"
5
- timestamp: 1710948089.685135
6
  torch: 2.1.0
7
  yaml_files:
8
  train_config: exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/config.yaml
 
1
  espnet: '202402'
2
  files:
3
+ model_file: exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/95epoch.pth
4
  python: "3.9.18 (main, Sep 11 2023, 13:41:44) \n[GCC 11.2.0]"
5
+ timestamp: 1713830406.80243
6
  torch: 2.1.0
7
  yaml_files:
8
  train_config: exp/enh_train_enh_tfgridnetv2_tf_lr-patience3_patience5_I_1_J_1_D_128_batch_8_raw/config.yaml