karthik commited on
Commit
79fdc0a
β€’
1 Parent(s): d1638d0
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. exp/asr_train_asr_raw_en_word/126epoch.pth +0 -3
  2. exp/asr_train_asr_raw_en_word/137epoch.pth +0 -3
  3. exp/asr_train_asr_raw_en_word/154epoch.pth +0 -3
  4. exp/asr_train_asr_raw_en_word/174epoch.pth +0 -3
  5. exp/asr_train_asr_raw_en_word/192epoch.pth +0 -3
  6. exp/asr_train_asr_raw_en_word/195epoch.pth +0 -3
  7. exp/asr_train_asr_raw_en_word/197epoch.pth +0 -3
  8. exp/asr_train_asr_raw_en_word/198epoch.pth +0 -3
  9. exp/asr_train_asr_raw_en_word/199epoch.pth +0 -3
  10. exp/asr_train_asr_raw_en_word/200epoch.pth +0 -3
  11. exp/asr_train_asr_raw_en_word/66epoch.pth +0 -3
  12. exp/asr_train_asr_raw_en_word/67epoch.pth +0 -3
  13. exp/asr_train_asr_raw_en_word/68epoch.pth +0 -3
  14. exp/asr_train_asr_raw_en_word/70epoch.pth +0 -3
  15. exp/asr_train_asr_raw_en_word/71epoch.pth +0 -3
  16. exp/asr_train_asr_raw_en_word/train.1.log +0 -0
  17. exp/asr_train_asr_raw_en_word/train.2.log +0 -557
  18. exp/asr_train_asr_raw_en_word/train.acc.ave.pth +0 -1
  19. exp/asr_train_asr_raw_en_word/train.acc.ave_1best.pth +0 -1
  20. exp/asr_train_asr_raw_en_word/train.acc.ave_5best.pth +0 -3
  21. exp/asr_train_asr_raw_en_word/train.acc.best.pth +0 -1
  22. exp/asr_train_asr_raw_en_word/train.log +0 -0
  23. exp/asr_train_asr_raw_en_word/train.loss.ave.pth +0 -1
  24. exp/asr_train_asr_raw_en_word/train.loss.ave_1best.pth +0 -1
  25. exp/asr_train_asr_raw_en_word/train.loss.ave_5best.pth +0 -3
  26. exp/asr_train_asr_raw_en_word/train.loss.best.pth +0 -1
  27. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/RESULTS.md +0 -0
  28. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.100ep.png +0 -0
  29. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.101ep.png +0 -0
  30. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.102ep.png +0 -0
  31. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.103ep.png +0 -0
  32. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.104ep.png +0 -0
  33. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.105ep.png +0 -0
  34. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.106ep.png +0 -0
  35. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.107ep.png +0 -0
  36. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.108ep.png +0 -0
  37. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.109ep.png +0 -0
  38. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.10ep.png +0 -0
  39. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.110ep.png +0 -0
  40. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.111ep.png +0 -0
  41. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.112ep.png +0 -0
  42. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.113ep.png +0 -0
  43. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.114ep.png +0 -0
  44. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.115ep.png +0 -0
  45. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.116ep.png +0 -0
  46. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.117ep.png +0 -0
  47. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.118ep.png +0 -0
  48. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.119ep.png +0 -0
  49. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.11ep.png +0 -0
  50. exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.120ep.png +0 -0
exp/asr_train_asr_raw_en_word/126epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:524df69c03273cdd0593cc5b62f2c82b2b18b86853a9390252f255f5713b4fba
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/137epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c89eb2c81fd8181534ddff02232ee375ccd26f28ab7a7831344e45c8cbfbc7a
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/154epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:99cfe5a5d6abdd10797e678903097cbb28a7d7734d047c44d1ebc1d06b14fb87
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/174epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:deee35c14d5235e2b506147b84b0889071fb45ee13c8aa185a24fd98ce4d5b25
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/192epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:be20c8f3c0d747b7597ea0f4dee8f4aa1ef30329179ecec20cbdce4b24ee865e
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/195epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ba689f98c1511c7dbbbf79841b918d7b52831197ee439176c4526344eb97bff
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/197epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a94a5974c14fe9f71f5be435473bac0059da139d1901f42a11e41fa6166f8485
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/198epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bf19a35ffb6767da69e0f7fdd562d362fd2bbd99de0171b5c00c4de0f49bff4
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/199epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7398df97ab7cd040dd143e7d863502a4617ccfe6a914fef59d78ca7887fb8745
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/200epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e05edb01164041f1cc97611a5fcc69a192e4cf26eec351f06ba21da40b702133
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/66epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d01c3e6b22c6a3ea3a69d1cbf4a9fcd2f331bc52c2a43ab857326f1dd8213c74
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/67epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb468260a5db3ed6bb528e94fc27b67e750d99d37482c5d01d25a5242cdd2607
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/68epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b41e2bc1ba70e18bb8982e54332df7ce3b25f2ae74aed3238ea7ac22a4bf0bc
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/70epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:839f0de480a2fb929682d232346c2f1b996991979a260c4abc03443224dd11d7
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/71epoch.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e986ca9e9d613f4a2ab154f573a8d229dee39923b28ee1e5bda676d4c1366fb
3
- size 110488525
 
 
 
 
exp/asr_train_asr_raw_en_word/train.1.log DELETED
The diff for this file is too large to render. See raw diff
 
exp/asr_train_asr_raw_en_word/train.2.log DELETED
@@ -1,557 +0,0 @@
1
- # Running on v030.ib.bridges2.psc.edu
2
- # Started at Mon Jan 31 01:55:52 EST 2022
3
- # SLURMD_NODENAME=v030
4
- # SLURM_CLUSTER_NAME=bridges2
5
- # SLURM_CONF=/var/spool/slurm/d/conf-cache/slurm.conf
6
- # SLURM_CPUS_ON_NODE=5
7
- # SLURM_CPUS_PER_TASK=1
8
- # SLURM_EXPORT_ENV=PATH
9
- # SLURM_GET_USER_ENV=1
10
- # SLURM_GTIDS=0
11
- # SLURM_JOBID=6473234
12
- # SLURM_JOB_ACCOUNT=cis210027p
13
- # SLURM_JOB_CPUS_PER_NODE=5
14
- # SLURM_JOB_GID=24886
15
- # SLURM_JOB_GPUS=5
16
- # SLURM_JOB_ID=6473234
17
- # SLURM_JOB_NAME=exp/asr_train_asr_raw_en_word/train.log
18
- # SLURM_JOB_NODELIST=v030
19
- # SLURM_JOB_NUM_NODES=1
20
- # SLURM_JOB_PARTITION=GPU-shared
21
- # SLURM_JOB_QOS=gpu
22
- # SLURM_JOB_UID=82326
23
- # SLURM_JOB_USER=ganesank
24
- # SLURM_LOCALID=0
25
- # SLURM_NNODES=1
26
- # SLURM_NODEID=0
27
- # SLURM_NODELIST=v030
28
- # SLURM_NODE_ALIASES='(null)'
29
- # SLURM_OPEN_MODE=a
30
- # SLURM_PRIO_PROCESS=0
31
- # SLURM_PROCID=0
32
- # SLURM_SUBMIT_DIR=/ocean/projects/cis210027p/ganesank/karthik_new/espnet/egs2/dstc2/asr2
33
- # SLURM_SUBMIT_HOST=br012.ib.bridges2.psc.edu
34
- # SLURM_TASKS_PER_NODE=5
35
- # SLURM_TASK_PID=54262
36
- # SLURM_TOPOLOGY_ADDR=v030
37
- # SLURM_TOPOLOGY_ADDR_PATTERN=node
38
- # SLURM_WORKING_CLUSTER=bridges2:br003:6814:9216:109
39
- # python3 -m espnet2.bin.asr_train --use_preprocessor true --bpemodel none --token_type word --token_list data/en_token_list/word/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/valid/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/valid/text,text,text --valid_shape_file exp/asr_stats_raw_en_word/valid/speech_shape --valid_shape_file exp/asr_stats_raw_en_word/valid/text_shape.word --resume true --fold_length 80000 --fold_length 150 --output_dir exp/asr_train_asr_raw_en_word --config conf/train_asr.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz --train_data_path_and_name_and_type dump/raw/train/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train/text,text,text --train_shape_file exp/asr_stats_raw_en_word/train/speech_shape --train_shape_file exp/asr_stats_raw_en_word/train/text_shape.word --ngpu 1 --multiprocessing_distributed True
40
- /ocean/projects/cis210027p/ganesank/karthik_new/espnet/tools/venv/bin/python3 /ocean/projects/cis210027p/ganesank/karthik_new/espnet/espnet2/bin/asr_train.py --use_preprocessor true --bpemodel none --token_type word --token_list data/en_token_list/word/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/valid/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/valid/text,text,text --valid_shape_file exp/asr_stats_raw_en_word/valid/speech_shape --valid_shape_file exp/asr_stats_raw_en_word/valid/text_shape.word --resume true --fold_length 80000 --fold_length 150 --output_dir exp/asr_train_asr_raw_en_word --config conf/train_asr.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz --train_data_path_and_name_and_type dump/raw/train/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train/text,text,text --train_shape_file exp/asr_stats_raw_en_word/train/speech_shape --train_shape_file exp/asr_stats_raw_en_word/train/text_shape.word --ngpu 1 --multiprocessing_distributed True
41
- [v030] 2022-01-31 01:56:49,667 (asr:382) INFO: Vocabulary size: 613
42
- [v030] 2022-01-31 01:57:10,900 (abs_task:1132) INFO: pytorch.version=1.8.1+cu102, cuda.available=True, cudnn.version=7605, cudnn.benchmark=False, cudnn.deterministic=True
43
- [v030] 2022-01-31 01:57:11,231 (abs_task:1133) INFO: Model structure:
44
- ESPnetASRModel(
45
- (frontend): DefaultFrontend(
46
- (stft): Stft(n_fft=512, win_length=512, hop_length=128, center=True, normalized=False, onesided=True)
47
- (frontend): Frontend()
48
- (logmel): LogMel(sr=16000, n_fft=512, n_mels=80, fmin=0, fmax=8000.0, htk=False)
49
- )
50
- (specaug): SpecAug(
51
- (time_warp): TimeWarp(window=5, mode=bicubic)
52
- (freq_mask): MaskAlongAxis(mask_width_range=[0, 30], num_mask=2, axis=freq)
53
- (time_mask): MaskAlongAxis(mask_width_range=[0, 40], num_mask=2, axis=time)
54
- )
55
- (normalize): GlobalMVN(stats_file=exp/asr_stats_raw_en_word/train/feats_stats.npz, norm_means=True, norm_vars=True)
56
- (encoder): TransformerEncoder(
57
- (embed): Conv2dSubsampling(
58
- (conv): Sequential(
59
- (0): Conv2d(1, 256, kernel_size=(3, 3), stride=(2, 2))
60
- (1): ReLU()
61
- (2): Conv2d(256, 256, kernel_size=(3, 3), stride=(2, 2))
62
- (3): ReLU()
63
- )
64
- (out): Sequential(
65
- (0): Linear(in_features=4864, out_features=256, bias=True)
66
- (1): PositionalEncoding(
67
- (dropout): Dropout(p=0.1, inplace=False)
68
- )
69
- )
70
- )
71
- (encoders): MultiSequential(
72
- (0): EncoderLayer(
73
- (self_attn): MultiHeadedAttention(
74
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
75
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
76
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
77
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
78
- (dropout): Dropout(p=0.0, inplace=False)
79
- )
80
- (feed_forward): PositionwiseFeedForward(
81
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
82
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
83
- (dropout): Dropout(p=0.1, inplace=False)
84
- (activation): ReLU()
85
- )
86
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
87
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
88
- (dropout): Dropout(p=0.1, inplace=False)
89
- )
90
- (1): EncoderLayer(
91
- (self_attn): MultiHeadedAttention(
92
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
93
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
94
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
95
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
96
- (dropout): Dropout(p=0.0, inplace=False)
97
- )
98
- (feed_forward): PositionwiseFeedForward(
99
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
100
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
101
- (dropout): Dropout(p=0.1, inplace=False)
102
- (activation): ReLU()
103
- )
104
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
105
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
106
- (dropout): Dropout(p=0.1, inplace=False)
107
- )
108
- (2): EncoderLayer(
109
- (self_attn): MultiHeadedAttention(
110
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
111
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
112
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
113
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
114
- (dropout): Dropout(p=0.0, inplace=False)
115
- )
116
- (feed_forward): PositionwiseFeedForward(
117
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
118
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
119
- (dropout): Dropout(p=0.1, inplace=False)
120
- (activation): ReLU()
121
- )
122
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
123
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
124
- (dropout): Dropout(p=0.1, inplace=False)
125
- )
126
- (3): EncoderLayer(
127
- (self_attn): MultiHeadedAttention(
128
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
129
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
130
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
131
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
132
- (dropout): Dropout(p=0.0, inplace=False)
133
- )
134
- (feed_forward): PositionwiseFeedForward(
135
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
136
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
137
- (dropout): Dropout(p=0.1, inplace=False)
138
- (activation): ReLU()
139
- )
140
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
141
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
142
- (dropout): Dropout(p=0.1, inplace=False)
143
- )
144
- (4): EncoderLayer(
145
- (self_attn): MultiHeadedAttention(
146
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
147
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
148
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
149
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
150
- (dropout): Dropout(p=0.0, inplace=False)
151
- )
152
- (feed_forward): PositionwiseFeedForward(
153
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
154
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
155
- (dropout): Dropout(p=0.1, inplace=False)
156
- (activation): ReLU()
157
- )
158
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
159
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
160
- (dropout): Dropout(p=0.1, inplace=False)
161
- )
162
- (5): EncoderLayer(
163
- (self_attn): MultiHeadedAttention(
164
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
165
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
166
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
167
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
168
- (dropout): Dropout(p=0.0, inplace=False)
169
- )
170
- (feed_forward): PositionwiseFeedForward(
171
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
172
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
173
- (dropout): Dropout(p=0.1, inplace=False)
174
- (activation): ReLU()
175
- )
176
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
177
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
178
- (dropout): Dropout(p=0.1, inplace=False)
179
- )
180
- (6): EncoderLayer(
181
- (self_attn): MultiHeadedAttention(
182
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
183
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
184
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
185
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
186
- (dropout): Dropout(p=0.0, inplace=False)
187
- )
188
- (feed_forward): PositionwiseFeedForward(
189
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
190
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
191
- (dropout): Dropout(p=0.1, inplace=False)
192
- (activation): ReLU()
193
- )
194
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
195
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
196
- (dropout): Dropout(p=0.1, inplace=False)
197
- )
198
- (7): EncoderLayer(
199
- (self_attn): MultiHeadedAttention(
200
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
201
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
202
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
203
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
204
- (dropout): Dropout(p=0.0, inplace=False)
205
- )
206
- (feed_forward): PositionwiseFeedForward(
207
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
208
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
209
- (dropout): Dropout(p=0.1, inplace=False)
210
- (activation): ReLU()
211
- )
212
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
213
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
214
- (dropout): Dropout(p=0.1, inplace=False)
215
- )
216
- (8): EncoderLayer(
217
- (self_attn): MultiHeadedAttention(
218
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
219
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
220
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
221
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
222
- (dropout): Dropout(p=0.0, inplace=False)
223
- )
224
- (feed_forward): PositionwiseFeedForward(
225
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
226
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
227
- (dropout): Dropout(p=0.1, inplace=False)
228
- (activation): ReLU()
229
- )
230
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
231
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
232
- (dropout): Dropout(p=0.1, inplace=False)
233
- )
234
- (9): EncoderLayer(
235
- (self_attn): MultiHeadedAttention(
236
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
237
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
238
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
239
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
240
- (dropout): Dropout(p=0.0, inplace=False)
241
- )
242
- (feed_forward): PositionwiseFeedForward(
243
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
244
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
245
- (dropout): Dropout(p=0.1, inplace=False)
246
- (activation): ReLU()
247
- )
248
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
249
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
250
- (dropout): Dropout(p=0.1, inplace=False)
251
- )
252
- (10): EncoderLayer(
253
- (self_attn): MultiHeadedAttention(
254
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
255
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
256
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
257
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
258
- (dropout): Dropout(p=0.0, inplace=False)
259
- )
260
- (feed_forward): PositionwiseFeedForward(
261
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
262
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
263
- (dropout): Dropout(p=0.1, inplace=False)
264
- (activation): ReLU()
265
- )
266
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
267
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
268
- (dropout): Dropout(p=0.1, inplace=False)
269
- )
270
- (11): EncoderLayer(
271
- (self_attn): MultiHeadedAttention(
272
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
273
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
274
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
275
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
276
- (dropout): Dropout(p=0.0, inplace=False)
277
- )
278
- (feed_forward): PositionwiseFeedForward(
279
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
280
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
281
- (dropout): Dropout(p=0.1, inplace=False)
282
- (activation): ReLU()
283
- )
284
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
285
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
286
- (dropout): Dropout(p=0.1, inplace=False)
287
- )
288
- )
289
- (after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
290
- )
291
- (decoder): TransformerDecoder(
292
- (embed): Sequential(
293
- (0): Embedding(613, 256)
294
- (1): PositionalEncoding(
295
- (dropout): Dropout(p=0.1, inplace=False)
296
- )
297
- )
298
- (after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
299
- (output_layer): Linear(in_features=256, out_features=613, bias=True)
300
- (decoders): MultiSequential(
301
- (0): DecoderLayer(
302
- (self_attn): MultiHeadedAttention(
303
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
304
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
305
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
306
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
307
- (dropout): Dropout(p=0.0, inplace=False)
308
- )
309
- (src_attn): MultiHeadedAttention(
310
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
311
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
312
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
313
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
314
- (dropout): Dropout(p=0.0, inplace=False)
315
- )
316
- (feed_forward): PositionwiseFeedForward(
317
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
318
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
319
- (dropout): Dropout(p=0.1, inplace=False)
320
- (activation): ReLU()
321
- )
322
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
323
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
324
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
325
- (dropout): Dropout(p=0.1, inplace=False)
326
- )
327
- (1): DecoderLayer(
328
- (self_attn): MultiHeadedAttention(
329
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
330
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
331
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
332
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
333
- (dropout): Dropout(p=0.0, inplace=False)
334
- )
335
- (src_attn): MultiHeadedAttention(
336
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
337
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
338
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
339
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
340
- (dropout): Dropout(p=0.0, inplace=False)
341
- )
342
- (feed_forward): PositionwiseFeedForward(
343
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
344
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
345
- (dropout): Dropout(p=0.1, inplace=False)
346
- (activation): ReLU()
347
- )
348
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
349
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
350
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
351
- (dropout): Dropout(p=0.1, inplace=False)
352
- )
353
- (2): DecoderLayer(
354
- (self_attn): MultiHeadedAttention(
355
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
356
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
357
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
358
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
359
- (dropout): Dropout(p=0.0, inplace=False)
360
- )
361
- (src_attn): MultiHeadedAttention(
362
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
363
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
364
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
365
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
366
- (dropout): Dropout(p=0.0, inplace=False)
367
- )
368
- (feed_forward): PositionwiseFeedForward(
369
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
370
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
371
- (dropout): Dropout(p=0.1, inplace=False)
372
- (activation): ReLU()
373
- )
374
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
375
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
376
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
377
- (dropout): Dropout(p=0.1, inplace=False)
378
- )
379
- (3): DecoderLayer(
380
- (self_attn): MultiHeadedAttention(
381
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
382
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
383
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
384
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
385
- (dropout): Dropout(p=0.0, inplace=False)
386
- )
387
- (src_attn): MultiHeadedAttention(
388
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
389
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
390
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
391
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
392
- (dropout): Dropout(p=0.0, inplace=False)
393
- )
394
- (feed_forward): PositionwiseFeedForward(
395
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
396
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
397
- (dropout): Dropout(p=0.1, inplace=False)
398
- (activation): ReLU()
399
- )
400
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
401
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
402
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
403
- (dropout): Dropout(p=0.1, inplace=False)
404
- )
405
- (4): DecoderLayer(
406
- (self_attn): MultiHeadedAttention(
407
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
408
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
409
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
410
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
411
- (dropout): Dropout(p=0.0, inplace=False)
412
- )
413
- (src_attn): MultiHeadedAttention(
414
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
415
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
416
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
417
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
418
- (dropout): Dropout(p=0.0, inplace=False)
419
- )
420
- (feed_forward): PositionwiseFeedForward(
421
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
422
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
423
- (dropout): Dropout(p=0.1, inplace=False)
424
- (activation): ReLU()
425
- )
426
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
427
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
428
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
429
- (dropout): Dropout(p=0.1, inplace=False)
430
- )
431
- (5): DecoderLayer(
432
- (self_attn): MultiHeadedAttention(
433
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
434
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
435
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
436
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
437
- (dropout): Dropout(p=0.0, inplace=False)
438
- )
439
- (src_attn): MultiHeadedAttention(
440
- (linear_q): Linear(in_features=256, out_features=256, bias=True)
441
- (linear_k): Linear(in_features=256, out_features=256, bias=True)
442
- (linear_v): Linear(in_features=256, out_features=256, bias=True)
443
- (linear_out): Linear(in_features=256, out_features=256, bias=True)
444
- (dropout): Dropout(p=0.0, inplace=False)
445
- )
446
- (feed_forward): PositionwiseFeedForward(
447
- (w_1): Linear(in_features=256, out_features=2048, bias=True)
448
- (w_2): Linear(in_features=2048, out_features=256, bias=True)
449
- (dropout): Dropout(p=0.1, inplace=False)
450
- (activation): ReLU()
451
- )
452
- (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
453
- (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
454
- (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
455
- (dropout): Dropout(p=0.1, inplace=False)
456
- )
457
- )
458
- )
459
- (ctc): CTC(
460
- (ctc_lo): Linear(in_features=256, out_features=613, bias=True)
461
- (ctc_loss): CTCLoss()
462
- )
463
- (criterion_att): LabelSmoothingLoss(
464
- (criterion): KLDivLoss()
465
- )
466
- )
467
-
468
- Model summary:
469
- Class Name: ESPnetASRModel
470
- Total Number of model parameters: 27.56 M
471
- Number of trainable parameters: 27.56 M (100.0%)
472
- Size: 110.26 MB
473
- Type: torch.float32
474
- [v030] 2022-01-31 01:57:11,231 (abs_task:1136) INFO: Optimizer:
475
- Adam (
476
- Parameter Group 0
477
- amsgrad: False
478
- betas: (0.9, 0.999)
479
- eps: 1e-08
480
- initial_lr: 0.0002
481
- lr: 8e-09
482
- weight_decay: 0
483
- )
484
- [v030] 2022-01-31 01:57:11,232 (abs_task:1137) INFO: Scheduler: WarmupLR(warmup_steps=25000)
485
- [v030] 2022-01-31 01:57:11,236 (abs_task:1146) INFO: Saving the configuration in exp/asr_train_asr_raw_en_word/config.yaml
486
- [v030] 2022-01-31 01:57:11,750 (abs_task:1493) INFO: [train] dataset:
487
- ESPnetDataset(
488
- speech: {"path": "dump/raw/train/wav.scp", "type": "sound"}
489
- text: {"path": "dump/raw/train/text", "type": "text"}
490
- preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a71f8310>)
491
- [v030] 2022-01-31 01:57:11,751 (abs_task:1494) INFO: [train] Batch sampler: FoldedBatchSampler(N-batch=547, batch_size=20, shape_files=['exp/asr_stats_raw_en_word/train/speech_shape', 'exp/asr_stats_raw_en_word/train/text_shape.word'], sort_in_batch=descending, sort_batch=descending)
492
- [v030] 2022-01-31 01:57:11,787 (abs_task:1495) INFO: [train] mini-batch sizes summary: N-batch=547, mean=19.9, min=6, max=20
493
- [v030] 2022-01-31 01:57:12,032 (abs_task:1493) INFO: [valid] dataset:
494
- ESPnetDataset(
495
- speech: {"path": "dump/raw/valid/wav.scp", "type": "sound"}
496
- text: {"path": "dump/raw/valid/text", "type": "text"}
497
- preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a3c2d520>)
498
- [v030] 2022-01-31 01:57:12,033 (abs_task:1494) INFO: [valid] Batch sampler: FoldedBatchSampler(N-batch=178, batch_size=20, shape_files=['exp/asr_stats_raw_en_word/valid/speech_shape', 'exp/asr_stats_raw_en_word/valid/text_shape.word'], sort_in_batch=descending, sort_batch=descending)
499
- [v030] 2022-01-31 01:57:12,033 (abs_task:1495) INFO: [valid] mini-batch sizes summary: N-batch=178, mean=20.0, min=19, max=20
500
- [v030] 2022-01-31 01:57:12,059 (abs_task:1493) INFO: [plot_att] dataset:
501
- ESPnetDataset(
502
- speech: {"path": "dump/raw/valid/wav.scp", "type": "sound"}
503
- text: {"path": "dump/raw/valid/text", "type": "text"}
504
- preprocess: <espnet2.train.preprocessor.CommonPreprocessor object at 0x1459a3c2dfa0>)
505
- [v030] 2022-01-31 01:57:12,059 (abs_task:1494) INFO: [plot_att] Batch sampler: UnsortedBatchSampler(N-batch=3559, batch_size=1, key_file=exp/asr_stats_raw_en_word/valid/speech_shape,
506
- [v030] 2022-01-31 01:57:12,059 (abs_task:1495) INFO: [plot_att] mini-batch sizes summary: N-batch=3, mean=1.0, min=1, max=1
507
- [v030] 2022-01-31 01:57:12,109 (trainer:273) INFO: 1/2epoch started
508
- [v030] 2022-01-31 01:57:30,639 (trainer:653) INFO: 1epoch:train:1-27batch: iter_time=0.230, forward_time=0.249, loss=141.855, loss_att=47.067, loss_ctc=236.642, acc=4.953e-04, backward_time=0.036, optim_step_time=0.026, optim0_lr0=1.200e-07, train_time=0.632
509
- [v030] 2022-01-31 01:57:39,688 (trainer:653) INFO: 1epoch:train:28-54batch: iter_time=0.188, forward_time=0.069, loss=142.019, loss_att=49.139, loss_ctc=234.898, acc=2.572e-04, backward_time=0.030, optim_step_time=0.024, optim0_lr0=3.360e-07, train_time=0.373
510
- [v030] 2022-01-31 01:57:45,424 (trainer:653) INFO: 1epoch:train:55-81batch: iter_time=0.082, forward_time=0.058, loss=136.567, loss_att=49.707, loss_ctc=223.427, acc=4.240e-04, backward_time=0.029, optim_step_time=0.023, optim0_lr0=5.520e-07, train_time=0.213
511
- [v030] 2022-01-31 01:57:51,546 (trainer:653) INFO: 1epoch:train:82-108batch: iter_time=0.103, forward_time=0.054, loss=140.269, loss_att=55.069, loss_ctc=225.468, acc=3.225e-04, backward_time=0.026, optim_step_time=0.022, optim0_lr0=7.680e-07, train_time=0.227
512
- [v030] 2022-01-31 01:57:57,839 (trainer:653) INFO: 1epoch:train:109-135batch: iter_time=0.136, forward_time=0.034, loss=103.266, loss_att=50.612, loss_ctc=155.921, acc=2.684e-04, backward_time=0.023, optim_step_time=0.022, optim0_lr0=9.840e-07, train_time=0.233
513
- [v030] 2022-01-31 01:58:07,335 (trainer:653) INFO: 1epoch:train:136-162batch: iter_time=0.133, forward_time=0.138, loss=86.321, loss_att=51.425, loss_ctc=121.217, acc=2.486e-04, backward_time=0.030, optim_step_time=0.022, optim0_lr0=1.200e-06, train_time=0.351
514
- [v030] 2022-01-31 01:58:14,180 (trainer:653) INFO: 1epoch:train:163-189batch: iter_time=0.118, forward_time=0.068, loss=59.836, loss_att=46.904, loss_ctc=72.767, acc=2.503e-04, backward_time=0.026, optim_step_time=0.022, optim0_lr0=1.416e-06, train_time=0.253
515
- [v030] 2022-01-31 01:58:19,882 (trainer:653) INFO: 1epoch:train:190-216batch: iter_time=0.110, forward_time=0.035, loss=62.973, loss_att=55.924, loss_ctc=70.023, acc=2.639e-04, backward_time=0.023, optim_step_time=0.022, optim0_lr0=1.632e-06, train_time=0.211
516
- [v030] 2022-01-31 01:58:25,457 (trainer:653) INFO: 1epoch:train:217-243batch: iter_time=0.080, forward_time=0.058, loss=50.057, loss_att=50.607, loss_ctc=49.508, acc=0.000e+00, backward_time=0.026, optim_step_time=0.023, optim0_lr0=1.848e-06, train_time=0.206
517
- [v030] 2022-01-31 01:58:40,212 (trainer:653) INFO: 1epoch:train:244-270batch: iter_time=0.090, forward_time=0.381, loss=42.156, loss_att=44.508, loss_ctc=39.803, acc=0.001, backward_time=0.027, optim_step_time=0.023, optim0_lr0=2.064e-06, train_time=0.546
518
- [v030] 2022-01-31 01:58:49,595 (trainer:653) INFO: 1epoch:train:271-297batch: iter_time=0.165, forward_time=0.104, loss=35.478, loss_att=38.536, loss_ctc=32.420, acc=0.004, backward_time=0.031, optim_step_time=0.025, optim0_lr0=2.280e-06, train_time=0.347
519
- [v030] 2022-01-31 01:58:55,079 (trainer:653) INFO: 1epoch:train:298-324batch: iter_time=0.107, forward_time=0.033, loss=40.375, loss_att=43.577, loss_ctc=37.174, acc=0.014, backward_time=0.023, optim_step_time=0.022, optim0_lr0=2.496e-06, train_time=0.203
520
- [v030] 2022-01-31 01:59:05,215 (trainer:653) INFO: 1epoch:train:325-351batch: iter_time=0.110, forward_time=0.102, loss=51.612, loss_att=55.003, loss_ctc=48.221, acc=0.023, backward_time=0.095, optim_step_time=0.032, optim0_lr0=2.712e-06, train_time=0.375
521
- [v030] 2022-01-31 01:59:16,439 (trainer:653) INFO: 1epoch:train:352-378batch: iter_time=0.198, forward_time=0.133, loss=43.941, loss_att=47.359, loss_ctc=40.524, acc=0.052, backward_time=0.035, optim_step_time=0.025, optim0_lr0=2.928e-06, train_time=0.415
522
- [v030] 2022-01-31 01:59:22,380 (trainer:653) INFO: 1epoch:train:379-405batch: iter_time=0.055, forward_time=0.095, loss=40.773, loss_att=44.045, loss_ctc=37.502, acc=0.088, backward_time=0.027, optim_step_time=0.023, optim0_lr0=3.144e-06, train_time=0.220
523
- [v030] 2022-01-31 01:59:28,375 (trainer:653) INFO: 1epoch:train:406-432batch: iter_time=0.121, forward_time=0.037, loss=41.249, loss_att=44.500, loss_ctc=37.998, acc=0.114, backward_time=0.023, optim_step_time=0.022, optim0_lr0=3.360e-06, train_time=0.222
524
- [v030] 2022-01-31 01:59:40,639 (trainer:653) INFO: 1epoch:train:433-459batch: iter_time=0.094, forward_time=0.278, loss=47.537, loss_att=50.761, loss_ctc=44.314, acc=0.119, backward_time=0.035, optim_step_time=0.025, optim0_lr0=3.576e-06, train_time=0.454
525
- [v030] 2022-01-31 02:00:15,522 (trainer:653) INFO: 1epoch:train:460-486batch: iter_time=0.126, forward_time=1.014, loss=37.846, loss_att=40.791, loss_ctc=34.901, acc=0.159, backward_time=0.079, optim_step_time=0.031, optim0_lr0=3.792e-06, train_time=1.292
526
- [v030] 2022-01-31 02:00:29,360 (trainer:653) INFO: 1epoch:train:487-513batch: iter_time=0.110, forward_time=0.262, loss=48.077, loss_att=51.027, loss_ctc=45.127, acc=0.149, backward_time=0.061, optim_step_time=0.033, optim0_lr0=4.008e-06, train_time=0.512
527
- [v030] 2022-01-31 02:00:38,109 (trainer:653) INFO: 1epoch:train:514-540batch: iter_time=0.097, forward_time=0.136, loss=35.719, loss_att=38.304, loss_ctc=33.134, acc=0.214, backward_time=0.036, optim_step_time=0.027, optim0_lr0=4.224e-06, train_time=0.324
528
- [v030] 2022-01-31 02:02:20,408 (trainer:328) INFO: 1epoch results: [train] iter_time=0.124, forward_time=0.165, loss=68.972, loss_att=47.724, loss_ctc=90.219, acc=0.049, backward_time=0.036, optim_step_time=0.025, optim0_lr0=2.200e-06, train_time=0.379, time=3 minutes and 28.1 seconds, total_count=547, gpu_max_cached_mem_GB=4.266, [valid] loss=42.143, loss_att=42.969, loss_ctc=41.317, acc=0.226, cer=0.917, wer=1.000, cer_ctc=1.000, time=34.01 seconds, total_count=178, gpu_max_cached_mem_GB=4.891, [att_plot] time=1 minute and 6.1 seconds, total_count=0, gpu_max_cached_mem_GB=4.891
529
- [v030] 2022-01-31 02:02:24,238 (trainer:375) INFO: The best model has been updated: train.loss, valid.loss, train.acc, valid.acc
530
- [v030] 2022-01-31 02:02:24,290 (trainer:261) INFO: 2/2epoch started. Estimated time to finish: 5 minutes and 12.15 seconds
531
- [v030] 2022-01-31 02:02:39,772 (trainer:653) INFO: 2epoch:train:1-27batch: iter_time=0.200, forward_time=0.227, loss=37.230, loss_att=39.722, loss_ctc=34.738, acc=0.219, backward_time=0.046, optim_step_time=0.026, optim0_lr0=4.496e-06, train_time=0.540
532
- [v030] 2022-01-31 02:02:46,635 (trainer:653) INFO: 2epoch:train:28-54batch: iter_time=0.143, forward_time=0.046, loss=35.185, loss_att=37.229, loss_ctc=33.140, acc=0.229, backward_time=0.024, optim_step_time=0.022, optim0_lr0=4.712e-06, train_time=0.254
533
- [v030] 2022-01-31 02:02:54,463 (trainer:653) INFO: 2epoch:train:55-81batch: iter_time=0.136, forward_time=0.073, loss=42.493, loss_att=44.580, loss_ctc=40.405, acc=0.204, backward_time=0.026, optim_step_time=0.032, optim0_lr0=4.928e-06, train_time=0.290
534
- [v030] 2022-01-31 02:03:02,383 (trainer:653) INFO: 2epoch:train:82-108batch: iter_time=0.100, forward_time=0.122, loss=44.804, loss_att=46.801, loss_ctc=42.807, acc=0.193, backward_time=0.026, optim_step_time=0.023, optim0_lr0=5.144e-06, train_time=0.293
535
- [v030] 2022-01-31 02:03:10,308 (trainer:653) INFO: 2epoch:train:109-135batch: iter_time=0.091, forward_time=0.111, loss=37.952, loss_att=39.864, loss_ctc=36.040, acc=0.221, backward_time=0.045, optim_step_time=0.023, optim0_lr0=5.360e-06, train_time=0.293
536
- [v030] 2022-01-31 02:03:29,729 (trainer:653) INFO: 2epoch:train:136-162batch: iter_time=0.116, forward_time=0.490, loss=36.738, loss_att=38.327, loss_ctc=35.149, acc=0.230, backward_time=0.046, optim_step_time=0.026, optim0_lr0=5.576e-06, train_time=0.719
537
- [v030] 2022-01-31 02:03:41,965 (trainer:653) INFO: 2epoch:train:163-189batch: iter_time=0.154, forward_time=0.226, loss=33.661, loss_att=34.931, loss_ctc=32.390, acc=0.232, backward_time=0.028, optim_step_time=0.023, optim0_lr0=5.792e-06, train_time=0.453
538
- [v030] 2022-01-31 02:03:52,572 (trainer:653) INFO: 2epoch:train:190-216batch: iter_time=0.108, forward_time=0.164, loss=37.235, loss_att=38.509, loss_ctc=35.961, acc=0.236, backward_time=0.069, optim_step_time=0.023, optim0_lr0=6.008e-06, train_time=0.393
539
- [v030] 2022-01-31 02:04:09,632 (trainer:653) INFO: 2epoch:train:217-243batch: iter_time=0.137, forward_time=0.341, loss=36.192, loss_att=37.130, loss_ctc=35.254, acc=0.226, backward_time=0.074, optim_step_time=0.031, optim0_lr0=6.224e-06, train_time=0.631
540
- [v030] 2022-01-31 02:04:24,555 (trainer:653) INFO: 2epoch:train:244-270batch: iter_time=0.165, forward_time=0.303, loss=34.811, loss_att=35.738, loss_ctc=33.883, acc=0.241, backward_time=0.029, optim_step_time=0.029, optim0_lr0=6.440e-06, train_time=0.552
541
- [v030] 2022-01-31 02:04:41,937 (trainer:653) INFO: 2epoch:train:271-297batch: iter_time=0.122, forward_time=0.326, loss=35.547, loss_att=36.079, loss_ctc=35.016, acc=0.238, backward_time=0.109, optim_step_time=0.036, optim0_lr0=6.656e-06, train_time=0.644
542
- [v030] 2022-01-31 02:04:54,885 (trainer:653) INFO: 2epoch:train:298-324batch: iter_time=0.096, forward_time=0.271, loss=34.077, loss_att=34.718, loss_ctc=33.436, acc=0.243, backward_time=0.042, optim_step_time=0.026, optim0_lr0=6.872e-06, train_time=0.479
543
- [v030] 2022-01-31 02:05:05,565 (trainer:653) INFO: 2epoch:train:325-351batch: iter_time=0.225, forward_time=0.076, loss=35.986, loss_att=36.355, loss_ctc=35.616, acc=0.242, backward_time=0.042, optim_step_time=0.024, optim0_lr0=7.088e-06, train_time=0.395
544
- [v030] 2022-01-31 02:05:14,712 (trainer:653) INFO: 2epoch:train:352-378batch: iter_time=0.117, forward_time=0.152, loss=36.101, loss_att=36.323, loss_ctc=35.880, acc=0.243, backward_time=0.025, optim_step_time=0.023, optim0_lr0=7.304e-06, train_time=0.338
545
- [v030] 2022-01-31 02:05:31,872 (trainer:653) INFO: 2epoch:train:379-405batch: iter_time=0.080, forward_time=0.434, loss=31.234, loss_att=31.494, loss_ctc=30.973, acc=0.279, backward_time=0.056, optim_step_time=0.027, optim0_lr0=7.520e-06, train_time=0.635
546
- [v030] 2022-01-31 02:05:51,754 (trainer:653) INFO: 2epoch:train:406-432batch: iter_time=0.093, forward_time=0.395, loss=34.236, loss_att=34.248, loss_ctc=34.225, acc=0.278, backward_time=0.135, optim_step_time=0.037, optim0_lr0=7.736e-06, train_time=0.735
547
- [v030] 2022-01-31 02:06:02,624 (trainer:653) INFO: 2epoch:train:433-459batch: iter_time=0.127, forward_time=0.178, loss=30.387, loss_att=30.207, loss_ctc=30.567, acc=0.282, backward_time=0.037, optim_step_time=0.027, optim0_lr0=7.952e-06, train_time=0.403
548
- [v030] 2022-01-31 02:06:13,226 (trainer:653) INFO: 2epoch:train:460-486batch: iter_time=0.101, forward_time=0.167, loss=34.295, loss_att=34.131, loss_ctc=34.460, acc=0.281, backward_time=0.060, optim_step_time=0.030, optim0_lr0=8.168e-06, train_time=0.392
549
- [v030] 2022-01-31 02:06:23,189 (trainer:653) INFO: 2epoch:train:487-513batch: iter_time=0.091, forward_time=0.160, loss=31.108, loss_att=30.714, loss_ctc=31.502, acc=0.286, backward_time=0.057, optim_step_time=0.032, optim0_lr0=8.384e-06, train_time=0.368
550
- [v030] 2022-01-31 02:06:30,758 (trainer:653) INFO: 2epoch:train:514-540batch: iter_time=0.154, forward_time=0.051, loss=34.146, loss_att=33.576, loss_ctc=34.716, acc=0.262, backward_time=0.026, optim_step_time=0.028, optim0_lr0=8.600e-06, train_time=0.280
551
- [v030] 2022-01-31 02:08:56,243 (trainer:328) INFO: 2epoch results: [train] iter_time=0.128, forward_time=0.235, loss=35.545, loss_att=36.390, loss_ctc=34.699, acc=0.245, backward_time=0.054, optim_step_time=0.028, optim0_lr0=6.576e-06, train_time=0.479, time=4 minutes and 23.23 seconds, total_count=1094, gpu_max_cached_mem_GB=4.891, [valid] loss=34.559, loss_att=31.596, loss_ctc=37.522, acc=0.298, cer=0.977, wer=1.000, cer_ctc=1.000, time=1 minute and 19.1 seconds, total_count=356, gpu_max_cached_mem_GB=4.891, [att_plot] time=49.47 seconds, total_count=0, gpu_max_cached_mem_GB=4.891
552
- [v030] 2022-01-31 02:08:59,763 (trainer:375) INFO: The best model has been updated: train.loss, valid.loss, train.acc, valid.acc
553
- [v030] 2022-01-31 02:08:59,768 (trainer:431) INFO: The training was finished at 2 epochs
554
- # Accounting: begin_time=1643612152
555
- # Accounting: end_time=1643612955
556
- # Accounting: time=803 threads=1
557
- # Finished at Mon Jan 31 02:09:15 EST 2022 with status 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
exp/asr_train_asr_raw_en_word/train.acc.ave.pth DELETED
@@ -1 +0,0 @@
1
- train.acc.ave_5best.pth
 
 
exp/asr_train_asr_raw_en_word/train.acc.ave_1best.pth DELETED
@@ -1 +0,0 @@
1
- 2epoch.pth
 
 
exp/asr_train_asr_raw_en_word/train.acc.ave_5best.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6856ac64a7105bd9c0809ace8b562e044d2375eab41818f8c205a5cafefab1c
3
- size 110483405
 
 
 
 
exp/asr_train_asr_raw_en_word/train.acc.best.pth DELETED
@@ -1 +0,0 @@
1
- 195epoch.pth
 
 
exp/asr_train_asr_raw_en_word/train.log DELETED
The diff for this file is too large to render. See raw diff
 
exp/asr_train_asr_raw_en_word/train.loss.ave.pth DELETED
@@ -1 +0,0 @@
1
- train.loss.ave_5best.pth
 
 
exp/asr_train_asr_raw_en_word/train.loss.ave_1best.pth DELETED
@@ -1 +0,0 @@
1
- 2epoch.pth
 
 
exp/asr_train_asr_raw_en_word/train.loss.ave_5best.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b49bebd712995835c9df6a9b32244362261deef949cde007de65204d0ea1473b
3
- size 110483405
 
 
 
 
exp/asr_train_asr_raw_en_word/train.loss.best.pth DELETED
@@ -1 +0,0 @@
1
- 200epoch.pth
 
 
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/RESULTS.md RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.100ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.101ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.102ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.103ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.104ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.105ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.106ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.107ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.108ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.109ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.10ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.110ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.111ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.112ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.113ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.114ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.115ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.116ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.117ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.118ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.119ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.11ep.png RENAMED
File without changes
exp/{asr_train_asr_raw_en_word β†’ asr_train_asr_transformer}/att_ws/BrX8aDqK2cLZRYl_voip-10beae627f-20130328_122107_pt344x_0001704_0001832.wav/decoder.decoders.0.self_attn.120ep.png RENAMED
File without changes