File size: 93,936 Bytes
f54b36b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
# python3 -m espnet2.bin.asr_train --collect_stats true --use_preprocessor true --bpemodel data/token_list/bpe_unigram2000/bpe.model --token_type bpe --token_list data/token_list/bpe_unigram2000/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --train_shape_file exp/asr_stats_raw_bpe2000_sp/logdir/train.1.scp --valid_shape_file exp/asr_stats_raw_bpe2000_sp/logdir/valid.1.scp --output_dir exp/asr_stats_raw_bpe2000_sp/logdir/stats.1 --config conf/train.yaml --frontend_conf fs=16k --train_data_path_and_name_and_type dump/raw/train_sp/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/dev/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train_sp/text,text,text --valid_data_path_and_name_and_type dump/raw/dev/text,text,text 
# Started at Fri May 12 13:12:20 CST 2023
#
/mnt/bd/khassan-volume3/tools/espent_KSC_recipe_test/tools/miniconda/envs/espnet/bin/python3 /mnt/bd/khassan-volume3/tools/espent_KSC_recipe_test/espnet2/bin/asr_train.py --collect_stats true --use_preprocessor true --bpemodel data/token_list/bpe_unigram2000/bpe.model --token_type bpe --token_list data/token_list/bpe_unigram2000/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --train_shape_file exp/asr_stats_raw_bpe2000_sp/logdir/train.1.scp --valid_shape_file exp/asr_stats_raw_bpe2000_sp/logdir/valid.1.scp --output_dir exp/asr_stats_raw_bpe2000_sp/logdir/stats.1 --config conf/train.yaml --frontend_conf fs=16k --train_data_path_and_name_and_type dump/raw/train_sp/wav.scp,speech,sound --valid_data_path_and_name_and_type dump/raw/dev/wav.scp,speech,sound --train_data_path_and_name_and_type dump/raw/train_sp/text,text,text --valid_data_path_and_name_and_type dump/raw/dev/text,text,text
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:28,839 (asr:500) INFO: Vocabulary size: 2000
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.embed.conv.0.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.embed.conv.2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.embed.out.0.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.0.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,523 (initialize:88) INFO: Initialize encoder.encoders.1.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.1.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.1.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.2.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.3.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.4.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,524 (initialize:88) INFO: Initialize encoder.encoders.5.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.5.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.5.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.6.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.7.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.8.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,525 (initialize:88) INFO: Initialize encoder.encoders.9.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.9.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.10.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.encoders.11.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize encoder.after_norm.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.after_norm.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.output_layer.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,526 (initialize:88) INFO: Initialize decoder.decoders.0.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.1.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.2.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,527 (initialize:88) INFO: Initialize decoder.decoders.3.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.3.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.4.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.self_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.self_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.self_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.self_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.src_attn.linear_q.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.src_attn.linear_k.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.src_attn.linear_v.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.src_attn.linear_out.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.feed_forward.w_1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.feed_forward.w_2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.norm1.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.norm2.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,528 (initialize:88) INFO: Initialize decoder.decoders.5.norm3.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,529 (initialize:88) INFO: Initialize ctc.ctc_lo.bias to zeros
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,563 (abs_task:1201) INFO: pytorch.version=1.13.1, cuda.available=True, cudnn.version=8500, cudnn.benchmark=False, cudnn.deterministic=True
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,568 (abs_task:1202) INFO: Model structure:
ESPnetASRModel(
  (frontend): DefaultFrontend(
    (stft): Stft(n_fft=512, win_length=512, hop_length=128, center=True, normalized=False, onesided=True)
    (frontend): Frontend()
    (logmel): LogMel(sr=16000, n_fft=512, n_mels=80, fmin=0, fmax=8000.0, htk=False)
  )
  (specaug): SpecAug(
    (time_warp): TimeWarp(window=5, mode=bicubic)
    (freq_mask): MaskAlongAxis(mask_width_range=[0, 27], num_mask=2, axis=freq)
    (time_mask): MaskAlongAxisVariableMaxWidth(mask_width_ratio_range=[0.0, 0.05], num_mask=10, axis=time)
  )
  (normalize): UtteranceMVN(norm_means=True, norm_vars=False)
  (encoder): TransformerEncoder(
    (embed): Conv2dSubsampling(
      (conv): Sequential(
        (0): Conv2d(1, 256, kernel_size=(3, 3), stride=(2, 2))
        (1): ReLU()
        (2): Conv2d(256, 256, kernel_size=(3, 3), stride=(2, 2))
        (3): ReLU()
      )
      (out): Sequential(
        (0): Linear(in_features=4864, out_features=256, bias=True)
        (1): PositionalEncoding(
          (dropout): Dropout(p=0.1, inplace=False)
        )
      )
    )
    (encoders): MultiSequential(
      (0): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (1): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (2): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (3): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (4): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (5): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (6): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (7): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (8): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (9): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (10): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (11): EncoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
    )
    (after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
  )
  (decoder): TransformerDecoder(
    (embed): Sequential(
      (0): Embedding(2000, 256)
      (1): PositionalEncoding(
        (dropout): Dropout(p=0.1, inplace=False)
      )
    )
    (after_norm): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
    (output_layer): Linear(in_features=256, out_features=2000, bias=True)
    (decoders): MultiSequential(
      (0): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (1): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (2): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (3): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (4): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (5): DecoderLayer(
        (self_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (src_attn): MultiHeadedAttention(
          (linear_q): Linear(in_features=256, out_features=256, bias=True)
          (linear_k): Linear(in_features=256, out_features=256, bias=True)
          (linear_v): Linear(in_features=256, out_features=256, bias=True)
          (linear_out): Linear(in_features=256, out_features=256, bias=True)
          (dropout): Dropout(p=0.0, inplace=False)
        )
        (feed_forward): PositionwiseFeedForward(
          (w_1): Linear(in_features=256, out_features=2048, bias=True)
          (w_2): Linear(in_features=2048, out_features=256, bias=True)
          (dropout): Dropout(p=0.1, inplace=False)
          (activation): ReLU()
        )
        (norm1): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm2): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (norm3): LayerNorm((256,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
    )
  )
  (criterion_att): LabelSmoothingLoss(
    (criterion): KLDivLoss()
  )
  (ctc): CTC(
    (ctc_lo): Linear(in_features=256, out_features=2000, bias=True)
    (ctc_loss): CTCLoss()
  )
)

Model summary:
    Class Name: ESPnetASRModel
    Total Number of model parameters: 28.63 M
    Number of trainable parameters: 28.63 M (100.0%)
    Size: 114.53 MB
    Type: torch.float32
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,568 (abs_task:1205) INFO: Optimizer:
Adam (
Parameter Group 0
    amsgrad: False
    betas: (0.9, 0.999)
    capturable: False
    differentiable: False
    eps: 1e-08
    foreach: None
    fused: False
    initial_lr: 0.0001
    lr: 3.3333333333333334e-09
    maximize: False
    weight_decay: 0
)
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,568 (abs_task:1206) INFO: Scheduler: WarmupLR(warmup_steps=30000)
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,570 (abs_task:1215) INFO: Saving the configuration in exp/asr_stats_raw_bpe2000_sp/logdir/stats.1/config.yaml
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,617 (abs_task:1226) INFO: Namespace(accum_grad=1, allow_variable_data_keys=False, aux_ctc_tasks=[], batch_bins=1000000, batch_size=128, batch_type='folded', best_model_criterion=[['valid', 'acc', 'max']], bpemodel='data/token_list/bpe_unigram2000/bpe.model', chunk_excluded_key_prefixes=[], chunk_length=500, chunk_shift_ratio=0.5, cleaner=None, collect_stats=True, config='conf/train.yaml', create_graph_in_tensorboard=False, ctc_conf={'dropout_rate': 0.0, 'ctc_type': 'builtin', 'reduce': True, 'ignore_nan_grad': None, 'zero_infinity': True}, cudnn_benchmark=False, cudnn_deterministic=True, cudnn_enabled=True, decoder='transformer', decoder_conf={'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 6, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'self_attention_dropout_rate': 0.0, 'src_attention_dropout_rate': 0.0}, detect_anomaly=False, dist_backend='nccl', dist_init_method='env://', dist_launcher=None, dist_master_addr=None, dist_master_port=None, dist_rank=None, dist_world_size=None, distributed=False, dry_run=False, early_stopping_criterion=('valid', 'loss', 'min'), encoder='transformer', encoder_conf={'output_size': 256, 'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 12, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'attention_dropout_rate': 0.0, 'input_layer': 'conv2d', 'normalize_before': True}, exclude_weight_decay=False, exclude_weight_decay_conf={}, fold_length=[], freeze_param=[], frontend='default', frontend_conf={'fs': '16k'}, g2p=None, grad_clip=5.0, grad_clip_type=2.0, grad_noise=False, ignore_init_mismatch=False, init='xavier_uniform', init_param=[], input_size=None, iterator_type='sequence', joint_net_conf=None, keep_nbest_models=10, local_rank=None, log_interval=None, log_level='INFO', max_cache_fd=32, max_cache_size=0.0, max_epoch=100, model='espnet', model_conf={'ctc_weight': 0.3, 'lsm_weight': 0.1, 'length_normalized_loss': False}, multiple_iterator=False, multiprocessing_distributed=False, nbest_averaging_interval=0, ngpu=0, no_forward_run=False, noise_apply_prob=1.0, noise_db_range='13_15', noise_scp=None, non_linguistic_symbols=None, normalize='utterance_mvn', normalize_conf={}, num_att_plot=3, num_cache_chunks=1024, num_iters_per_epoch=None, num_workers=1, optim='adam', optim_conf={'lr': 0.0001}, output_dir='exp/asr_stats_raw_bpe2000_sp/logdir/stats.1', patience=5, postencoder=None, postencoder_conf={}, preencoder=None, preencoder_conf={}, preprocessor='default', preprocessor_conf={}, pretrain_path=None, print_config=False, required=['output_dir', 'token_list'], resume=False, rir_apply_prob=1.0, rir_scp=None, scheduler='warmuplr', scheduler_conf={'warmup_steps': 30000}, seed=0, sharded_ddp=False, short_noise_thres=0.5, sort_batch='descending', sort_in_batch='descending', specaug='specaug', specaug_conf={'apply_time_warp': True, 'time_warp_window': 5, 'time_warp_mode': 'bicubic', 'apply_freq_mask': True, 'freq_mask_width_range': [0, 27], 'num_freq_mask': 2, 'apply_time_mask': True, 'time_mask_width_ratio_range': [0.0, 0.05], 'num_time_mask': 10}, speech_volume_normalize=None, token_list=['<blank>', '<unk>', 'н', 'р', 'м', 'у', 'п', 'ға', 'ды', 'ы', 'і', 'т', 'да', 'с', '▁', 'ді', '▁бір', 'й', 'к', 'л', 'ның', 'ге', '▁да', 'ын', 'де', 'ты', '▁де', 'з', 'қа', 'ған', 'е', 'ма', 'ң', 'қ', 'ып', '▁мен', 'мен', 'та', 'ін', '▁ал', 'ке', 'на', 'ғы', 'ген', 'гі', 'сы', '▁деп', 'а', '▁бұл', 'ра', 'ті', 'дың', '▁бар', 'лық', 'дан', 'нің', 'іп', '▁қа', 'те', '▁осы', 'шы', '▁ол', '▁а', '▁екі', 'ына', 'ша', 'са', '▁жүз', 'тың', 'ым', 'ш', 'лар', 'и', 'дық', 'ла', 'лы', '▁ай', '▁ақ', '▁мың', 'іне', 'тар', 'тер', '▁және', 'се', '▁деген', '▁ма', 'ында', 'ба', 'сі', 'ре', 'дар', 'па', 'дің', 'д', '▁жоқ', 'я', 'ия', 'інің', 'ны', '▁сол', 'ден', '▁та', 'тан', 'б', '▁еді', '▁жан', 'нан', '▁о', '▁өз', '▁қара', '▁са', 'дай', '▁он', '▁ба', 'лік', 'ық', '▁жыл', 'лі', '▁қол', '▁үшін', 'тің', 'сын', 'мыз', '▁емес', '▁екен', 'ме', 'ші', 'не', '▁жол', 'лер', '▁не', 'дағы', '▁бас', '▁отыр', '▁адам', 'інде', 'дер', '▁болған', '▁жер', '▁болып', '▁ата', 'о', '▁оның', '▁ел', 'г', '▁қала', '▁со', '▁тұр', '▁ой', '▁с', 'қан', 'тық', '▁ша', 'ының', 'қы', '▁қазақ', 'ов', 'дей', '▁үй', 'ыс', '▁тіл', 'дік', 'ше', '▁көп', '▁күн', '▁жұмыс', '▁керек', 'сіз', 'сыз', 'ыл', 'еді', 'ар', '▁и', '▁құ', '▁өзі', 'за', '▁бала', '▁ке', '▁су', '▁ме', '▁үш', '▁бірақ', 'уға', 'ту', '▁болды', 'ыр', '▁е', '▁өмір', 'ұ', 'ле', 'етін', 'мын', 'кі', '▁жүр', 'ім', 'атын', '▁сөз', '▁ғана', 'тен', '▁жатқан', 'ер', '▁алып', 'ру', '▁тұ', '▁деді', '▁қар', '▁жа', 'ады', '▁енді', '▁ауыл', '▁жас', '▁басқа', '▁қор', 'ан', 'ли', 'сін', '▁бес', 'ж', '▁қазақстан', 'пен', '▁сал', '▁ө', '▁ар', 'ынан', '▁жаңа', '▁тү', '▁ер', 'нен', '▁кө', '▁ас', '▁ә', 'ри', '▁то', '▁болса', '▁соң', 'бай', '▁қай', 'лары', '▁жатыр', '▁көз', 'ә', '▁сен', 'лап', '▁кезде', '▁біз', 'інен', '▁кейін', '▁болады', '▁бұ', 'бе', '▁жылы', '▁к', 'кен', '▁құр', 'ғ', 'ль', '▁ғой', '▁қал', '▁жеті', 'міз', 'тып', 'қты', 'ни', 'ал', 'ай', '▁олар', 'тай', '▁тоғыз', '▁ең', '▁ала', 'ғын', '▁бе', '▁келген', '▁қыз', 'ір', '▁аз', '▁те', '▁па', 'уы', 'май', 'ылған', '▁оны', 'мін', '▁ұ', '▁ат', 'уге', '▁шығар', '▁п', 'тары', '▁ү', 'сына', '▁жай', 'бы', 'бек', 'ымыз', '▁же', '▁қой', '▁қайта', 'ро', 'ақ', '▁жиырма', '▁тағы', 'гер', '▁дейін', 'іл', '▁на', '▁тө', '▁ана', '▁жақсы', '▁ара', '▁қалған', '▁халық', 'х', '▁қалды', '▁әр', 'ік', 'тік', 'іміз', '▁тек', '▁қазір', '▁қос', 'лан', 'ымен', '▁жаз', 'тын', '▁төрт', 'ти', 'дегі', '▁кү', '▁көрсет', '▁қызмет', '▁жұ', '▁менің', '▁біл', '▁бер', 'ек', '▁жағдай', 'нда', 'он', 'рт', 'лардың', 'іс', 'ө', 'ні', '▁ау', '▁бойынша', '▁кел', '▁алды', 'ф', 'лады', '▁айт', '▁туралы', 'лған', 'сының', '▁үлкен', '▁өте', '▁келе', 'сында', '▁пен', 'ю', 'ен', '▁тұрған', '▁бет', '▁бай', '▁бәрі', 'йды', '▁се', '▁жүрген', '▁ту', '▁көр', '▁бері', 'мақ', 'стан', '▁бірі', '▁күй', '▁дейді', 'лері', '▁келеді', 'ылып', '▁ше', '▁г', 'лып', 'сың', '▁істе', '▁орта', '▁ақын', '▁келіп', '▁біздің', 'ди', '▁қарай', 'ор', 'дары', '▁ті', '▁алмай', '▁уақыт', '▁ш', 'йтын', '▁дүние', '▁қи', 'тері', '▁ко', 'пай', 'ры', 'ат', '▁сала', 'лен', 'ст', '▁кеткен', 'ындағы', '▁кетті', 'тты', '▁жаса', 'лігі', 'жа', 'сің', '▁әлем', '▁бірінші', 'лығы', 'дері', '▁ка', '▁бо', '▁мұ', '▁алған', '▁э', '▁күні', '▁т', '▁жар', '▁көк', '▁өлең', 'лды', 'імен', '▁іс', 'ымды', '▁жерде', '▁сұра', '▁екінші', '▁өзге', '▁у', '▁болатын', 'тау', '▁бөл', 'ет', '▁кір', 'рек', 'ка', 'лау', '▁сіз', 'кті', '▁жү', '▁алты', '▁абай', 'лас', 'тор', '▁кеш', '▁шын', 'лай', '▁өткен', 'ман', '▁әйел', '▁отырған', '▁оған', 'ылды', 'ч', '▁пе', '▁ұлы', 'ыш', '▁болма', 'жан', '▁облыс', '▁шо', '▁жал', '▁қалай', '▁түс', '▁аудан', '▁әлі', '▁мына', '▁олардың', 'ция', 'тін', 'в', 'ңыз', 'ің', '▁рет', '▁сондықтан', '▁мемлекет', '▁ішінде', '▁тә', '▁өр', 'ың', 'қын', 'ми', '▁тау', 'мей', '▁қатар', '▁кез', '▁жара', 'імді', '▁аға', '▁сана', '▁дә', '▁си', '▁өнер', '▁сегіз', 'кер', 'ев', '▁би', '▁келді', '▁қаз', '▁әрі', '▁отыз', 'сып', '▁жұрт', '▁тарт', '▁жо', '▁кісі', '▁жат', '▁по', '▁балалар', '▁әке', '▁жау', '▁кітап', 'шылық', '▁пайда', '▁сыр', 'ь', '▁күл', '▁жина', '▁қы', '▁берді', 'ика', 'сті', '▁нұр', '▁білім', '▁іш', '▁шыққан', '▁арқылы', 'ло', '▁ре', '▁қарсы', '▁әдебиет', '▁көңіл', 'сіне', 'дап', 'ү', '▁ми', '▁тара', '▁ет', '▁сер', '▁таста', '▁ки', '▁оқу', '▁шығарма', 'тардың', '▁сүй', '▁ж', 'хан', '▁д', 'сіп', '▁маған', '▁тап', '▁көзі', '▁мүмкін', '▁заң', 'дым', '▁түр', '▁сияқты', '▁кім', '▁ғылым', '▁өл', '▁шал', '▁айтып', 'дардың', '▁бұрын', '▁ұлттық', '▁сөйле', '▁дәл', '▁ұш', '▁жыр', '▁бойы', '▁жібер', '▁айта', '▁ти', '▁айтқан', '▁әңгіме', '▁есе', 'тті', '▁баға', '▁осындай', 'маған', '▁біреу', '▁өзен', '▁арт', '▁мо', '▁бүгін', 'ріп', 'латын', 'ушы', '▁заман', 'бір', '▁сын', '▁хан', '▁түн', '▁таң', '▁апа', '▁ди', '▁барлық', 'іңіз', '▁тіпті', '▁басты', '▁қан', '▁ра', '▁з', '▁үн', 'ха', '▁шақыр', '▁ән', '▁сондай', 'лердің', '▁қырық', '▁сау', '▁жоғары', '▁мақсат', '▁ұр', 'сынан', '▁берген', '▁байқа', 'ғына', '▁ұста', '▁қыр', 'тқан', '▁қо', 'то', '▁бол', '▁кең', 'рып', 'ыра', '▁ая', '▁қандай', '▁жы', '▁қарап', '▁жігіт', '▁оқы', 'ыңыз', '▁есті', 'тал', '▁күш', 'э', '▁екенін', '▁еңбек', '▁ерекше', 'дігі', '▁бола', '▁шар', '▁біле', '▁әлде', '▁мә', '▁адамдар', '▁сы', 'лді', '▁ха', 'же', 'леді', '▁таны', '▁тоқта', 'ларды', 'уі', '▁мү', '▁елу', '▁б', '▁мектеп', '▁қу', 'ларын', 'тіп', '▁етіп', '▁сонда', '▁сырт', '▁тиіс', '▁аш', '▁ізде', '▁сай', '▁одан', '▁кеңес', '▁жеткіз', '▁мемлекеттік', '▁қажет', 'ва', '▁дос', 'мек', '▁жөн', '▁жауап', 'баев', '▁іл', '▁алматы', 'леп', '▁тал', 'сақ', '▁орын', 'нде', '▁себебі', '▁жалпы', '▁жоба', '▁көрген', '▁негіз', '▁қам', '▁бара', 'си', 'рап', '▁қазіргі', '▁соңғы', '▁шығып', 'ғым', '▁алу', 'тас', '▁барып', '▁кезінде', 'лу', '▁бө', '▁қыл', 'ілген', '▁тас', '▁өткіз', '▁қабылда', '▁қатты', '▁саны', '▁бере', 'ғыз', '▁бірге', '▁топ', 'тығы', '▁қыс', 'ңіз', '▁басшысы', '▁ес', '▁м', '▁байланысты', '▁жазушы', 'ол', '▁түсі', '▁ие', '▁қоя', 'бі', 'лес', '▁жары', 'ұр', '▁немесе', '▁бірнеше', 'ұлы', '▁бастап', 'тердің', '▁беру', '▁береді', '▁жолы', 'лаған', '▁әл', 'су', 'тап', '▁өзін', '▁мал', '▁мұндай', '▁тоқсан', 'тпе', '▁алады', '▁ем', 'тарын', '▁туған', '▁түрлі', '▁бу', '▁әуе', 'дығы', '▁өт', '▁орналасқан', '▁қайт', '▁бұр', '▁болар', '▁қолдан', 'би', '▁тол', 'зе', '▁қиын', '▁ауыр', '▁жастар', 'икалық', '▁жалғыз', '▁атқа', '▁табы', '▁шықты', 'тағы', '▁сез', '▁оң', '▁әлеуметті', '▁аң', '▁орында', '▁жеке', 'майды', 'сек', '▁көріп', '▁тапсыр', '▁қызы', 'шылар', '▁аудар', '▁өйткені', '▁салып', 'жы', '▁хабар', '▁сан', '▁отбасы', '▁анықта', '▁түскен', '▁боп', '▁ин', 'қай', '▁еске', '▁мені', '▁онда', '▁таба', '▁көрін', '▁бә', '▁алғашқы', '▁аты', '▁түсін', '▁таңда', 'быр', '▁құла', '▁теңге', 'ылады', '▁дү', 'ко', '▁белгі', '▁түсті', 'індегі', '▁егер', 'ци', '▁за', '▁оқиға', '▁бастады', '▁жағы', '▁алпыс', '▁жете', 'дарын', '▁ретінде', '▁қоғам', '▁арнайы', '▁адамның', '▁сонымен', '▁беріп', '▁сайын', '▁гүл', 'йді', '▁алдында', 'тегі', 'ғанда', 'во', 'зи', 'ларға', 'рат', '▁арман', 'тарды', '▁дала', '▁ақыл', 'еу', 'но', '▁отырып', '▁бек', '▁еш', '▁қанша', 'терін', '▁аман', '▁құра', '▁өңір', '▁түсіп', '▁қалып', '▁сый', '▁өзім', '▁сезім', '▁орыс', '▁жүріп', '▁айтты', '▁тұрып', '▁екеуі', '▁әй', '▁бей', 'хана', 'саң', '▁есі', 'бо', '▁сурет', '▁сақта', 'улы', '▁бастал', '▁кете', '▁бал', '▁аса', 'дердің', '▁жақ', 'пақ', 'ви', '▁белгілі', '▁қызыл', '▁десе', '▁әкім', '▁сүр', '▁май', '▁до', '▁алайда', 'сем', '▁ақша', '▁бү', 'ве', 'егін', 'йтін', 'лек', 'га', '▁содан', '▁төмен', 'ау', 'сқан', '▁дұрыс', '▁астана', '▁қон', '▁тір', '▁халқы', '▁сү', '▁республика', '▁міне', '▁кеше', '▁түрі', '▁міндет', 'лген', '▁төле', '▁батыр', 'теп', '▁көш', '▁биыл', '▁жи', '▁ф', '▁орны', '▁мамандар', '▁мән', '▁өздері', '▁сә', '▁үкімет', '▁бақ', 'бен', 'іш', '▁неге', 'майтын', '▁қаржы', '▁жүрегі', '▁арасында', 'пей', '▁аяқ', '▁соғыс', '▁құдай', '▁ойын', 'тей', '▁жел', 'терді', '▁жасы', '▁дауыс', '▁әзір', '▁ары', '▁бақыт', '▁әб', 'аралық', '▁жең', '▁шара', '▁баласы', '▁едім', '▁шыға', '▁біт', '▁орталық', '▁өзіне', 'мет', '▁ы', '▁адамды', '▁сексен', '▁ла', '▁ұлт', '▁болуы', '▁жап', '▁президент', '▁толық', 'сам', 'ова', '▁шай', '▁қайда', '▁барлығы', '▁шеш', 'лей', '▁көре', 'ілді', '▁уақытта', '▁миллион', 'мейді', 'ліп', '▁күнде', '▁азамат', 'бет', '▁жетпіс', 'це', '▁от', '▁ешкім', 'ация', '▁көбі', 'леу', '▁есім', '▁алтын', '▁секілді', '▁бой', 'ткен', '▁пікір', '▁қатысты', '▁шығыс', 'ушылар', 'ист', '▁өзің', '▁кетіп', '▁көмек', '▁тер', '▁ұл', '▁құл', '▁көше', '▁бәрін', '▁нәрсе', 'бер', '▁жақын', '▁мар', '▁тұра', '▁сары', '▁соның', '▁қаш', 'қыр', '▁себеп', '▁шеше', '▁бүгінгі', '▁таза', '▁шыр', 'іліп', 'деу', '▁көлік', '▁хат', 'даған', '▁арна', 'лерін', '▁яғ', '▁жаға', '▁жарық', '▁деңгей', '▁шаруа', '▁жерге', 'дім', 'меген', 'бас', 'ки', '▁темір', '▁болу', '▁асыр', '▁үстіне', 'шақ', '▁сәт', '▁әсер', 'дерін', '▁атты', '▁бейне', '▁мұра', '▁қасиет', '▁тар', '▁автор', '▁үшінші', '▁бауыр', '▁қазақтың', '▁мол', '▁өтті', '▁білдір', 'спе', '▁ете', '▁алыс', '▁орталығы', '▁кейбір', '▁бөлім', '▁х', '▁мәселе', '▁ірі', '▁қытай', '▁бұған', 'тарға', '▁ан', 'ц', '▁зе', '▁фа', '▁километр', '▁астам', '▁маман', '▁дәрігер', '▁кем', '▁жүрек', '▁қойды', '▁мекен', '▁сағат', 'шілер', '▁көрме', '▁тұрғындар', '▁ғасыр', '▁орал', 'лықтар', '▁ит', 'бей', 'ланып', 'дарды', 'гілікті', '▁мәдениет', '▁ойна', 'до', 'ология', '▁елдің', '▁жалға', '▁шө', '▁қасым', '▁даму', '▁ақпарат', '▁ресей', 'лігін', '▁жоспар', '▁басы', 'го', '▁облысы', '▁талай', '▁ва', '▁жә', 'ланған', 'лығын', '▁мұны', 'лықты', '▁қана', '▁айнала', '▁мұнда', '▁әлгі', '▁көтеріл', '▁кезең', '▁біраз', 'ейін', '▁сенің', '▁биік', '▁алаң', '▁құрылыс', '▁аяғы', '▁басып', '▁тарих', '▁келтір', '▁ора', '▁қауіп', '▁ұсын', 'шілік', '▁бұрынғы', '▁сүйе', '▁кен', '▁ұзақ', '▁әрине', '▁ұмыт', '▁ұя', '▁аузы', '▁жағына', 'сындағы', '▁тоғызыншы', '▁басын', '▁жылдан', '▁қап', 'деп', '▁жеткен', '▁ерте', 'уына', '▁талап', '▁ашыл', '▁ауыз', 'ілде', '▁бел', '▁бүтін', '▁қойған', '▁осылай', '▁әсіресе', '▁тын', 'уші', '▁көрді', 'лерге', '▁білмей', '▁шық', '▁қонақ', 'гіне', '▁тай', '▁қас', '▁той', '▁кон', 'иялық', '▁тағдыр', '▁құс', '▁арқа', '▁кеп', '▁негізгі', '▁жаман', '▁тұлға', '▁бөлме', '▁ұйымдастыр', '▁зор', '▁іске', '▁мүмкіндік', 'изм', 'һ', '▁дене', '▁әкел', '▁үміт', '▁қуан', '▁болмаса', '▁бүкіл', '▁сенім', '▁жатты', '▁рас', '▁төбе', '▁етті', 'лерді', '▁ши', '▁терең', 'тігін', '▁тәрбие', 'кір', '▁тұрады', 'сінде', '▁про', '▁сө', '▁но', '▁елбасы', '▁шешім', '▁сәл', '▁арналған', '▁құжат', 'тады', '▁құй', '▁жаза', 'қыл', '▁жазу', '▁маңызды', '▁жасау', '▁түрде', 'сінің', '▁спорт', '▁ли', '▁кә', '▁көл', '▁үйге', 'дерді', '▁зерттеу', '▁өтіп', '▁ішкі', '▁орнына', '▁отан', '▁әкесі', '▁бізге', '▁мін', '▁қалыптас', '▁кетеді', '▁еткен', '▁еле', '▁болсын', '▁алатын', '▁іші', '▁мұң', '▁басым', '▁тез', 'ыңды', '▁батыс', '▁жұмыстары', '▁ертең', 'айын', 'шыл', '▁пә', '▁көптеген', '▁тұрғындары', '▁әрекет', 'тырып', '▁алмас', 'хи', '▁үйрен', 'імде', '▁із', '▁елдер', '▁аспан', '▁шама', '▁тең', '▁сөздер', '▁ақыры', '▁бастаған', '▁оқып', 'жай', '▁нақты', '▁жу', '▁көркем', '▁тыңда', 'жи', 'дығын', '▁дегенмен', 'қтап', 'шысы', '▁мінез', '▁мақта', '▁теңіз', '▁жиі', '▁сауда', '▁сені', '▁ауа', 'фи', '▁тұратын', 'сар', '▁азаматтар', '▁әрбір', '▁рухани', '▁қаласы', '▁жасап', '▁күз', '▁сөзін', '▁сақ', '▁айналды', '▁жазып', '▁қалың', '▁қызық', '▁фо', '▁ету', '▁қойып', '▁әдет', '▁дәстүр', '▁жеңіл', '▁қысқа', '▁жазған', '▁университет', '▁кәсіпкер', '▁шаралар', '▁барады', '▁жайлы', 'инг', '▁бөлек', '▁жылдары', '▁төртінші', '▁шетел', '▁ауру', '▁ашық', '▁қарасты', '▁ағаш', '▁билік', '▁құрал', '▁жүйе', '▁есік', '▁республикасының', '▁басына', '▁бесінші', '▁қойма', '▁артық', '▁иә', 'асты', 'дырып', '▁табыс', '▁кейде', '▁студент', '▁күр', '▁ұйым', '▁таныс', '▁компания', '▁келетін', '▁шет', '▁жатады', '▁дайында', 'мейтін', '▁қоса', '▁киім', 'іңді', '▁мәселелер', '▁жүзеге', '▁шу', '▁өнім', '▁үлгі', '▁желі', '▁жұмса', '▁сұлтан', '▁тауып', '▁анық', '▁кіріс', '▁ұзын', '▁еркін', '▁енгіз', '▁жомарт', '▁қуаныш', '▁тартып', '▁ұстап', '▁қайыр', '▁жақта', '▁жетінші', '▁дін', '▁ұқса', '▁қатынас', '▁кар', '▁мұның', '▁өкілдері', '▁президенті', '▁мыс', '▁болмайды', '▁жарияла', '▁халықтың', '▁қалады', '▁бітір', '▁пайыз', '▁сегізінші', '▁оқ', '▁аралас', 'дарының', '▁алдына', '▁құрмет', '▁алтыншы', '▁сапар', '▁тарихи', '▁жарты', '▁солтүстік', '▁зар', '▁жуық', '▁баян', '▁айтады', 'ымның', '▁қолдау', '▁назар', '▁үйде', '▁оңтүстік', '▁ду', '▁роман', '▁шағын', '▁тым', '▁тобы', '▁мекеме', '▁шың', '▁фи', '▁дайын', '▁тәуелсіз', '▁сот', '▁елді', '▁күнге', '▁жүгір', '▁тура', 'құл', '▁алғаш', '▁солай', '▁былай', '▁кезек', '▁жақында', '▁иесі', '▁аумағы', 'қо', '▁махаббат', '▁апта', '▁беретін', '▁мынау', '▁бөлігі', '▁болашақ', '▁шебер', '▁тұрғыны', '▁жылғы', '▁жылдар', '▁тұрақты', '▁алдымен', '▁саяси', '▁саған', '▁сабақ', '▁күрес', '▁әскер', '▁бос', '▁тән', 'лайды', 'қтары', '▁айту', '▁сұлу', '▁оқырман', '▁нә', 'луы', '▁үстінде', '▁ашу', '▁мәлім', '▁бағытта', '▁бұлар', '▁келгенде', '▁ашып', '▁арасындағы', '▁табиғат', '▁көңілі', '▁өң', '▁қоғамдық', '▁патша', '▁медицина', '▁тірі', '▁орган', '▁дәрі', '▁қайғы', '▁бүгінде', '▁жүргізу', '▁қағаз', '▁ұрпақ', '▁ви', '▁сұрақ', '▁есепте', 'скен', '▁сәуле', '▁елімізде', '▁партия', '▁түсінді', '▁сәтте', '▁ауыс', '▁жасаған', '▁кеме', '▁аула', '▁шақ', '▁газет', '▁бағыт', '▁келіс', '▁айтпа', '▁алмады', '▁етеді', '▁еуропа', '▁саясат', '▁техника', '▁ыр', '▁байланыс', '▁шығады', '▁көрінеді', '▁биле', '▁тырыс', '▁кетпе', '▁жұма', '▁мәселен', '▁соны', '▁жасалған', '▁теле', '▁анасы', '▁бізде', '▁сақтау', '▁шекте', '▁ештеңе', '▁көлемі', '▁бағдарлама', '▁қоз', '▁құрайды', '▁жұлдыз', '▁түбі', '▁сосын', '▁әскери', '▁ру', '▁министр', '▁саты', '▁орынбасары', '▁мысалы', '▁жатып', '▁істер', '▁басында', '▁экономикалық', '▁толы', '▁жөнел', '▁дамыту', '▁жөнінде', '▁қанат', '▁жиын', '▁орай', '▁қылмыс', '▁салған', '▁әбден', '▁қожа', '▁қазақстандық', '▁көбей', '▁қоңыр', '▁тамақ', '▁тоқаев', '▁театр', '▁туынды', '▁жетті', '▁оңай', '▁сөйтіп', '▁келісім', '▁үзі', '▁аймақта', '▁метр', '▁мүлде', '▁әділ', '▁сайлау', '▁мұғалім', '▁сағын', '▁кіріп', '▁жүйесі', '▁тәжірибе', '▁әже', '▁журналист', '▁кино', '▁асыл', '▁кіші', '▁білген', '▁ескі', '▁қауым', '▁туыс', '▁ұстаз', '▁нысан', '▁үлес', '▁кей', '▁қаласында', '▁көктем', '▁қатысу', '▁жазылған', '▁жылқы', '▁ойлап', '▁үйіне', '▁келесі', '▁боса', '▁тыныш', '▁жіберді', '▁өлім', '▁амал', '▁көші', '▁қажетті', '▁жаққа', '▁шаруашылығы', '▁бағы', '▁оқи', '▁фильм', '▁өзіміз', 'фа', '▁дейтін', 'ья', '▁министрі', '▁қызметкерлер', '▁мәңгі', '▁бақылау', '▁тарихы', '▁кездесу', '▁аталған', '▁көтеріп', '▁тапқан', '▁төрағасы', '▁еліміздің', '▁адамдардың', '▁түйі', '▁тұңғыш', '▁шығу', '▁үздік', '▁орнат', '▁дамы', '▁мырза', '▁аясында', '▁айтуынша', '▁тақырып', '▁ұза', '▁сәйкес', '▁мереке', '▁тұрмыс', '▁назарбаев', '▁ве', '▁сына', 'ығып', '▁қарағанда', '▁түседі', '▁миллиард', 'екең', '▁жоғал', '▁мәселесі', '▁институт', '▁кішкентай', '▁экономика', '▁сіздің', '▁жанар', '▁еңбегі', '▁бағдарламасы', '▁аттан', '▁бермей', 'ыңа', '▁талқыла', '▁айналып', '▁өзгеріс', '▁қыздар', '▁көрші', '▁аурухана', '▁келсе', '▁жерлер', '▁мақала', '▁аяқтал', '▁қашан', '▁соғ', '▁иіс', '▁қалпы', '▁айналыс', '▁ыстық', '▁нәтиже', '▁керемет', '▁осылайша', '▁жігіттер', '▁білетін', '▁мағына', '▁түгел', '▁қуат', '▁өндіріс', '▁ерік', '▁поэзия', '▁дыбыс', '▁қамтамасыз', '▁ішіне', '▁кейіпкер', '▁шә', '▁адал', '▁га', '▁тамаша', '▁депутат', '▁кесі', '▁арқасында', '▁айқын', '▁полиция', '▁мәрте', '▁қош', '▁хо', '▁баяғы', '▁аңғар', '▁банк', '▁дәуір', '▁машина', '▁құнды', '▁сипат', '▁нұрсұлтан', '▁ешқандай', '▁ұшыра', '▁пәтер', '▁әдеби', '▁газ', '▁өлтір', '▁тәрізді', '▁тегін', '▁айырыл', '▁музыка', '▁қымбат', '▁жаңғыр', '▁авто', '▁көрініс', '▁қабырға', '▁бастама', 'қызы', '▁орындары', '▁өскен', '▁терезе', '▁қарапайым', '▁басқармасы', '▁қабат', '▁ойлан', '▁телефон', '▁қаламгер', '▁әдіс', '▁әдемі', '▁жүргізіл', '▁астында', '▁берілген', '▁барысында', '▁тыс', '▁шыда', '▁көпшілік', '▁кітабы', '▁кеуде', '▁бизнес', '▁округі', '▁кәрі', '▁қою', '▁құлағы', '▁тіршілік', '▁түркістан', '▁түркі', '▁ғалымдар', '▁кемпір', '▁ауданында', '▁директоры', '▁әкімшілік', '▁өкіл', '▁әнші', '▁нәтижесін', '▁қақпа', '▁құтқар', '▁әмір', '▁ғұмыр', '▁даусы', '▁шү', '▁түсінік', '▁ұғым', '▁күшей', '▁мұнай', '▁қимыл', '▁ғалым', '▁ұйқы', '▁кері', 'ылдап', '▁комитет', '▁қашықтықта', '▁наурыз', '▁технология', '▁ірге', '▁күтіп', '▁министрлігі', '▁белсенді', '▁елорда', '▁ұмтыл', '▁мүмкіндігі', '▁сонша', '▁ниет', '▁жалғасты', 'дырған', '▁заттар', '▁нұсқа', 'гүл', '▁мықты', 'тивті', '▁сәби', '▁тәртіп', '▁түзе', '▁мәліметтер', '▁шекара', '▁тиімді', '▁теріс', '▁муниципалитет', '▁асқар', '▁сыйла', '▁ақындар', '▁ғимарат', '▁сахна', '▁ұсыныс', '▁тамыр', '▁тұста', '▁зертте', '▁денсаулық', '▁әйтеуір', '▁тікелей', '▁рөл', '▁орман', '▁қозғал', '▁бойында', '▁іні', '▁америка', '▁қосымша', '▁идея', '▁екеуміз', '▁пайдалану', '▁өндір', '▁материал', '▁кәсіпорын', '▁мойында', '▁рух', '▁кенет', '▁бюджет', '▁ағылшын', '▁науқас', '▁кінә', '▁салынған', '▁құқық', '▁сарай', 'ейік', '▁былтыр', '▁ұйықта', '▁айналған', '▁оқушылар', '▁ойлаймын', '▁қозғалыс', '▁түйе', '▁ресми', '▁қырғыз', '▁рұқсат', '▁қарағанды', '▁қадам', '▁шамасында', '▁маусым', '▁адамзат', '▁жетекші', '▁ағайын', '▁мәдени', '▁мәскеу', '▁ахмет', '▁өсір', '▁үнемі', '▁орналас', '▁төтенше', '▁жәрдем', '▁маңдай', '▁жоғарғы', '▁астына', '▁бірқатар', '▁академия', '▁қабілет', '▁тигіз', '▁шығын', '▁көмектес', '▁талант', '▁жеміс', '▁домбыра', '▁мөлшер', '▁сағасы', '▁электр', '▁кәсіп', '▁көзқарас', '▁процес', '▁жаңбыр', '▁құтыл', '▁деймін', '▁зауыт', '▁еркек', '▁құбылыс', '▁дәреже', '▁ғалам', '▁бастайды', '▁қорытынды', '▁бейбіт', '▁қасқыр', '▁тіркел', '▁базар', '▁парламент', 'олог', '▁ешқашан', '▁чемпион', '▁германия', '▁мәжіліс', '▁табиғи', '▁өкініш', '▁қаражат', '▁тұрғыда', '▁жағалауынан', '▁жетпей', '▁алғыс', '▁шынайы', 'ский', '▁тонна', '▁күрделі', '▁топырақ', '▁жедел', '▁лайық', '▁доллар', '▁райы', '▁экс', '▁риза', '▁ажыра', '▁деректер', '▁комиссия', 'ъ', '▁коронавирус', '▁тәсіл', '▁сапасы', '▁танымал', '▁өтініш', '▁намыс', '▁журнал', '▁шаһар', '▁дөң', '▁көрермен', '▁демалыс', '▁салыстыр', '▁қайраткер', '▁тәңір', '▁өкпе', '▁оян', '▁көмегі', '▁мезгіл', '▁форма', '▁жамбыл', '▁штат', '▁одақ', '▁болмыс', '▁жабыл', '▁насихат', '▁ыңғай', '▁ғашық', '▁қараңғы', '▁бірдеңе', '▁тарапынан', '▁тергеу', '▁минут', '▁әріптес', '▁мерзім', '▁қамқор', '▁ендеше', '▁тексері', '▁бұйыр', '▁жапырақ', '▁көрсеткіш', '▁құрамында', '▁мойны', '▁департаменті', '▁премьер', '▁футбол', '▁сіңір', '▁мүшесі', 'щ', '▁көпбалалы', '▁ереже', '▁мектебі', '▁ықпал', '▁кедергі', 'дәулет', '▁құрамына', '▁қауіпсіздік', 'аяқ', '▁өсімдік', 'ё', '<sos/eos>'], token_type='bpe', train_data_path_and_name_and_type=[('dump/raw/train_sp/wav.scp', 'speech', 'sound'), ('dump/raw/train_sp/text', 'text', 'text')], train_dtype='float32', train_shape_file=['exp/asr_stats_raw_bpe2000_sp/logdir/train.1.scp'], unused_parameters=False, use_amp=True, use_matplotlib=True, use_preprocessor=True, use_tensorboard=True, use_wandb=False, val_scheduler_criterion=('valid', 'loss'), valid_batch_bins=None, valid_batch_size=None, valid_batch_type=None, valid_data_path_and_name_and_type=[('dump/raw/dev/wav.scp', 'speech', 'sound'), ('dump/raw/dev/text', 'text', 'text')], valid_max_cache_size=None, valid_shape_file=['exp/asr_stats_raw_bpe2000_sp/logdir/valid.1.scp'], version='202304', wandb_entity=None, wandb_id=None, wandb_model_log_interval=-1, wandb_name=None, wandb_project=None, write_collected_feats=False)
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,803 (asr:471) INFO: Optional Data Names: ('text_spk2', 'text_spk3', 'text_spk4')
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:12:29,823 (asr:471) INFO: Optional Data Names: ('text_spk2', 'text_spk3', 'text_spk4')
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:16:49,549 (collect_stats:107) INFO: Niter: 100
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:21:22,516 (collect_stats:107) INFO: Niter: 200
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:25:30,003 (collect_stats:107) INFO: Niter: 300
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:29:36,879 (collect_stats:107) INFO: Niter: 400
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:33:35,571 (collect_stats:107) INFO: Niter: 500
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:37:24,160 (collect_stats:107) INFO: Niter: 600
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:41:13,172 (collect_stats:107) INFO: Niter: 700
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:44:46,295 (collect_stats:107) INFO: Niter: 800
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:48:15,175 (collect_stats:107) INFO: Niter: 900
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:51:59,969 (collect_stats:107) INFO: Niter: 1000
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:52:51,717 (preprocessor:336) WARNING: The length of the text output exceeds 100, which may cause OOM on the GPU.Please ensure that the data processing is correct and verify it.
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:56:00,377 (collect_stats:107) INFO: Niter: 1100
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 13:59:57,946 (collect_stats:107) INFO: Niter: 1200
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:05:03,411 (collect_stats:107) INFO: Niter: 1300
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:09:46,755 (collect_stats:107) INFO: Niter: 1400
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:14:08,451 (collect_stats:107) INFO: Niter: 1500
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:18:21,971 (collect_stats:107) INFO: Niter: 1600
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:22:22,231 (collect_stats:107) INFO: Niter: 1700
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:26:18,394 (collect_stats:107) INFO: Niter: 1800
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:30:18,387 (collect_stats:107) INFO: Niter: 1900
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:34:13,006 (collect_stats:107) INFO: Niter: 2000
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:38:13,910 (collect_stats:107) INFO: Niter: 2100
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:41:22,834 (preprocessor:336) WARNING: The length of the text output exceeds 100, which may cause OOM on the GPU.Please ensure that the data processing is correct and verify it.
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:42:42,797 (collect_stats:107) INFO: Niter: 2200
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:47:01,489 (collect_stats:107) INFO: Niter: 2300
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:50:42,055 (collect_stats:107) INFO: Niter: 2400
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:54:36,834 (collect_stats:107) INFO: Niter: 2500
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 14:58:06,106 (collect_stats:107) INFO: Niter: 2600
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:01:39,637 (collect_stats:107) INFO: Niter: 2700
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:05:09,506 (collect_stats:107) INFO: Niter: 2800
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:08:35,702 (collect_stats:107) INFO: Niter: 2900
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:12:06,380 (collect_stats:107) INFO: Niter: 3000
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:15:36,267 (collect_stats:107) INFO: Niter: 3100
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:18:59,110 (collect_stats:107) INFO: Niter: 3200
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:22:30,986 (collect_stats:107) INFO: Niter: 3300
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:23:18,906 (preprocessor:336) WARNING: The length of the text output exceeds 100, which may cause OOM on the GPU.Please ensure that the data processing is correct and verify it.
[mlxlabq1l19yow63f8475a-20230224051258-1mabjw-np2rqw-worker] 2023-05-12 15:26:13,081 (collect_stats:107) INFO: Niter: 3400
# Accounting: time=8200 threads=1
# Ended (code 0) at Fri May 12 15:29:00 CST 2023, elapsed time 8200 seconds