mizoru commited on
Commit
e3dd71f
1 Parent(s): 23a804a

Training in progress, step 500

Browse files
config.json CHANGED
@@ -23,7 +23,20 @@
23
  "encoder_layerdrop": 0.0,
24
  "encoder_layers": 4,
25
  "eos_token_id": 50257,
26
- "forced_decoder_ids": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  "init_std": 0.02,
28
  "is_encoder_decoder": true,
29
  "max_length": 448,
@@ -34,9 +47,96 @@
34
  "num_mel_bins": 80,
35
  "pad_token_id": 50257,
36
  "scale_embedding": false,
37
- "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  "torch_dtype": "float32",
39
  "transformers_version": "4.26.0.dev0",
40
- "use_cache": true,
41
  "vocab_size": 51865
42
  }
 
23
  "encoder_layerdrop": 0.0,
24
  "encoder_layers": 4,
25
  "eos_token_id": 50257,
26
+ "forced_decoder_ids": [
27
+ [
28
+ 1,
29
+ 50259
30
+ ],
31
+ [
32
+ 2,
33
+ 50359
34
+ ],
35
+ [
36
+ 3,
37
+ 50363
38
+ ]
39
+ ],
40
  "init_std": 0.02,
41
  "is_encoder_decoder": true,
42
  "max_length": 448,
 
47
  "num_mel_bins": 80,
48
  "pad_token_id": 50257,
49
  "scale_embedding": false,
50
+ "suppress_tokens": [
51
+ 1,
52
+ 2,
53
+ 7,
54
+ 8,
55
+ 9,
56
+ 10,
57
+ 14,
58
+ 25,
59
+ 26,
60
+ 27,
61
+ 28,
62
+ 29,
63
+ 31,
64
+ 58,
65
+ 59,
66
+ 60,
67
+ 61,
68
+ 62,
69
+ 63,
70
+ 90,
71
+ 91,
72
+ 92,
73
+ 93,
74
+ 359,
75
+ 503,
76
+ 522,
77
+ 542,
78
+ 873,
79
+ 893,
80
+ 902,
81
+ 918,
82
+ 922,
83
+ 931,
84
+ 1350,
85
+ 1853,
86
+ 1982,
87
+ 2460,
88
+ 2627,
89
+ 3246,
90
+ 3253,
91
+ 3268,
92
+ 3536,
93
+ 3846,
94
+ 3961,
95
+ 4183,
96
+ 4667,
97
+ 6585,
98
+ 6647,
99
+ 7273,
100
+ 9061,
101
+ 9383,
102
+ 10428,
103
+ 10929,
104
+ 11938,
105
+ 12033,
106
+ 12331,
107
+ 12562,
108
+ 13793,
109
+ 14157,
110
+ 14635,
111
+ 15265,
112
+ 15618,
113
+ 16553,
114
+ 16604,
115
+ 18362,
116
+ 18956,
117
+ 20075,
118
+ 21675,
119
+ 22520,
120
+ 26130,
121
+ 26161,
122
+ 26435,
123
+ 28279,
124
+ 29464,
125
+ 31650,
126
+ 32302,
127
+ 32470,
128
+ 36865,
129
+ 42863,
130
+ 47425,
131
+ 49870,
132
+ 50254,
133
+ 50258,
134
+ 50360,
135
+ 50361,
136
+ 50362
137
+ ],
138
  "torch_dtype": "float32",
139
  "transformers_version": "4.26.0.dev0",
140
+ "use_cache": false,
141
  "vocab_size": 51865
142
  }
preprocessor_config.json CHANGED
@@ -5,7 +5,7 @@
5
  "hop_length": 160,
6
  "mel_filters": [
7
  [
8
- 0.0,
9
  0.02486259490251541,
10
  0.0,
11
  0.0,
 
5
  "hop_length": 160,
6
  "mel_filters": [
7
  [
8
+ -0.0,
9
  0.02486259490251541,
10
  0.0,
11
  0.0,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:465475659f18dde8aa2c1a3a5ba06be498f1a1cc839983f0c7acd24cada6c46e
3
- size 151098921
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a7cc1a8e87999784b75fbf721c783c47fa3c6e363fa9af5eefd5c5331f8b7c9
3
+ size 151097331
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25c3a5f76ac6a2d0463f3b9fb39c424bcd9b015c983da62efdb458ed542b21fa
3
- size 3643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7886f5995da873e3b3ac58a33ef481ba3e5ead4654206bf20d7915beb9243995
3
+ size 3567