ArthurZ HF staff commited on
Commit
d13d0e4
1 Parent(s): 80bf224

Upload config

Browse files
Files changed (1) hide show
  1. config.json +73 -71
config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "_name_or_path": "openai/whisper-tiny.en",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
- "architectures": [
6
- "WhisperForConditionalGeneration"
7
- ],
8
  "attention_dropout": 0.0,
 
 
 
 
9
  "bos_token_id": 50257,
10
  "d_model": 384,
11
  "decoder_attention_heads": 6,
@@ -19,14 +19,16 @@
19
  "encoder_layerdrop": 0.0,
20
  "encoder_layers": 4,
21
  "eos_token_id": 50256,
22
- "feature_size": 1,
23
  "init_std": 0.02,
24
- "input_channels": 1,
25
  "is_encoder_decoder": true,
26
  "max_source_positions": 1500,
27
  "max_target_positions": 448,
28
  "model_type": "whisper",
29
- "non_speech_tokens": [
 
 
 
 
30
  1,
31
  2,
32
  6,
@@ -52,71 +54,71 @@
52
  91,
53
  92,
54
  93,
55
- 359,
56
- 503,
57
- 522,
58
- 542,
59
- 873,
60
- 893,
61
- 902,
62
- 918,
63
- 922,
64
- 931,
65
- 1350,
66
- 1853,
67
- 1982,
68
- 2460,
69
- 2627,
70
- 3246,
71
- 3253,
72
- 3268,
73
- 3536,
74
- 3846,
75
- 3961,
76
- 4183,
77
- 4667,
78
- 6585,
79
- 6647,
80
- 7273,
81
- 9061,
82
- 9383,
83
- 10428,
84
- 10929,
85
- 11938,
86
- 12033,
87
- 12331,
88
- 12562,
89
- 13793,
90
- 14157,
91
- 14635,
92
- 15265,
93
- 15618,
94
- 16553,
95
- 16604,
96
- 18362,
97
- 18956,
98
- 20075,
99
- 21675,
100
- 22520,
101
- 26130,
102
- 26161,
103
- 26435,
104
- 28279,
105
- 29464,
106
- 31650,
107
- 32302,
108
- 32470,
109
- 36865,
110
- 42863,
111
- 47425,
112
- 49870,
113
- 50254
 
 
 
 
 
114
  ],
115
- "num_hidden_layers": 4,
116
- "num_mel_bins": 80,
117
- "pad_token_id": 0,
118
- "scale_embedding": false,
119
- "torch_dtype": "float32",
120
  "transformers_version": "4.23.0.dev0",
121
  "use_cache": true,
122
  "vocab_size": 51864
 
1
  {
 
2
  "activation_dropout": 0.0,
3
  "activation_function": "gelu",
 
 
 
4
  "attention_dropout": 0.0,
5
+ "begin_suppress_tokens": [
6
+ 220,
7
+ 50256
8
+ ],
9
  "bos_token_id": 50257,
10
  "d_model": 384,
11
  "decoder_attention_heads": 6,
 
19
  "encoder_layerdrop": 0.0,
20
  "encoder_layers": 4,
21
  "eos_token_id": 50256,
 
22
  "init_std": 0.02,
 
23
  "is_encoder_decoder": true,
24
  "max_source_positions": 1500,
25
  "max_target_positions": 448,
26
  "model_type": "whisper",
27
+ "num_hidden_layers": 4,
28
+ "num_mel_bins": 80,
29
+ "pad_token_id": 0,
30
+ "scale_embedding": false,
31
+ "suppress_tokens": [
32
  1,
33
  2,
34
  6,
 
54
  91,
55
  92,
56
  93,
57
+ 357,
58
+ 366,
59
+ 438,
60
+ 532,
61
+ 685,
62
+ 705,
63
+ 796,
64
+ 930,
65
+ 1058,
66
+ 1220,
67
+ 1267,
68
+ 1279,
69
+ 1303,
70
+ 1343,
71
+ 1377,
72
+ 1391,
73
+ 1635,
74
+ 1782,
75
+ 1875,
76
+ 2162,
77
+ 2361,
78
+ 2488,
79
+ 3467,
80
+ 4008,
81
+ 4211,
82
+ 4600,
83
+ 4808,
84
+ 5299,
85
+ 5855,
86
+ 6329,
87
+ 7203,
88
+ 9609,
89
+ 9959,
90
+ 10563,
91
+ 10786,
92
+ 11420,
93
+ 11709,
94
+ 11907,
95
+ 13163,
96
+ 13697,
97
+ 13700,
98
+ 14808,
99
+ 15306,
100
+ 16410,
101
+ 16791,
102
+ 17992,
103
+ 19203,
104
+ 19510,
105
+ 20724,
106
+ 22305,
107
+ 22935,
108
+ 27007,
109
+ 30109,
110
+ 30420,
111
+ 33409,
112
+ 34949,
113
+ 40283,
114
+ 40493,
115
+ 40549,
116
+ 47282,
117
+ 49146,
118
+ 50257,
119
+ 50359,
120
+ 50360
121
  ],
 
 
 
 
 
122
  "transformers_version": "4.23.0.dev0",
123
  "use_cache": true,
124
  "vocab_size": 51864