voidful commited on
Commit
8b1605a
1 Parent(s): d6c348e

increase model

Browse files
Files changed (2) hide show
  1. config.json +18 -3
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -10,15 +10,30 @@
10
  "classifier_proj_size": 256,
11
  "conv_bias": false,
12
  "conv_dim": [
 
 
 
 
 
13
  512,
14
  512
15
  ],
16
  "conv_kernel": [
17
  10,
18
- 3
 
 
 
 
 
19
  ],
20
  "conv_stride": [
21
  5,
 
 
 
 
 
22
  2
23
  ],
24
  "ctc_loss_reduction": "sum",
@@ -47,11 +62,11 @@
47
  "num_attention_heads": 12,
48
  "num_conv_pos_embedding_groups": 16,
49
  "num_conv_pos_embeddings": 128,
50
- "num_feat_extract_layers": 2,
51
  "num_hidden_layers": 1,
52
  "pad_token_id": 0,
53
  "torch_dtype": "float32",
54
- "transformers_version": "4.27.0.dev0",
55
  "use_weighted_layer_sum": false,
56
  "vocab_size": 32
57
  }
 
10
  "classifier_proj_size": 256,
11
  "conv_bias": false,
12
  "conv_dim": [
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
  512,
19
  512
20
  ],
21
  "conv_kernel": [
22
  10,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 3,
27
+ 2,
28
+ 2
29
  ],
30
  "conv_stride": [
31
  5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
  2
38
  ],
39
  "ctc_loss_reduction": "sum",
 
62
  "num_attention_heads": 12,
63
  "num_conv_pos_embedding_groups": 16,
64
  "num_conv_pos_embeddings": 128,
65
+ "num_feat_extract_layers": 7,
66
  "num_hidden_layers": 1,
67
  "pad_token_id": 0,
68
  "torch_dtype": "float32",
69
+ "transformers_version": "4.26.1",
70
  "use_weighted_layer_sum": false,
71
  "vocab_size": 32
72
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00d79d8e070bebb2872331165bc7df1ed41b3fd02848e90ce2c91d062b77e43b
3
- size 51993669
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97cce607ce1d2a01dda6deca8b1398eaa2d10d52cc28d4d8fae9bfb6b711cf4f
3
+ size 65626772