hoangbinhmta99 commited on
Commit
10b672b
1 Parent(s): b6bc685

Update my model

Browse files
3gram.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e1c2d9b5f5a4f0dbe60c6b2fe288df84705c7e0ce601b3720757967a1a49ac1
3
+ size 337389627
4gram.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e2fcfc875e22de373f738eb1fd1dc52a6f9070a00a9f6f4dd34e9048ee2ac24
3
+ size 889960226
audio-test/t1_0001-00010.wav ADDED
Binary file (120 kB). View file
 
audio-test/t1_utt000000042.wav ADDED
Binary file (76.8 kB). View file
 
audio-test/t2_0000006682.wav ADDED
Binary file (49.6 kB). View file
 
config.json CHANGED
@@ -1,15 +1,11 @@
1
  {
2
  "activation_dropout": 0.1,
3
- "adapter_kernel_size": 3,
4
- "adapter_stride": 2,
5
- "add_adapter": false,
6
  "apply_spec_augment": true,
7
  "architectures": [
8
  "Wav2Vec2ForPreTraining"
9
  ],
10
  "attention_dropout": 0.1,
11
  "bos_token_id": 1,
12
- "classifier_proj_size": 256,
13
  "codevector_dim": 256,
14
  "contrastive_logits_temperature": 0.1,
15
  "conv_bias": false,
@@ -47,9 +43,10 @@
47
  "eos_token_id": 2,
48
  "feat_extract_activation": "gelu",
49
  "feat_extract_norm": "group",
50
- "feat_proj_dropout": 0.0,
51
  "feat_quantizer_dropout": 0.0,
52
  "final_dropout": 0.1,
 
53
  "hidden_act": "gelu",
54
  "hidden_dropout": 0.1,
55
  "hidden_size": 768,
@@ -58,13 +55,10 @@
58
  "layer_norm_eps": 1e-05,
59
  "layerdrop": 0.1,
60
  "mask_feature_length": 10,
61
- "mask_feature_min_masks": 0,
62
  "mask_feature_prob": 0.0,
63
  "mask_time_length": 10,
64
- "mask_time_min_masks": 2,
65
  "mask_time_prob": 0.05,
66
  "model_type": "wav2vec2",
67
- "num_adapter_layers": 3,
68
  "num_attention_heads": 12,
69
  "num_codevector_groups": 2,
70
  "num_codevectors_per_group": 320,
@@ -73,33 +67,9 @@
73
  "num_feat_extract_layers": 7,
74
  "num_hidden_layers": 12,
75
  "num_negatives": 100,
76
- "output_hidden_size": 768,
77
  "pad_token_id": 0,
78
  "proj_codevector_dim": 256,
79
- "tdnn_dilation": [
80
- 1,
81
- 2,
82
- 3,
83
- 1,
84
- 1
85
- ],
86
- "tdnn_dim": [
87
- 512,
88
- 512,
89
- 512,
90
- 512,
91
- 1500
92
- ],
93
- "tdnn_kernel": [
94
- 5,
95
- 3,
96
- 3,
97
- 1,
98
- 1
99
- ],
100
  "torch_dtype": "float32",
101
- "transformers_version": "4.15.0",
102
- "use_weighted_layer_sum": false,
103
- "vocab_size": 32,
104
- "xvector_output_dim": 512
105
  }
 
1
  {
2
  "activation_dropout": 0.1,
 
 
 
3
  "apply_spec_augment": true,
4
  "architectures": [
5
  "Wav2Vec2ForPreTraining"
6
  ],
7
  "attention_dropout": 0.1,
8
  "bos_token_id": 1,
 
9
  "codevector_dim": 256,
10
  "contrastive_logits_temperature": 0.1,
11
  "conv_bias": false,
 
43
  "eos_token_id": 2,
44
  "feat_extract_activation": "gelu",
45
  "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
  "feat_quantizer_dropout": 0.0,
48
  "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
  "hidden_act": "gelu",
51
  "hidden_dropout": 0.1,
52
  "hidden_size": 768,
 
55
  "layer_norm_eps": 1e-05,
56
  "layerdrop": 0.1,
57
  "mask_feature_length": 10,
 
58
  "mask_feature_prob": 0.0,
59
  "mask_time_length": 10,
 
60
  "mask_time_prob": 0.05,
61
  "model_type": "wav2vec2",
 
62
  "num_attention_heads": 12,
63
  "num_codevector_groups": 2,
64
  "num_codevectors_per_group": 320,
 
67
  "num_feat_extract_layers": 7,
68
  "num_hidden_layers": 12,
69
  "num_negatives": 100,
 
70
  "pad_token_id": 0,
71
  "proj_codevector_dim": 256,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
  "torch_dtype": "float32",
73
+ "transformers_version": "4.9.2",
74
+ "vocab_size": 32
 
 
75
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd7cbbde57f35aeb4397b8c692e121b1b5f74cf9638e98d2209f4ea4edad4847
3
  size 380261837
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7ba6a0b8b9710efbd751fb872e89ddcbb50e62c08f0002fdef6b155ac4ac6f
3
  size 380261837
vocab.json CHANGED
@@ -1 +1 @@
1
- {"<s>": 1, "<pad>": 0, "</s>": 2, "<unk>": 3, "|": 4, "E": 5, "T": 6, "A": 7, "O": 8, "N": 9, "I": 10, "H": 11, "S": 12, "R": 13, "D": 14, "L": 15, "U": 16, "M": 17, "W": 18, "C": 19, "F": 20, "G": 21, "Y": 22, "P": 23, "B": 24, "V": 25, "K": 26, "'": 27, "X": 28, "J": 29, "Q": 30, "Z": 31}
 
1
+ {"<s>": 1, "<pad>": 0, "</s>": 2, "<unk>": 3, "N": 4, "H": 5, "T": 6, "I": 7, "C": 8, "G": 9, "A": 10, "M": 11, "U": 12, "Đ": 13, "À": 14, "O": 15, "V": 16, "L": 17, "Ư": 18, "R": 19, "Á": 20, "Y": 21, "B": 22, "P": 23, "K": 24, "Ô": 25, "S": 26, "Ó": 27, "": 28, "": 29, "": 30, "Ờ": 31, "Ì": 32, "Ê": 33, "Ả": 34, "Ệ": 35, "D": 36, "Â": 37, "Ố": 38, "Ớ": 39, "Ấ": 40, "Ơ": 41, "Ề": 42, "Q": 43, "Ủ": 44, "Ể": 45, "Ă": 46, "Ợ": 47, "Ị": 48, "E": 49, "Ậ": 50, "Í": 51, "X": 52, "Ầ": 53, "Ú": 54, "Ự": 55, "Ữ": 56, "Ọ": 57, "Ứ": 58, "Ở": 59, "Ã": 60, "Ồ": 61, "Ắ": 62, "Ụ": 63, "Ừ": 64, "Ò": 65, "Ổ": 66, "Ù": 67, "Ũ": 68, "Ặ": 69, "Ẽ": 70, "Ý": 71, "Ỉ": 72, "Ỏ": 73, "Ử": 74, "Ằ": 75, "É": 76, "Ĩ": 77, "Ễ": 78, "Ẩ": 79, "<": 80, ">": 81, "Ẫ": 82, "Ỗ": 83, "Ẹ": 84, "Ỹ": 85, "Ẻ": 86, "Ỳ": 87, "È": 88, "Õ": 89, "Ỡ": 90, "Ẳ": 91, "Ỷ": 92, "Ẵ": 93, "F": 94, "W": 95, "Z": 96, "J": 97, "Ỵ": 98, "'": 99}