kellywong commited on
Commit
67cdd43
1 Parent(s): 9436224

upload model and tokenizer

Browse files
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CSGModel"
4
+ ],
5
+ "beam_size": 12,
6
+ "ctx_max_seq_len": 1024,
7
+ "dropout": 0.2,
8
+ "embedding_dim": 500,
9
+ "eos_idx": 2,
10
+ "hidden_dim": 1024,
11
+ "initializer_range": 0.02,
12
+ "kernel_size": 3,
13
+ "num_aux_encoders": 3,
14
+ "num_decoders": 7,
15
+ "num_encoders": 7,
16
+ "padding_idx": 1,
17
+ "source_vocab_size": 30004,
18
+ "src_max_seq_len": 1024,
19
+ "target_vocab_size": 30004,
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.10.2",
22
+ "trg_max_seq_len": 1024
23
+ }
ctx_tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
ctx_tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<pad>"}
ctx_tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "pad_token": "<pad>", "tokenizer_class": "CSGTokenizer"}
ctx_tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c044cff3e715c2afb2dd7c89e07f3cc4e931810213f75470c40fbe3723893ea4
3
+ size 802816685
src_tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
src_tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<pad>"}
src_tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "pad_token": "<pad>", "tokenizer_class": "CSGTokenizer"}
src_tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
tgt_tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tgt_tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<pad>"}
tgt_tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "pad_token": "<pad>", "tokenizer_class": "CSGTokenizer"}
tgt_tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff