shi-zheng-qxhs commited on
Commit
1fbcd63
1 Parent(s): c22eb15

End of training

Browse files
README.md CHANGED
@@ -41,7 +41,7 @@ The following hyperparameters were used during training:
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - lr_scheduler_warmup_steps: 1000
44
- - num_epochs: 10
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
 
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - lr_scheduler_warmup_steps: 1000
44
+ - num_epochs: 20
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|assistant|>": 50258,
3
+ "<|end|>": 50259,
4
+ "<|user|>": 50257
5
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a2385a1e1b43f6e9509e1960824cd63d58c00a80eec8ec19490a76412036b98
3
  size 497783424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6622b8293cafa420ddb16018bea8b80f708eef57b54a7e4eb08400a68466b9ef
3
  size 497783424
special_tokens_map.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|user|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|assistant|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<|end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ ],
25
+ "bos_token": "<|endoftext|>",
26
+ "eos_token": "<|endoftext|>",
27
+ "pad_token": "<|endoftext|>",
28
+ "unk_token": "<|endoftext|>"
29
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "50257": {
13
+ "content": "<|user|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "50258": {
21
+ "content": "<|assistant|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "50259": {
29
+ "content": "<|end|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ }
36
+ },
37
+ "additional_special_tokens": [
38
+ "<|user|>",
39
+ "<|assistant|>",
40
+ "<|end|>"
41
+ ],
42
+ "bos_token": "<|endoftext|>",
43
+ "clean_up_tokenization_spaces": true,
44
+ "eos_token": "<|endoftext|>",
45
+ "model_max_length": 1024,
46
+ "pad_token": "<|endoftext|>",
47
+ "tokenizer_class": "GPT2Tokenizer",
48
+ "unk_token": "<|endoftext|>"
49
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14ac7e96a73e85f57f3c59fdfbe1c8088ad20217736e61b40756da453784dec1
3
  size 4219
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5996e56c3426d31bb70a71b8368922aa4c3ba2c6a278a886bc7dd471ec565b4d
3
  size 4219
vocab.json ADDED
The diff for this file is too large to render. See raw diff