m3hrdadfi commited on
Commit
3005219
1 Parent(s): f8786ee
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
1
+ step*
2
+ epoch*
added_tokens.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<bos>": 50257,
3
+ "<context>": 50263,
4
+ "<dialog>": 50264,
5
+ "<eos>": 50258,
6
+ "<mask>": 50260,
7
+ "<pad>": 50259,
8
+ "<persona>": 50262,
9
+ "<sep>": 50261
10
+ }
config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2DoubleHeadsModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "aux_num_labels": 50265,
9
+ "bos_token_id": 50257,
10
+ "class_names": [],
11
+ "embd_pdrop": 0.1,
12
+ "eos_token_id": 50258,
13
+ "id2label": {
14
+ "0": "LABEL_0"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_epsilon": 1e-05,
21
+ "model_type": "gpt2",
22
+ "n_ctx": 1024,
23
+ "n_embd": 1024,
24
+ "n_head": 16,
25
+ "n_inner": null,
26
+ "n_layer": 24,
27
+ "n_positions": 1024,
28
+ "n_special": 0,
29
+ "pad_token_id": 50259,
30
+ "predict_special_tokens": true,
31
+ "reorder_and_upcast_attn": false,
32
+ "resid_pdrop": 0.1,
33
+ "scale_attn_by_inverse_layer_idx": false,
34
+ "scale_attn_weights": true,
35
+ "summary_activation": null,
36
+ "summary_first_dropout": 0.1,
37
+ "summary_proj_to_labels": true,
38
+ "summary_type": "cls_index",
39
+ "summary_use_proj": true,
40
+ "task_specific_params": {
41
+ "text-generation": {
42
+ "do_sample": true,
43
+ "max_length": 50
44
+ }
45
+ },
46
+ "torch_dtype": "float32",
47
+ "transformers_version": "4.26.0",
48
+ "use_cache": true,
49
+ "vocab_size": 50265
50
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50257,
4
+ "eos_token_id": 50258,
5
+ "pad_token_id": 50259,
6
+ "transformers_version": "4.26.0"
7
+ }
logs/AuxGPT2-Research-Alvis-alvis-pc-umt-gpt2-medium-random/version_05-02-2023--08-53-37/events.out.tfevents.1675583622.alvis4-42 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6b21b593d28da76c5efe25afef1a26b64e60aae7f7506dab01912618b5efd83
3
+ size 1258159
logs/AuxGPT2-Research-Alvis-alvis-pc-umt-gpt2-medium-random/version_05-02-2023--08-53-37/events.out.tfevents.1675622340.alvis4-42 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:add116036a80a49af9d8163c1cb35fd4338f286356963b4c168a13d8fb434469
3
+ size 1751
logs/AuxGPT2-Research-Alvis-alvis-pc-umt-gpt2-medium-random/version_05-02-2023--08-53-37/hparams.yaml ADDED
The diff for this file is too large to render. See raw diff
merges.txt ADDED
The diff for this file is too large to render. See raw diff
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f5f90c2c8a377f17ff63d287767f9a139d5ea7c9dc38dd4473a99e2d67d528
3
+ size 1654897269
special_tokens_map.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<sep>",
4
+ "<persona>",
5
+ "<context>",
6
+ "<dialog>"
7
+ ],
8
+ "bos_token": "<bos>",
9
+ "eos_token": "<eos>",
10
+ "mask_token": "<mask>",
11
+ "pad_token": "<pad>",
12
+ "unk_token": "<|endoftext|>"
13
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "model_max_length": 1024,
6
+ "name_or_path": "gpt2",
7
+ "special_tokens_map_file": null,
8
+ "tokenizer_class": "GPT2Tokenizer",
9
+ "unk_token": "<|endoftext|>"
10
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff