m3hrdadfi commited on
Commit
5142f2a
1 Parent(s): c2bd648
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ step*
2
+ epoch*
added_tokens.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<bos>": 50257,
3
+ "<context>": 50263,
4
+ "<dialog>": 50264,
5
+ "<eos>": 50258,
6
+ "<mask>": 50260,
7
+ "<pad>": 50259,
8
+ "<persona>": 50262,
9
+ "<sep>": 50261
10
+ }
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2DoubleHeadsModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "aux_num_labels": 2,
9
+ "bos_token_id": 50257,
10
+ "class_names": [],
11
+ "embd_pdrop": 0.1,
12
+ "eos_token_id": 50258,
13
+ "id2label": {
14
+ "0": "LABEL_0"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_epsilon": 1e-05,
21
+ "model_type": "gpt2",
22
+ "n_ctx": 1024,
23
+ "n_embd": 768,
24
+ "n_head": 12,
25
+ "n_inner": null,
26
+ "n_layer": 12,
27
+ "n_positions": 1024,
28
+ "pad_token_id": 50259,
29
+ "reorder_and_upcast_attn": false,
30
+ "resid_pdrop": 0.1,
31
+ "scale_attn_by_inverse_layer_idx": false,
32
+ "scale_attn_weights": true,
33
+ "summary_activation": null,
34
+ "summary_first_dropout": 0.1,
35
+ "summary_proj_to_labels": true,
36
+ "summary_type": "cls_index",
37
+ "summary_use_proj": true,
38
+ "task_specific_params": {
39
+ "text-generation": {
40
+ "do_sample": true,
41
+ "max_length": 50
42
+ }
43
+ },
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.25.1",
46
+ "use_cache": true,
47
+ "vocab_size": 50265
48
+ }
logs/AuxGPT2-Research-Alvis-alvis-dd-urb-gpt2-small-context/version_08-02-2023--09-28-58/events.out.tfevents.1675844940.alvis4-31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ffb42bce0f31297cf97c38d49b3bc464151d11bf0b24ed323b5f9f2fd0737a1
3
+ size 109372
logs/AuxGPT2-Research-Alvis-alvis-dd-urb-gpt2-small-context/version_08-02-2023--09-28-58/events.out.tfevents.1675846749.alvis4-31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d06f7471ab037362bfe72643c3eaadb512766776b2bb7b47dc58a99e4a0427b
3
+ size 1739
logs/AuxGPT2-Research-Alvis-alvis-dd-urb-gpt2-small-context/version_08-02-2023--09-28-58/hparams.yaml ADDED
The diff for this file is too large to render. See raw diff
 
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6de213223d92e3c5a77abceee95e590bd76fdd6f24931cf0db4c2d2d0aefda1b
3
+ size 512799253
special_tokens_map.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<sep>",
4
+ "<persona>",
5
+ "<context>",
6
+ "<dialog>"
7
+ ],
8
+ "bos_token": "<bos>",
9
+ "eos_token": "<eos>",
10
+ "mask_token": "<mask>",
11
+ "pad_token": "<pad>",
12
+ "unk_token": "<|endoftext|>"
13
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "model_max_length": 1024,
6
+ "name_or_path": "gpt2",
7
+ "special_tokens_map_file": null,
8
+ "tokenizer_class": "GPT2Tokenizer",
9
+ "unk_token": "<|endoftext|>"
10
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff