emre commited on
Commit
a22698a
1 Parent(s): 5a2429c

Initial commit

Browse files
all_results.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 25.0,
3
+ "eval_loss": 1.7208493947982788,
4
+ "eval_mem_cpu_alloc_delta": 104295,
5
+ "eval_mem_cpu_peaked_delta": 513049,
6
+ "eval_mem_gpu_alloc_delta": 0,
7
+ "eval_mem_gpu_peaked_delta": 5843399680,
8
+ "eval_runtime": 2.9265,
9
+ "eval_samples": 20,
10
+ "eval_samples_per_second": 6.834,
11
+ "init_mem_cpu_alloc_delta": 336056,
12
+ "init_mem_cpu_peaked_delta": 18306,
13
+ "init_mem_gpu_alloc_delta": 509558784,
14
+ "init_mem_gpu_peaked_delta": 0,
15
+ "perplexity": 5.58927394881408,
16
+ "train_mem_cpu_alloc_delta": 846748,
17
+ "train_mem_cpu_peaked_delta": 752929,
18
+ "train_mem_gpu_alloc_delta": 1501306368,
19
+ "train_mem_gpu_peaked_delta": 3372484096,
20
+ "train_runtime": 1074.2277,
21
+ "train_samples": 184,
22
+ "train_samples_per_second": 4.282
23
+ }
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/gpt2/finetuned",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 0,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 2,
11
+ "gradient_checkpointing": false,
12
+ "initializer_range": 0.02,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "model_type": "gpt2",
15
+ "n_ctx": 1024,
16
+ "n_embd": 768,
17
+ "n_head": 12,
18
+ "n_inner": null,
19
+ "n_layer": 12,
20
+ "n_positions": 1024,
21
+ "resid_pdrop": 0.1,
22
+ "summary_activation": null,
23
+ "summary_first_dropout": 0.1,
24
+ "summary_proj_to_labels": true,
25
+ "summary_type": "cls_index",
26
+ "summary_use_proj": true,
27
+ "transformers_version": "4.4.2",
28
+ "use_cache": true,
29
+ "vocab_size": 50000
30
+ }
eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 25.0,
3
+ "eval_loss": 1.7208493947982788,
4
+ "eval_mem_cpu_alloc_delta": 104295,
5
+ "eval_mem_cpu_peaked_delta": 513049,
6
+ "eval_mem_gpu_alloc_delta": 0,
7
+ "eval_mem_gpu_peaked_delta": 5843399680,
8
+ "eval_runtime": 2.9265,
9
+ "eval_samples": 20,
10
+ "eval_samples_per_second": 6.834,
11
+ "perplexity": 5.58927394881408
12
+ }
merges.docx ADDED
Binary file (988 kB). View file
merges.txt ADDED
The diff for this file is too large to render. See raw diff
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d010aed8fecf2bb774f1fb809f4b77039c4d78b708c9ec20741e87054965a47
3
+ size 509618811
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>"}
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f6dc4ce5d994738a0b5ac09e67066a7c31dec4b2980c3911b49cce91823c1ec
3
+ size 497144144
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"errors": "replace", "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "special_tokens_map_file": null, "name_or_path": "/content/drive/MyDrive/gpt2/finetuned", "tokenizer_file": null}
train_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 25.0,
3
+ "init_mem_cpu_alloc_delta": 336056,
4
+ "init_mem_cpu_peaked_delta": 18306,
5
+ "init_mem_gpu_alloc_delta": 509558784,
6
+ "init_mem_gpu_peaked_delta": 0,
7
+ "train_mem_cpu_alloc_delta": 846748,
8
+ "train_mem_cpu_peaked_delta": 752929,
9
+ "train_mem_gpu_alloc_delta": 1501306368,
10
+ "train_mem_gpu_peaked_delta": 3372484096,
11
+ "train_runtime": 1074.2277,
12
+ "train_samples": 184,
13
+ "train_samples_per_second": 4.282
14
+ }
trainer_state.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 25.0,
5
+ "global_step": 4600,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 2.72,
12
+ "learning_rate": 4.456521739130435e-05,
13
+ "loss": 2.6463,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 5.43,
18
+ "learning_rate": 3.91304347826087e-05,
19
+ "loss": 1.5589,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 8.15,
24
+ "learning_rate": 3.369565217391305e-05,
25
+ "loss": 1.107,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 10.87,
30
+ "learning_rate": 2.826086956521739e-05,
31
+ "loss": 0.8178,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 13.59,
36
+ "learning_rate": 2.282608695652174e-05,
37
+ "loss": 0.6328,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 16.3,
42
+ "learning_rate": 1.739130434782609e-05,
43
+ "loss": 0.515,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 19.02,
48
+ "learning_rate": 1.1956521739130435e-05,
49
+ "loss": 0.4357,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 21.74,
54
+ "learning_rate": 6.521739130434783e-06,
55
+ "loss": 0.3796,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 24.46,
60
+ "learning_rate": 1.0869565217391306e-06,
61
+ "loss": 0.3478,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 25.0,
66
+ "step": 4600,
67
+ "total_flos": 3511389310156800.0,
68
+ "train_runtime": 1074.2277,
69
+ "train_samples_per_second": 4.282
70
+ }
71
+ ],
72
+ "max_steps": 4600,
73
+ "num_train_epochs": 25,
74
+ "total_flos": 3511389310156800.0,
75
+ "trial_name": null,
76
+ "trial_params": null
77
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e28a6df1626a16a0a69053e61921133f6c89a402aaf1f0c97b9703214bbb28
3
+ size 2287
vocab.json ADDED
The diff for this file is too large to render. See raw diff