mathemakitten commited on
Commit
9ab78a3
1 Parent(s): 77968c9

Training in progress, step 20000

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2-medium",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 1024,
16
+ "n_head": 16,
17
+ "n_inner": null,
18
+ "n_layer": 24,
19
+ "n_positions": 1024,
20
+ "n_special": 0,
21
+ "predict_special_tokens": true,
22
+ "reorder_and_upcast_attn": false,
23
+ "resid_pdrop": 0.1,
24
+ "scale_attn_by_inverse_layer_idx": false,
25
+ "scale_attn_weights": true,
26
+ "summary_activation": null,
27
+ "summary_first_dropout": 0.1,
28
+ "summary_proj_to_labels": true,
29
+ "summary_type": "cls_index",
30
+ "summary_use_proj": true,
31
+ "task_specific_params": {
32
+ "text-generation": {
33
+ "do_sample": true,
34
+ "max_length": 50
35
+ }
36
+ },
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.24.0",
39
+ "use_cache": true,
40
+ "vocab_size": 50265
41
+ }
logs/1669193348.7292657/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07382bf39b63c0ab960d00d78f146408b9298e6ae64df2b6cffce3fbf26ae37
3
+ size 5539
logs/1669193545.402253/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b29f2e1a346c2ae1db43290fa4d9ebfee3595424032c556b82d72abae8b7cf90
3
+ size 5546
logs/1669194183.4841528/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c485b7ecad3816dcfaa86795efee3e60597a1097cd05fb45eb720d78d9336c3
3
+ size 5546
logs/1669195418.7645125/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f35bc0ab696c1411e0d89bb3897f5fed60001a53d0987be0374ed182dfb5c6b6
3
+ size 5546
logs/1669195541.478093/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4099a90c5f86ac19a823800fac21b092caec170ca68ee879ec2c4aeb17640c72
3
+ size 5546
logs/1669195669.2123573/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1c04008000a774adb0f85cf1405039e57243de639cc4e38688e9ec6d8eed450
3
+ size 5546
logs/1669195797.3350801/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab4b6ef490d4264ce96c7619e297319d8fef66302cbc0c5ffafae81ee127f39
3
+ size 5546
logs/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:715012a7f299b4d2d533341a9287e74353ae92c928bb51e8a392a68ee042b6c3
3
+ size 4085
logs/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30dcd6333f966006ed70279a7168ac4e286709b093832629ac6660ff4cd0eb49
3
+ size 4083
logs/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3796397e51d448f0813e0c5bd682f79990015515a7e0567296f08f71c7a291b1
3
+ size 4085
logs/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6998db7e2ad03fa382cfd1e1aaaf121e39bd09ffc71a30c2b32e3498408f396
3
+ size 4085
logs/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6aa5458fa5b7195c7aa689087fae5571f6be29e7aba8d463734e256489157843
3
+ size 4085
logs/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e751a577360b9d5ba2086dded324fbf9a18fabeeed2477951dcf97dab1116702
3
+ size 4085
logs/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:950541d6c9e4103b40f33bf35a088895d5ab9400d509a8beccf12fe0352cef3e
3
+ size 35593
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7703f1818f03191a55d3155509a6be7c2fda52e89468175813d9f6d3c2560cc2
3
+ size 1444602141
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "errors": "replace",
7
+ "mask_token": {
8
+ "__type": "AddedToken",
9
+ "content": "<mask>",
10
+ "lstrip": true,
11
+ "normalized": false,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ },
15
+ "model_max_length": 512,
16
+ "name_or_path": "Tristan/olm-tokenizer",
17
+ "pad_token": "<pad>",
18
+ "sep_token": "</s>",
19
+ "special_tokens_map_file": null,
20
+ "tokenizer_class": "RobertaTokenizer",
21
+ "trim_offsets": true,
22
+ "unk_token": "<unk>"
23
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0013a3044db87f1dde2c04d9ba3f5fe8f90b8f0bdbad2fd399ebe799e005e205
3
+ size 3451
vocab.json ADDED
The diff for this file is too large to render. See raw diff