mathemakitten
commited on
Commit
•
9ab78a3
1
Parent(s):
77968c9
Training in progress, step 20000
Browse files- .gitignore +1 -0
- config.json +41 -0
- logs/1669193348.7292657/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.1 +3 -0
- logs/1669193545.402253/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.1 +3 -0
- logs/1669194183.4841528/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.1 +3 -0
- logs/1669195418.7645125/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.1 +3 -0
- logs/1669195541.478093/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.1 +3 -0
- logs/1669195669.2123573/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.1 +3 -0
- logs/1669195797.3350801/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.1 +3 -0
- logs/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.0 +3 -0
- logs/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.0 +3 -0
- logs/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.0 +3 -0
- logs/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.0 +3 -0
- logs/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.0 +3 -0
- logs/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.0 +3 -0
- logs/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.0 +3 -0
- merges.txt +0 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +15 -0
- tokenizer.json +0 -0
- tokenizer_config.json +23 -0
- training_args.bin +3 -0
- vocab.json +0 -0
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoint-*/
|
config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "gpt2-medium",
|
3 |
+
"activation_function": "gelu_new",
|
4 |
+
"architectures": [
|
5 |
+
"GPT2LMHeadModel"
|
6 |
+
],
|
7 |
+
"attn_pdrop": 0.1,
|
8 |
+
"bos_token_id": 50256,
|
9 |
+
"embd_pdrop": 0.1,
|
10 |
+
"eos_token_id": 50256,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"layer_norm_epsilon": 1e-05,
|
13 |
+
"model_type": "gpt2",
|
14 |
+
"n_ctx": 1024,
|
15 |
+
"n_embd": 1024,
|
16 |
+
"n_head": 16,
|
17 |
+
"n_inner": null,
|
18 |
+
"n_layer": 24,
|
19 |
+
"n_positions": 1024,
|
20 |
+
"n_special": 0,
|
21 |
+
"predict_special_tokens": true,
|
22 |
+
"reorder_and_upcast_attn": false,
|
23 |
+
"resid_pdrop": 0.1,
|
24 |
+
"scale_attn_by_inverse_layer_idx": false,
|
25 |
+
"scale_attn_weights": true,
|
26 |
+
"summary_activation": null,
|
27 |
+
"summary_first_dropout": 0.1,
|
28 |
+
"summary_proj_to_labels": true,
|
29 |
+
"summary_type": "cls_index",
|
30 |
+
"summary_use_proj": true,
|
31 |
+
"task_specific_params": {
|
32 |
+
"text-generation": {
|
33 |
+
"do_sample": true,
|
34 |
+
"max_length": 50
|
35 |
+
}
|
36 |
+
},
|
37 |
+
"torch_dtype": "float32",
|
38 |
+
"transformers_version": "4.24.0",
|
39 |
+
"use_cache": true,
|
40 |
+
"vocab_size": 50265
|
41 |
+
}
|
logs/1669193348.7292657/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f07382bf39b63c0ab960d00d78f146408b9298e6ae64df2b6cffce3fbf26ae37
|
3 |
+
size 5539
|
logs/1669193545.402253/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b29f2e1a346c2ae1db43290fa4d9ebfee3595424032c556b82d72abae8b7cf90
|
3 |
+
size 5546
|
logs/1669194183.4841528/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c485b7ecad3816dcfaa86795efee3e60597a1097cd05fb45eb720d78d9336c3
|
3 |
+
size 5546
|
logs/1669195418.7645125/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f35bc0ab696c1411e0d89bb3897f5fed60001a53d0987be0374ed182dfb5c6b6
|
3 |
+
size 5546
|
logs/1669195541.478093/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4099a90c5f86ac19a823800fac21b092caec170ca68ee879ec2c4aeb17640c72
|
3 |
+
size 5546
|
logs/1669195669.2123573/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1c04008000a774adb0f85cf1405039e57243de639cc4e38688e9ec6d8eed450
|
3 |
+
size 5546
|
logs/1669195797.3350801/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ab4b6ef490d4264ce96c7619e297319d8fef66302cbc0c5ffafae81ee127f39
|
3 |
+
size 5546
|
logs/events.out.tfevents.1669193348.helen-pytorch-xla-vm-vm.36480.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:715012a7f299b4d2d533341a9287e74353ae92c928bb51e8a392a68ee042b6c3
|
3 |
+
size 4085
|
logs/events.out.tfevents.1669193545.helen-pytorch-xla-vm-vm.37254.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30dcd6333f966006ed70279a7168ac4e286709b093832629ac6660ff4cd0eb49
|
3 |
+
size 4083
|
logs/events.out.tfevents.1669194183.helen-pytorch-xla-vm-vm.38388.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3796397e51d448f0813e0c5bd682f79990015515a7e0567296f08f71c7a291b1
|
3 |
+
size 4085
|
logs/events.out.tfevents.1669195418.helen-pytorch-xla-vm-vm.40518.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a6998db7e2ad03fa382cfd1e1aaaf121e39bd09ffc71a30c2b32e3498408f396
|
3 |
+
size 4085
|
logs/events.out.tfevents.1669195541.helen-pytorch-xla-vm-vm.41308.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6aa5458fa5b7195c7aa689087fae5571f6be29e7aba8d463734e256489157843
|
3 |
+
size 4085
|
logs/events.out.tfevents.1669195669.helen-pytorch-xla-vm-vm.42090.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e751a577360b9d5ba2086dded324fbf9a18fabeeed2477951dcf97dab1116702
|
3 |
+
size 4085
|
logs/events.out.tfevents.1669195797.helen-pytorch-xla-vm-vm.42877.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:950541d6c9e4103b40f33bf35a088895d5ab9400d509a8beccf12fe0352cef3e
|
3 |
+
size 35593
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7703f1818f03191a55d3155509a6be7c2fda52e89468175813d9f6d3c2560cc2
|
3 |
+
size 1444602141
|
special_tokens_map.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "<s>",
|
3 |
+
"cls_token": "<s>",
|
4 |
+
"eos_token": "</s>",
|
5 |
+
"mask_token": {
|
6 |
+
"content": "<mask>",
|
7 |
+
"lstrip": true,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"pad_token": "<pad>",
|
13 |
+
"sep_token": "</s>",
|
14 |
+
"unk_token": "<unk>"
|
15 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"bos_token": "<s>",
|
4 |
+
"cls_token": "<s>",
|
5 |
+
"eos_token": "</s>",
|
6 |
+
"errors": "replace",
|
7 |
+
"mask_token": {
|
8 |
+
"__type": "AddedToken",
|
9 |
+
"content": "<mask>",
|
10 |
+
"lstrip": true,
|
11 |
+
"normalized": false,
|
12 |
+
"rstrip": false,
|
13 |
+
"single_word": false
|
14 |
+
},
|
15 |
+
"model_max_length": 512,
|
16 |
+
"name_or_path": "Tristan/olm-tokenizer",
|
17 |
+
"pad_token": "<pad>",
|
18 |
+
"sep_token": "</s>",
|
19 |
+
"special_tokens_map_file": null,
|
20 |
+
"tokenizer_class": "RobertaTokenizer",
|
21 |
+
"trim_offsets": true,
|
22 |
+
"unk_token": "<unk>"
|
23 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0013a3044db87f1dde2c04d9ba3f5fe8f90b8f0bdbad2fd399ebe799e005e205
|
3 |
+
size 3451
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|