0xsuid commited on
Commit
ed9109c
1 Parent(s): 8e3761e

Remove Extra files

Browse files
Files changed (49) hide show
  1. experiments/.gitkeep +0 -0
  2. experiments/configs.json +0 -26
  3. experiments/deepspeed.json +0 -48
  4. experiments/final_checkpoint/config.json +0 -54
  5. experiments/final_checkpoint/pytorch_model.bin +0 -3
  6. experiments/final_checkpoint/tokenizer/merges.txt +0 -0
  7. experiments/final_checkpoint/tokenizer/special_tokens_map.json +0 -24
  8. experiments/final_checkpoint/tokenizer/tokenizer.json +0 -0
  9. experiments/final_checkpoint/tokenizer/tokenizer_config.json +0 -34
  10. experiments/final_checkpoint/tokenizer/vocab.json +0 -0
  11. experiments/logs/1676990060.264247/events.out.tfevents.1676990060.7b1eed3780f4.2829.2 +0 -3
  12. experiments/logs/1676990154.028005/events.out.tfevents.1676990154.7b1eed3780f4.5567.2 +0 -3
  13. experiments/logs/1676990712.4404476/events.out.tfevents.1676990712.7b1eed3780f4.10223.2 +0 -3
  14. experiments/logs/1677006993.51447/events.out.tfevents.1677006993.7b1eed3780f4.46131.2 +0 -3
  15. experiments/logs/1677007417.411844/events.out.tfevents.1677007417.7b1eed3780f4.51624.2 +0 -3
  16. experiments/logs/1677007657.7754965/events.out.tfevents.1677007657.7b1eed3780f4.56432.2 +0 -3
  17. experiments/logs/1677007781.5676754/events.out.tfevents.1677007781.7b1eed3780f4.59255.2 +0 -3
  18. experiments/logs/1677007927.0654259/events.out.tfevents.1677007927.7b1eed3780f4.62309.2 +0 -3
  19. experiments/logs/1677008561.9460855/events.out.tfevents.1677008561.7b1eed3780f4.3852.2 +0 -3
  20. experiments/logs/1677009134.735858/events.out.tfevents.1677009134.7b1eed3780f4.11652.2 +0 -3
  21. experiments/logs/events.out.tfevents.1676990060.7b1eed3780f4.2829.1 +0 -3
  22. experiments/logs/events.out.tfevents.1676990154.7b1eed3780f4.5567.1 +0 -3
  23. experiments/logs/events.out.tfevents.1676990712.7b1eed3780f4.10223.1 +0 -3
  24. experiments/logs/events.out.tfevents.1677006993.7b1eed3780f4.46131.1 +0 -3
  25. experiments/logs/events.out.tfevents.1677007417.7b1eed3780f4.51624.1 +0 -3
  26. experiments/logs/events.out.tfevents.1677007657.7b1eed3780f4.56432.1 +0 -3
  27. experiments/logs/events.out.tfevents.1677007781.7b1eed3780f4.59255.1 +0 -3
  28. experiments/logs/events.out.tfevents.1677007927.7b1eed3780f4.62309.1 +0 -3
  29. experiments/logs/events.out.tfevents.1677008561.7b1eed3780f4.3852.1 +0 -3
  30. experiments/logs/events.out.tfevents.1677009134.7b1eed3780f4.11652.1 +0 -3
  31. experiments/logs/train_neo/events.out.tfevents.1676989999.7b1eed3780f4.2829.0 +0 -3
  32. experiments/logs/train_neo/events.out.tfevents.1676990139.7b1eed3780f4.5567.0 +0 -3
  33. experiments/logs/train_neo/events.out.tfevents.1676990697.7b1eed3780f4.10223.0 +0 -3
  34. experiments/logs/train_neo/events.out.tfevents.1677006978.7b1eed3780f4.46131.0 +0 -3
  35. experiments/logs/train_neo/events.out.tfevents.1677007402.7b1eed3780f4.51624.0 +0 -3
  36. experiments/logs/train_neo/events.out.tfevents.1677007642.7b1eed3780f4.56432.0 +0 -3
  37. experiments/logs/train_neo/events.out.tfevents.1677007765.7b1eed3780f4.59255.0 +0 -3
  38. experiments/logs/train_neo/events.out.tfevents.1677007911.7b1eed3780f4.62309.0 +0 -3
  39. experiments/logs/train_neo/events.out.tfevents.1677008546.7b1eed3780f4.3852.0 +0 -3
  40. experiments/logs/train_neo/events.out.tfevents.1677009119.7b1eed3780f4.11652.0 +0 -3
  41. experiments/output.log +0 -0
  42. experiments/trainer_final_checkpoint/config.json +0 -54
  43. experiments/trainer_final_checkpoint/merges.txt +0 -0
  44. experiments/trainer_final_checkpoint/pytorch_model.bin +0 -3
  45. experiments/trainer_final_checkpoint/special_tokens_map.json +0 -24
  46. experiments/trainer_final_checkpoint/tokenizer.json +0 -0
  47. experiments/trainer_final_checkpoint/tokenizer_config.json +0 -34
  48. experiments/trainer_final_checkpoint/training_args.bin +0 -3
  49. experiments/trainer_final_checkpoint/vocab.json +0 -0
experiments/.gitkeep DELETED
File without changes
experiments/configs.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "output_dir": "./results",
3
- "evaluation_strategy": "no",
4
- "do_eval": false,
5
- "eval_steps": 0,
6
- "log_level": "info",
7
- "logging_first_step": true,
8
- "logging_steps": 5,
9
- "logging_dir": "./logs",
10
- "save_steps": 150,
11
- "save_total_limit": 1,
12
- "num_train_epochs": 10,
13
- "per_device_train_batch_size": 6,
14
- "optim": "adamw_torch",
15
- "gradient_accumulation_steps": 4,
16
- "dataloader_drop_last": true,
17
- "warmup_steps": 1000,
18
- "weight_decay": 0.1,
19
- "learning_rate": 0.0001,
20
- "deepspeed": "deepspeed.json",
21
- "local_rank": 0,
22
- "total_gpus": 8,
23
- "v_cpus": 64,
24
- "total_memory_in_gb": 257604.9765625,
25
- "dataset_limit": 0
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/deepspeed.json DELETED
@@ -1,48 +0,0 @@
1
- {
2
- "optimizer": {
3
- "type": "AdamW",
4
- "params": {
5
- "lr": 1e-4,
6
- "betas": [ 0.9, 0.999 ],
7
- "eps": 1e-8,
8
- "weight_decay": 0.1
9
- }
10
- },
11
-
12
- "scheduler": {
13
- "type": "WarmupLR",
14
- "params": {
15
- "warmup_min_lr": 1e-9,
16
- "warmup_max_lr": 1e-4,
17
- "warmup_num_steps": 1000
18
- }
19
- },
20
-
21
- "zero_optimization": {
22
- "stage": 2,
23
- "offload_optimizer": {
24
- "device": "cpu",
25
- "pin_memory": true
26
- },
27
- "allgather_partitions": true,
28
- "allgather_bucket_size": 5e8,
29
- "overlap_comm": true,
30
- "reduce_scatter": true,
31
- "reduce_bucket_size": 5e8,
32
- "contiguous_gradients": true
33
- },
34
-
35
- "tensorboard": {
36
- "enabled": true,
37
- "output_path": "logs/",
38
- "job_name": "train_neo"
39
- },
40
-
41
- "zero_allow_untested_optimizer": true,
42
- "gradient_accumulation_steps": "auto",
43
- "gradient_clipping": "auto",
44
- "steps_per_print": 2000,
45
- "train_batch_size": "auto",
46
- "train_micro_batch_size_per_gpu": "auto",
47
- "wall_clock_breakdown": false
48
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/final_checkpoint/config.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "_name_or_path": "EleutherAI/gpt-neo-125M",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPTNeoForCausalLM"
6
- ],
7
- "attention_dropout": 0,
8
- "attention_layers": [
9
- "global",
10
- "local",
11
- "global",
12
- "local",
13
- "global",
14
- "local",
15
- "global",
16
- "local",
17
- "global",
18
- "local",
19
- "global",
20
- "local"
21
- ],
22
- "attention_types": [
23
- [
24
- [
25
- "global",
26
- "local"
27
- ],
28
- 6
29
- ]
30
- ],
31
- "bos_token_id": 50256,
32
- "embed_dropout": 0,
33
- "eos_token_id": 50256,
34
- "gradient_checkpointing": false,
35
- "hidden_size": 768,
36
- "initializer_range": 0.02,
37
- "intermediate_size": null,
38
- "layer_norm_epsilon": 1e-05,
39
- "max_position_embeddings": 2048,
40
- "model_type": "gpt_neo",
41
- "num_heads": 12,
42
- "num_layers": 12,
43
- "resid_dropout": 0,
44
- "summary_activation": null,
45
- "summary_first_dropout": 0.1,
46
- "summary_proj_to_labels": true,
47
- "summary_type": "cls_index",
48
- "summary_use_proj": true,
49
- "torch_dtype": "float32",
50
- "transformers_version": "4.24.0",
51
- "use_cache": true,
52
- "vocab_size": 50257,
53
- "window_size": 256
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/final_checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6589651dd8c9588a8faa97daec7814ffcfe998ba2e481eabdcb3e75405761d5d
3
- size 551154684
 
 
 
 
experiments/final_checkpoint/tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
experiments/final_checkpoint/tokenizer/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/final_checkpoint/tokenizer/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
experiments/final_checkpoint/tokenizer/tokenizer_config.json DELETED
@@ -1,34 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "eos_token": {
13
- "__type": "AddedToken",
14
- "content": "<|endoftext|>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "errors": "replace",
21
- "model_max_length": 2048,
22
- "name_or_path": "EleutherAI/gpt-neo-125M",
23
- "pad_token": null,
24
- "special_tokens_map_file": null,
25
- "tokenizer_class": "GPT2Tokenizer",
26
- "unk_token": {
27
- "__type": "AddedToken",
28
- "content": "<|endoftext|>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false
33
- }
34
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/final_checkpoint/tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
experiments/logs/1676990060.264247/events.out.tfevents.1676990060.7b1eed3780f4.2829.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0a44882a6cf48f00bbae625e93c4eb24bf47ac0723da1b60bba90e403527478
3
- size 5418
 
 
 
 
experiments/logs/1676990154.028005/events.out.tfevents.1676990154.7b1eed3780f4.5567.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8cad42700950e422799625ed456e575bab05a5a8d59a135ad217bc04118fa3a
3
- size 5418
 
 
 
 
experiments/logs/1676990712.4404476/events.out.tfevents.1676990712.7b1eed3780f4.10223.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:604efe07f7a9c4930e77e489c3df413af1a8323edd27cdee3d2f1d128f4c1c89
3
- size 5418
 
 
 
 
experiments/logs/1677006993.51447/events.out.tfevents.1677006993.7b1eed3780f4.46131.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a22c2f7b4fbb4cd79ea6415d3c51f9a09dc73db339caa3f5f5eb586fe1302e8a
3
- size 5418
 
 
 
 
experiments/logs/1677007417.411844/events.out.tfevents.1677007417.7b1eed3780f4.51624.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8abe3ac2c851eefaa9ea98dbcabd1ddf10d0de80c47b521651f31e3a50b2b785
3
- size 5418
 
 
 
 
experiments/logs/1677007657.7754965/events.out.tfevents.1677007657.7b1eed3780f4.56432.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fd51f599cb26d493f515bc23292f9207bc3c12926fd96a21adef93e0d48ab7a
3
- size 5418
 
 
 
 
experiments/logs/1677007781.5676754/events.out.tfevents.1677007781.7b1eed3780f4.59255.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f59371ae067572e2bac707ffc12c5a6b195952c9f8a02efd1f13071407fdef34
3
- size 5418
 
 
 
 
experiments/logs/1677007927.0654259/events.out.tfevents.1677007927.7b1eed3780f4.62309.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:583410e80db49e1fd00d700134bc0c7b940a282a02db4ea1c04e343b2cf95d53
3
- size 5418
 
 
 
 
experiments/logs/1677008561.9460855/events.out.tfevents.1677008561.7b1eed3780f4.3852.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f9718ff8c295f15850eae9e52a47d89e4fe679205f0878dbbebae43b69e924e
3
- size 5418
 
 
 
 
experiments/logs/1677009134.735858/events.out.tfevents.1677009134.7b1eed3780f4.11652.2 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:de0699c3ccb09a5e8169901a5a82bf8626cdb9173b5ef41fa9b1ed6130d9e876
3
- size 5418
 
 
 
 
experiments/logs/events.out.tfevents.1676990060.7b1eed3780f4.2829.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d60506ce0259929b66d97f59ae51e9741a0439206a3e90b9eafafbb1d4e55d99
3
- size 4972
 
 
 
 
experiments/logs/events.out.tfevents.1676990154.7b1eed3780f4.5567.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c40f4b630877ec383adc308e311d8a0055d151bdba93a915ac8627ac21f703b
3
- size 7128
 
 
 
 
experiments/logs/events.out.tfevents.1676990712.7b1eed3780f4.10223.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc84a6719de976bab0c9c89d033c452a0d36fe758d4ea7b29b7744044ba17bd8
3
- size 93460
 
 
 
 
experiments/logs/events.out.tfevents.1677006993.7b1eed3780f4.46131.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f8491b10d8281d0ca19c6f340ada66b121750b085ed36267b6d7915a0a7d100
3
- size 5126
 
 
 
 
experiments/logs/events.out.tfevents.1677007417.7b1eed3780f4.51624.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9138322d0249bf3e3664827b969c7e1ba7ecfeed12561218b0107aff49d5db6
3
- size 4663
 
 
 
 
experiments/logs/events.out.tfevents.1677007657.7b1eed3780f4.56432.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:82770df39e009cdd002b40f4ef5a003d0c10180e1d9e9cff927bdfcc28829f0f
3
- size 4356
 
 
 
 
experiments/logs/events.out.tfevents.1677007781.7b1eed3780f4.59255.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:231b985f36d43f817df84e527d0a9e2a2c51f5fba78f1312ad0ce84cd167bd34
3
- size 4357
 
 
 
 
experiments/logs/events.out.tfevents.1677007927.7b1eed3780f4.62309.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9434bfc711ed73659875ea734d5ef6c0a3c4da8a9988dbfda3bc8d15d6eb18f4
3
- size 5280
 
 
 
 
experiments/logs/events.out.tfevents.1677008561.7b1eed3780f4.3852.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e598b68031d0e4b024d334bbdd8f9578605d37afad9e12883d40569898022704
3
- size 6357
 
 
 
 
experiments/logs/events.out.tfevents.1677009134.7b1eed3780f4.11652.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7beda0e65719bc42ea79ca8271a49e66236fe8a545d6df0f15e0bccf4d151cc6
3
- size 196022
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1676989999.7b1eed3780f4.2829.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2b32ba109e1360ab5b662d33639c8269cff54e5c6b2b3db72ddbd367a99fac3
3
- size 3282
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1676990139.7b1eed3780f4.5567.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fe69eaf4075333c1258fd65cc71b802cdf1bb6cbe92a208bbbccb6da2e45670
3
- size 11778
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1676990697.7b1eed3780f4.10223.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b135479273cad4c3731d46a46bbd33836789a12fbc8410939f6db61548beca21
3
- size 341626
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677006978.7b1eed3780f4.46131.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a7183fffe76771d6f41c9327c7a4e712703dd37770facb441f9ec4187e0f4d2
3
- size 4049
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677007402.7b1eed3780f4.51624.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fc001b4583c619b0ac7b8b1d9f87ee06c44e21bb8b608157e4ace4e3f213985
3
- size 2043
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677007642.7b1eed3780f4.56432.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2fc5f10ff92471189468976c3bfcdd39a381332cd35e72beddaac389a968102
3
- size 627
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677007765.7b1eed3780f4.59255.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9b1c3b9a91703f47c9076edd408eba70052a2c5b5377145092942f48daa0943
3
- size 863
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677007911.7b1eed3780f4.62309.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5aed7d3a1b9b70a512390aa06f65ee6f4ee67bb22476ba9262c40aa1cd9900c
3
- size 4167
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677008546.7b1eed3780f4.3852.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:753aebddba05c1ac14d703f0ff956fd7046abe63675889cde20723c61e83e6db
3
- size 8297
 
 
 
 
experiments/logs/train_neo/events.out.tfevents.1677009119.7b1eed3780f4.11652.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:060fae387dbb47c447acafec429237e7bfbdbb16d0fc159199e60e3f9fcf0155
3
- size 732346
 
 
 
 
experiments/output.log DELETED
The diff for this file is too large to render. See raw diff
 
experiments/trainer_final_checkpoint/config.json DELETED
@@ -1,54 +0,0 @@
1
- {
2
- "_name_or_path": "EleutherAI/gpt-neo-125M",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPTNeoForCausalLM"
6
- ],
7
- "attention_dropout": 0,
8
- "attention_layers": [
9
- "global",
10
- "local",
11
- "global",
12
- "local",
13
- "global",
14
- "local",
15
- "global",
16
- "local",
17
- "global",
18
- "local",
19
- "global",
20
- "local"
21
- ],
22
- "attention_types": [
23
- [
24
- [
25
- "global",
26
- "local"
27
- ],
28
- 6
29
- ]
30
- ],
31
- "bos_token_id": 50256,
32
- "embed_dropout": 0,
33
- "eos_token_id": 50256,
34
- "gradient_checkpointing": false,
35
- "hidden_size": 768,
36
- "initializer_range": 0.02,
37
- "intermediate_size": null,
38
- "layer_norm_epsilon": 1e-05,
39
- "max_position_embeddings": 2048,
40
- "model_type": "gpt_neo",
41
- "num_heads": 12,
42
- "num_layers": 12,
43
- "resid_dropout": 0,
44
- "summary_activation": null,
45
- "summary_first_dropout": 0.1,
46
- "summary_proj_to_labels": true,
47
- "summary_type": "cls_index",
48
- "summary_use_proj": true,
49
- "torch_dtype": "float32",
50
- "transformers_version": "4.24.0",
51
- "use_cache": true,
52
- "vocab_size": 50257,
53
- "window_size": 256
54
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/trainer_final_checkpoint/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
experiments/trainer_final_checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6589651dd8c9588a8faa97daec7814ffcfe998ba2e481eabdcb3e75405761d5d
3
- size 551154684
 
 
 
 
experiments/trainer_final_checkpoint/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/trainer_final_checkpoint/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
experiments/trainer_final_checkpoint/tokenizer_config.json DELETED
@@ -1,34 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "eos_token": {
13
- "__type": "AddedToken",
14
- "content": "<|endoftext|>",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "errors": "replace",
21
- "model_max_length": 2048,
22
- "name_or_path": "EleutherAI/gpt-neo-125M",
23
- "pad_token": null,
24
- "special_tokens_map_file": null,
25
- "tokenizer_class": "GPT2Tokenizer",
26
- "unk_token": {
27
- "__type": "AddedToken",
28
- "content": "<|endoftext|>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false
33
- }
34
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
experiments/trainer_final_checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:69255842ed78e689b179c20bae1499576b967ef499d84d9ec162b162d779e910
3
- size 4475
 
 
 
 
experiments/trainer_final_checkpoint/vocab.json DELETED
The diff for this file is too large to render. See raw diff