Muennighoff commited on
Commit
1d4971c
1 Parent(s): 42f0298
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. sp-1b-100m/checkpoint-100/config.json +33 -0
  2. sp-1b-100m/checkpoint-100/optimizer.pt +3 -0
  3. sp-1b-100m/checkpoint-100/pytorch_model.bin +3 -0
  4. sp-1b-100m/checkpoint-100/rng_state.pth +3 -0
  5. sp-1b-100m/checkpoint-100/scaler.pt +3 -0
  6. sp-1b-100m/checkpoint-100/scheduler.pt +3 -0
  7. sp-1b-100m/checkpoint-100/trainer_state.json +15 -0
  8. sp-1b-100m/checkpoint-100/training_args.bin +3 -0
  9. sp-1b-100m/checkpoint-200/config.json +33 -0
  10. sp-1b-100m/checkpoint-200/optimizer.pt +3 -0
  11. sp-1b-100m/checkpoint-200/pytorch_model.bin +3 -0
  12. sp-1b-100m/checkpoint-200/rng_state.pth +3 -0
  13. sp-1b-100m/checkpoint-200/scaler.pt +3 -0
  14. sp-1b-100m/checkpoint-200/scheduler.pt +3 -0
  15. sp-1b-100m/checkpoint-200/trainer_state.json +15 -0
  16. sp-1b-100m/checkpoint-200/training_args.bin +3 -0
  17. sp-1b-100m/checkpoint-300/config.json +33 -0
  18. sp-1b-100m/checkpoint-300/optimizer.pt +3 -0
  19. sp-1b-100m/checkpoint-300/pytorch_model.bin +3 -0
  20. sp-1b-100m/checkpoint-300/rng_state.pth +3 -0
  21. sp-1b-100m/checkpoint-300/scaler.pt +3 -0
  22. sp-1b-100m/checkpoint-300/scheduler.pt +3 -0
  23. sp-1b-100m/checkpoint-300/trainer_state.json +15 -0
  24. sp-1b-100m/checkpoint-300/training_args.bin +3 -0
  25. sp-1b-100m/config.json +33 -0
  26. sp-1b-100m/pytorch_model.bin +3 -0
  27. sp-1b-100m/training_args.bin +3 -0
  28. sp-200m-100m/checkpoint-100/config.json +33 -0
  29. sp-200m-100m/checkpoint-100/optimizer.pt +3 -0
  30. sp-200m-100m/checkpoint-100/pytorch_model.bin +3 -0
  31. sp-200m-100m/checkpoint-100/rng_state.pth +3 -0
  32. sp-200m-100m/checkpoint-100/scaler.pt +3 -0
  33. sp-200m-100m/checkpoint-100/scheduler.pt +3 -0
  34. sp-200m-100m/checkpoint-100/trainer_state.json +15 -0
  35. sp-200m-100m/checkpoint-100/training_args.bin +3 -0
  36. sp-200m-100m/checkpoint-200/config.json +33 -0
  37. sp-200m-100m/checkpoint-200/optimizer.pt +3 -0
  38. sp-200m-100m/checkpoint-200/pytorch_model.bin +3 -0
  39. sp-200m-100m/checkpoint-200/rng_state.pth +3 -0
  40. sp-200m-100m/checkpoint-200/scaler.pt +3 -0
  41. sp-200m-100m/checkpoint-200/scheduler.pt +3 -0
  42. sp-200m-100m/checkpoint-200/trainer_state.json +15 -0
  43. sp-200m-100m/checkpoint-200/training_args.bin +3 -0
  44. sp-200m-100m/checkpoint-300/config.json +33 -0
  45. sp-200m-100m/checkpoint-300/optimizer.pt +3 -0
  46. sp-200m-100m/checkpoint-300/pytorch_model.bin +3 -0
  47. sp-200m-100m/checkpoint-300/rng_state.pth +3 -0
  48. sp-200m-100m/checkpoint-300/scaler.pt +3 -0
  49. sp-200m-100m/checkpoint-300/scheduler.pt +3 -0
  50. sp-200m-100m/checkpoint-300/trainer_state.json +15 -0
sp-1b-100m/checkpoint-100/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 12288,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 3072,
15
+ "n_head": 64,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-1b-100m/checkpoint-100/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d33cc8ed49fde04377a4aa3b5c21aa69213a091f7f5eee135712782dea5c282c
3
+ size 12135889757
sp-1b-100m/checkpoint-100/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dbc01cfdc8748bf8b0a9fa8dce3ff0340d7e03b03df4b3c8864b1cb491dc0d5
3
+ size 6080539805
sp-1b-100m/checkpoint-100/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba78ebce0e6b5b3f4e93d2b800e3ebc1f9c36976d4963bf270e6783fb8c0606e
3
+ size 14575
sp-1b-100m/checkpoint-100/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfa44e8523f62833816d29aa6c576eaa7783e3bbdb3e132e248b1d8aaee6132b
3
+ size 557
sp-1b-100m/checkpoint-100/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7ea57e999fee966b1265f956366ee033f13ca6a2b608b4432ce2605ac0d88ed
3
+ size 627
sp-1b-100m/checkpoint-100/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.26309839468870116,
5
+ "global_step": 100,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 2.138214034833408e+17,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
sp-1b-100m/checkpoint-100/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feee864b9593fe1171ce64e635566351bc726e347de5c1946387ce337455125d
3
+ size 3387
sp-1b-100m/checkpoint-200/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 12288,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 3072,
15
+ "n_head": 64,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-1b-100m/checkpoint-200/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e916bfdb2400f3353cd7e42f749d2a3d9394cbf6a1b13445fc17d9ef4f86052
3
+ size 12135889757
sp-1b-100m/checkpoint-200/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a7e467ee0f58c2493d2e8f8afd547c0ca9ce47cbdd3a3328f22dbdea1f68f3a
3
+ size 6080539805
sp-1b-100m/checkpoint-200/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3b95dbfc53cdce2495f06e1c586eed0d4a66d02153a8ae6886b56fc37002ed2
3
+ size 14575
sp-1b-100m/checkpoint-200/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fccf0f9be1bb8f24861e4393745b3e09cc2687125a69e3757955fb0f0925ea5
3
+ size 557
sp-1b-100m/checkpoint-200/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c162896f6df3ba4475b56099d964ded55f9fe1ecbaa5d3b3cdb9215d9d04af81
3
+ size 627
sp-1b-100m/checkpoint-200/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.5261967893774023,
5
+ "global_step": 200,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 4.276428069666816e+17,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
sp-1b-100m/checkpoint-200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feee864b9593fe1171ce64e635566351bc726e347de5c1946387ce337455125d
3
+ size 3387
sp-1b-100m/checkpoint-300/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 12288,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 3072,
15
+ "n_head": 64,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-1b-100m/checkpoint-300/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:263c995c30ba2be19018d3b20db9aee7b61dcb9de7ff86807cee7d72ed0304f7
3
+ size 12135889885
sp-1b-100m/checkpoint-300/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32075e5f905116b3d1be94622c51f8fe46cc06b58ff08cafaa0a1b3b57f85af1
3
+ size 6080539805
sp-1b-100m/checkpoint-300/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2188bfd57d858ee450d8255287416c27d03305b15e592c12098f02f2ed02d870
3
+ size 14575
sp-1b-100m/checkpoint-300/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efafd90182e3d39d1b7c4a686f86e5913f5abc094dc3e2f827a6d479c6cef247
3
+ size 557
sp-1b-100m/checkpoint-300/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4546d9532f4bee54bea33da91e5c41d94c56696c6723d0717ee8575a4c90c582
3
+ size 627
sp-1b-100m/checkpoint-300/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.7892951840661034,
5
+ "global_step": 300,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 6.414642104500224e+17,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
sp-1b-100m/checkpoint-300/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feee864b9593fe1171ce64e635566351bc726e347de5c1946387ce337455125d
3
+ size 3387
sp-1b-100m/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 12288,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 3072,
15
+ "n_head": 64,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-1b-100m/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5304288c5cc05479cbe6d40c5f0bce44426cd1c6f53fb430d6cb290d31448fdc
3
+ size 6080539805
sp-1b-100m/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feee864b9593fe1171ce64e635566351bc726e347de5c1946387ce337455125d
3
+ size 3387
sp-200m-100m/checkpoint-100/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 1024,
15
+ "n_head": 32,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-200m-100m/checkpoint-100/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dca74745ea576c9ba7c6580733d0b9544223a6ac9dab358dd8838b71b6d1a50
3
+ size 1629434309
sp-200m-100m/checkpoint-100/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9de15ae087dea9bd556536a910cf273304f0d48353fca437be4a4f7df6061ad
3
+ size 827312701
sp-200m-100m/checkpoint-100/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66ebba15e85e086af7736fdcf05e9cd682b1a2e05cbcf39415aaa86a3b656a0f
3
+ size 14575
sp-200m-100m/checkpoint-100/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfa44e8523f62833816d29aa6c576eaa7783e3bbdb3e132e248b1d8aaee6132b
3
+ size 557
sp-200m-100m/checkpoint-100/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7ea57e999fee966b1265f956366ee033f13ca6a2b608b4432ce2605ac0d88ed
3
+ size 627
sp-200m-100m/checkpoint-100/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.2630929869275672,
5
+ "global_step": 100,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 2.37748988411904e+16,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
sp-200m-100m/checkpoint-100/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:334dd1c91498da0d2485dba103030ffcef7471025d2adba3a950d4a720fba08b
3
+ size 3387
sp-200m-100m/checkpoint-200/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 1024,
15
+ "n_head": 32,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-200m-100m/checkpoint-200/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e13a3e586d7bfbd082efe67bd40654dda1133a17b52c58f540d9d4c038ef0237
3
+ size 1629434309
sp-200m-100m/checkpoint-200/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c77811a3465bf64a82617298848e9a3d6b369e7af6cdb5fa0ca9a287315eaead
3
+ size 827312701
sp-200m-100m/checkpoint-200/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4851f3b302b7c4fee9d9420148364c058cf57804100694ae01ab586d33879732
3
+ size 14575
sp-200m-100m/checkpoint-200/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fccf0f9be1bb8f24861e4393745b3e09cc2687125a69e3757955fb0f0925ea5
3
+ size 557
sp-200m-100m/checkpoint-200/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c162896f6df3ba4475b56099d964ded55f9fe1ecbaa5d3b3cdb9215d9d04af81
3
+ size 627
sp-200m-100m/checkpoint-200/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.5261859738551344,
5
+ "global_step": 200,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 4.75497976823808e+16,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
sp-200m-100m/checkpoint-200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:334dd1c91498da0d2485dba103030ffcef7471025d2adba3a950d4a720fba08b
3
+ size 3387
sp-200m-100m/checkpoint-300/config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_embd": 1024,
15
+ "n_head": 32,
16
+ "n_inner": null,
17
+ "n_layer": 12,
18
+ "n_positions": 1024,
19
+ "num_layers": 12,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.25.1",
31
+ "use_cache": false,
32
+ "vocab_size": 50257
33
+ }
sp-200m-100m/checkpoint-300/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f6a6179d4f91596677cc4638dfb4fb9cc993e03f9098e4c09312c54c3427cd0
3
+ size 1629434437
sp-200m-100m/checkpoint-300/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60b9a2117ba0c4fa31230300d872c83eee01d63f134deb9a34d6edbb8833bb41
3
+ size 827312701
sp-200m-100m/checkpoint-300/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08950dc609e5df5b063c350f814f906ffba4fb0870677e17789bafc766cec4b7
3
+ size 14575
sp-200m-100m/checkpoint-300/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efafd90182e3d39d1b7c4a686f86e5913f5abc094dc3e2f827a6d479c6cef247
3
+ size 557
sp-200m-100m/checkpoint-300/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4546d9532f4bee54bea33da91e5c41d94c56696c6723d0717ee8575a4c90c582
3
+ size 627
sp-200m-100m/checkpoint-300/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.7892789607827017,
5
+ "global_step": 300,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 380,
11
+ "num_train_epochs": 1,
12
+ "total_flos": 7.13246965235712e+16,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }