taufiqdp commited on
Commit
a6db988
1 Parent(s): 1207906

Upload folder using huggingface_hub

Browse files
Files changed (31) hide show
  1. checkpoints/checkpoint-51000/config.json +29 -0
  2. checkpoints/checkpoint-51000/model.safetensors +3 -0
  3. checkpoints/checkpoint-51000/optimizer.pt +3 -0
  4. checkpoints/checkpoint-51000/rng_state.pth +3 -0
  5. checkpoints/checkpoint-51000/scheduler.pt +3 -0
  6. checkpoints/checkpoint-51000/trainer_state.json +0 -0
  7. checkpoints/checkpoint-51000/training_args.bin +3 -0
  8. checkpoints/checkpoint-52000/config.json +29 -0
  9. checkpoints/checkpoint-52000/model.safetensors +3 -0
  10. checkpoints/checkpoint-52000/optimizer.pt +3 -0
  11. checkpoints/checkpoint-52000/rng_state.pth +3 -0
  12. checkpoints/checkpoint-52000/scheduler.pt +3 -0
  13. checkpoints/checkpoint-52000/trainer_state.json +0 -0
  14. checkpoints/checkpoint-52000/training_args.bin +3 -0
  15. checkpoints/checkpoint-53000/config.json +29 -0
  16. checkpoints/checkpoint-53000/model.safetensors +3 -0
  17. checkpoints/checkpoint-53000/optimizer.pt +3 -0
  18. checkpoints/checkpoint-53000/rng_state.pth +3 -0
  19. checkpoints/checkpoint-53000/scheduler.pt +3 -0
  20. checkpoints/checkpoint-53000/trainer_state.json +0 -0
  21. checkpoints/checkpoint-53000/training_args.bin +3 -0
  22. checkpoints/checkpoint-54000/config.json +29 -0
  23. checkpoints/checkpoint-54000/model.safetensors +3 -0
  24. checkpoints/checkpoint-54000/optimizer.pt +3 -0
  25. checkpoints/checkpoint-54000/rng_state.pth +3 -0
  26. checkpoints/checkpoint-54000/scheduler.pt +3 -0
  27. checkpoints/checkpoint-54000/trainer_state.json +0 -0
  28. checkpoints/checkpoint-54000/training_args.bin +3 -0
  29. checkpoints/config.json +1 -1
  30. checkpoints/model.safetensors +1 -1
  31. checkpoints/training_args.bin +1 -1
checkpoints/checkpoint-51000/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
3
+ "architectures": [
4
+ "AlbertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "embedding_size": 128,
10
+ "eos_token_id": 3,
11
+ "hidden_act": "gelu_new",
12
+ "hidden_dropout_prob": 0,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "inner_group_num": 1,
16
+ "intermediate_size": 3072,
17
+ "layer_norm_eps": 1e-12,
18
+ "max_position_embeddings": 512,
19
+ "model_type": "albert",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_groups": 1,
22
+ "num_hidden_layers": 12,
23
+ "pad_token_id": 0,
24
+ "position_embedding_type": "absolute",
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.38.2",
27
+ "type_vocab_size": 2,
28
+ "vocab_size": 30000
29
+ }
checkpoints/checkpoint-51000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e658cc46bdd8a72cb52f403eef5d565dae45616f78bd3c57533755f9cb6041
3
+ size 44890256
checkpoints/checkpoint-51000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46dde2ed75d759c8e38c1336599c03c653f1b57bca5188d3ff869e9c9bef71fc
3
+ size 89797322
checkpoints/checkpoint-51000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d80778c8232421fd236e39309a51aaaa60f2e382c3a48206f6dbc992f76c370b
3
+ size 14244
checkpoints/checkpoint-51000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba7d5590af05df6e51ad422f658f48d4b533eb50367380ec9c9eff7012fe42eb
3
+ size 1064
checkpoints/checkpoint-51000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-51000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
3
+ size 5112
checkpoints/checkpoint-52000/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
3
+ "architectures": [
4
+ "AlbertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "embedding_size": 128,
10
+ "eos_token_id": 3,
11
+ "hidden_act": "gelu_new",
12
+ "hidden_dropout_prob": 0,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "inner_group_num": 1,
16
+ "intermediate_size": 3072,
17
+ "layer_norm_eps": 1e-12,
18
+ "max_position_embeddings": 512,
19
+ "model_type": "albert",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_groups": 1,
22
+ "num_hidden_layers": 12,
23
+ "pad_token_id": 0,
24
+ "position_embedding_type": "absolute",
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.38.2",
27
+ "type_vocab_size": 2,
28
+ "vocab_size": 30000
29
+ }
checkpoints/checkpoint-52000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fa518a02969a2a8f6fdf740701e61a39bf4ac87eb9e5c5e17200dc0bb43f990
3
+ size 44890256
checkpoints/checkpoint-52000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c087906600c9662b09a8ff962630a01cc747ba7cf6e7fa2eff7c901618e2c19c
3
+ size 89797322
checkpoints/checkpoint-52000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f0d6128daece7ee138488d0405ba20683554aebe56f4e3774203b1ff5e43d32
3
+ size 14244
checkpoints/checkpoint-52000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ee8f4ed8a0fd1b13d34b4eda8cf9645e38e4339d6fa5ca42244af1d707bfba
3
+ size 1064
checkpoints/checkpoint-52000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-52000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
3
+ size 5112
checkpoints/checkpoint-53000/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
3
+ "architectures": [
4
+ "AlbertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "embedding_size": 128,
10
+ "eos_token_id": 3,
11
+ "hidden_act": "gelu_new",
12
+ "hidden_dropout_prob": 0,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "inner_group_num": 1,
16
+ "intermediate_size": 3072,
17
+ "layer_norm_eps": 1e-12,
18
+ "max_position_embeddings": 512,
19
+ "model_type": "albert",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_groups": 1,
22
+ "num_hidden_layers": 12,
23
+ "pad_token_id": 0,
24
+ "position_embedding_type": "absolute",
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.38.2",
27
+ "type_vocab_size": 2,
28
+ "vocab_size": 30000
29
+ }
checkpoints/checkpoint-53000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17183bf2f2d1e3e15f79ae798c9ccc9dcae0bb2add266d7b63cd3c880a971844
3
+ size 44890256
checkpoints/checkpoint-53000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b32bf7b4072067bce6267388c3c9ab5bc404502e425eb13afa2bf854016f7aa
3
+ size 89797322
checkpoints/checkpoint-53000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5121d7ea8e1c63607830973f6a99e34ee3b048bf1d8313fd391e150255e789be
3
+ size 14244
checkpoints/checkpoint-53000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a151fdab51c034cd3fc8b0b65fee7659b99e5753fa0dff3edb44e998ba4aba41
3
+ size 1064
checkpoints/checkpoint-53000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-53000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
3
+ size 5112
checkpoints/checkpoint-54000/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
3
+ "architectures": [
4
+ "AlbertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "embedding_size": 128,
10
+ "eos_token_id": 3,
11
+ "hidden_act": "gelu_new",
12
+ "hidden_dropout_prob": 0,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "inner_group_num": 1,
16
+ "intermediate_size": 3072,
17
+ "layer_norm_eps": 1e-12,
18
+ "max_position_embeddings": 512,
19
+ "model_type": "albert",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_groups": 1,
22
+ "num_hidden_layers": 12,
23
+ "pad_token_id": 0,
24
+ "position_embedding_type": "absolute",
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.38.2",
27
+ "type_vocab_size": 2,
28
+ "vocab_size": 30000
29
+ }
checkpoints/checkpoint-54000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a906631ee85a758dafc7f566772038ce5c928a568c5a705ae27a2e9909adb9c1
3
+ size 44890256
checkpoints/checkpoint-54000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07d4d67af68d2e7f47482742124a750511f13e4de0acc548a578349dbdd3da24
3
+ size 89797322
checkpoints/checkpoint-54000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:719dcc21f701523b0563a2506cce2906abe179e40c26e43d9781cc2652dc8f6c
3
+ size 14244
checkpoints/checkpoint-54000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:276adbe1735ddeb94c55b460a02fb15f06f275511fd3941aed6564ac8b3e883a
3
+ size 1064
checkpoints/checkpoint-54000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/checkpoint-54000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
3
+ size 5112
checkpoints/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-38000",
3
  "architectures": [
4
  "AlbertForMaskedLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
3
  "architectures": [
4
  "AlbertForMaskedLM"
5
  ],
checkpoints/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6282ac12001508f3896fc3dfa89f981a5d570184187728288c152c09968c763b
3
  size 44890256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a906631ee85a758dafc7f566772038ce5c928a568c5a705ae27a2e9909adb9c1
3
  size 44890256
checkpoints/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8859a43f8f50f6d107f5edda559ce599497f70507f4c174a92c481b462156193
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
3
  size 5112