Upload folder using huggingface_hub
Browse files- checkpoints/checkpoint-59000/config.json +29 -0
- checkpoints/checkpoint-59000/model.safetensors +3 -0
- checkpoints/checkpoint-59000/optimizer.pt +3 -0
- checkpoints/checkpoint-59000/rng_state.pth +3 -0
- checkpoints/checkpoint-59000/scheduler.pt +3 -0
- checkpoints/checkpoint-59000/trainer_state.json +0 -0
- checkpoints/checkpoint-59000/training_args.bin +3 -0
- checkpoints/checkpoint-60000/config.json +29 -0
- checkpoints/checkpoint-60000/model.safetensors +3 -0
- checkpoints/checkpoint-60000/optimizer.pt +3 -0
- checkpoints/checkpoint-60000/rng_state.pth +3 -0
- checkpoints/checkpoint-60000/scheduler.pt +3 -0
- checkpoints/checkpoint-60000/trainer_state.json +0 -0
- checkpoints/checkpoint-60000/training_args.bin +3 -0
- checkpoints/checkpoint-61000/config.json +29 -0
- checkpoints/checkpoint-61000/model.safetensors +3 -0
- checkpoints/checkpoint-61000/optimizer.pt +3 -0
- checkpoints/checkpoint-61000/rng_state.pth +3 -0
- checkpoints/checkpoint-61000/scheduler.pt +3 -0
- checkpoints/checkpoint-61000/trainer_state.json +0 -0
- checkpoints/checkpoint-61000/training_args.bin +3 -0
- checkpoints/checkpoint-62000/config.json +29 -0
- checkpoints/checkpoint-62000/model.safetensors +3 -0
- checkpoints/checkpoint-62000/optimizer.pt +3 -0
- checkpoints/checkpoint-62000/rng_state.pth +3 -0
- checkpoints/checkpoint-62000/scheduler.pt +3 -0
- checkpoints/checkpoint-62000/trainer_state.json +0 -0
- checkpoints/checkpoint-62000/training_args.bin +3 -0
- checkpoints/checkpoint-63000/config.json +29 -0
- checkpoints/checkpoint-63000/model.safetensors +3 -0
- checkpoints/checkpoint-63000/optimizer.pt +3 -0
- checkpoints/checkpoint-63000/rng_state.pth +3 -0
- checkpoints/checkpoint-63000/scheduler.pt +3 -0
- checkpoints/checkpoint-63000/trainer_state.json +0 -0
- checkpoints/checkpoint-63000/training_args.bin +3 -0
- checkpoints/config.json +1 -1
- checkpoints/model.safetensors +1 -1
- checkpoints/training_args.bin +1 -1
checkpoints/checkpoint-59000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-59000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a078dbe78cb827099c5bae10ccfba8c1ddb7460fcbeea37f6ccad55903f4880
|
3 |
+
size 44890256
|
checkpoints/checkpoint-59000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ff5797bb1965eb82b4f8c48f18cf1273223f620fe0778592eef4792662b65b1
|
3 |
+
size 89797322
|
checkpoints/checkpoint-59000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ebe18649ff4574c2dec0242d26b29cce66bd05756ddbc29bcbf2a6f242206875
|
3 |
+
size 14244
|
checkpoints/checkpoint-59000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b08589dd37d469539a32630639e905643a5649b953c8a84938858247be0b95b6
|
3 |
+
size 1064
|
checkpoints/checkpoint-59000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-59000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
+
size 5112
|
checkpoints/checkpoint-60000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-60000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:612220fdd0ae03018775f63b04c15d2cbc906797215fa017bfe1b99f0f7aae35
|
3 |
+
size 44890256
|
checkpoints/checkpoint-60000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd4931161a6760c42904c78c77f126fda0a033d828d759a511007af36bb7c9c8
|
3 |
+
size 89797322
|
checkpoints/checkpoint-60000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37f1d75b826d2368dfe21037f7d56228356c8da58c3eb8ccaa90ed91aa82751b
|
3 |
+
size 14244
|
checkpoints/checkpoint-60000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2e048ddb01688e28f20f7d5bb39e32253d75e60ab2b4380ebabdb908bac39c5
|
3 |
+
size 1064
|
checkpoints/checkpoint-60000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-60000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
+
size 5112
|
checkpoints/checkpoint-61000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-61000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8f5b16a9c8812969ac5b79666d98808cc593b7deb3c93866fbb38867b1e92048
|
3 |
+
size 44890256
|
checkpoints/checkpoint-61000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f3e50d647f5611e6a8c1f1eb3bb6e2b4635b37ea9ad7f5a45b44851764fc22f
|
3 |
+
size 89797322
|
checkpoints/checkpoint-61000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:256d928092b6f549999dafaa485be23e7bc6703bb8155793a25c8ac6254f4f38
|
3 |
+
size 14244
|
checkpoints/checkpoint-61000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90783fc1a8dac263de0f461a7fcd915c156560f52a2d1c2d467f6d08d3f30798
|
3 |
+
size 1064
|
checkpoints/checkpoint-61000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-61000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
+
size 5112
|
checkpoints/checkpoint-62000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-62000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67b4d1705cdcc198716ed4f7722a4ed08a202fbf7d3584ed8924e25f50697917
|
3 |
+
size 44890256
|
checkpoints/checkpoint-62000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94e20b667552e8bc07364efa8eef9d26ad835339eff964183635379e39af3a91
|
3 |
+
size 89797322
|
checkpoints/checkpoint-62000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e74c1f337e7e7b6faed05338b6f975a5fb474f48008169b43efe77ca2117e3c
|
3 |
+
size 14244
|
checkpoints/checkpoint-62000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96b241f1ddb61534bab840149364db8b8aec297c5895b68c513ef224f7436e85
|
3 |
+
size 1064
|
checkpoints/checkpoint-62000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-62000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
+
size 5112
|
checkpoints/checkpoint-63000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-63000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9bdf2140400c28602e5b9c9350668395fa8f9253f444034541608c6c018cace3
|
3 |
+
size 44890256
|
checkpoints/checkpoint-63000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8cac7b107eb68cfa2e4f09a44af5837a6c013ca2030d688f393e310ca26a3d8a
|
3 |
+
size 89797322
|
checkpoints/checkpoint-63000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3808d59f4160d63bc77299817fde9ef210448d71fadc5af12ef57ec682704a37
|
3 |
+
size 14244
|
checkpoints/checkpoint-63000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15b86d2fc451862e21996f902735985ed99a5b81bd5ef38222783c82c7f729af
|
3 |
+
size 1064
|
checkpoints/checkpoint-63000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-63000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
+
size 5112
|
checkpoints/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-54000",
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
checkpoints/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 44890256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9bdf2140400c28602e5b9c9350668395fa8f9253f444034541608c6c018cace3
|
3 |
size 44890256
|
checkpoints/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3122098f6dd0b39207b2763c78d63fe9889378f88849a628ec4adff3d17a326
|
3 |
size 5112
|