daydrill commited on
Commit
9cade0a
1 Parent(s): b1ca3af

Training in progress, step 500

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
added_tokens.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<td-d>": 32004,
3
+ "<td-du>": 32007,
4
+ "<td-l>": 32003,
5
+ "<td-ld>": 32009,
6
+ "<td-ldu>": 32015,
7
+ "<td-lu>": 32011,
8
+ "<td-r>": 32002,
9
+ "<td-rd>": 32008,
10
+ "<td-rdu>": 32014,
11
+ "<td-rl>": 32006,
12
+ "<td-rld>": 32012,
13
+ "<td-rldu>": 32016,
14
+ "<td-rlu>": 32013,
15
+ "<td-ru>": 32010,
16
+ "<td-u>": 32005,
17
+ "<td>": 32001,
18
+ "<tr>": 32000
19
+ }
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "klue/bert-base",
3
+ "architectures": [
4
+ "BertForQuestionAnswering"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.25.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 32017
25
+ }
last-checkpoint/added_tokens.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<td-d>": 32004,
3
+ "<td-du>": 32007,
4
+ "<td-l>": 32003,
5
+ "<td-ld>": 32009,
6
+ "<td-ldu>": 32015,
7
+ "<td-lu>": 32011,
8
+ "<td-r>": 32002,
9
+ "<td-rd>": 32008,
10
+ "<td-rdu>": 32014,
11
+ "<td-rl>": 32006,
12
+ "<td-rld>": 32012,
13
+ "<td-rldu>": 32016,
14
+ "<td-rlu>": 32013,
15
+ "<td-ru>": 32010,
16
+ "<td-u>": 32005,
17
+ "<td>": 32001,
18
+ "<tr>": 32000
19
+ }
last-checkpoint/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "klue/bert-base",
3
+ "architectures": [
4
+ "BertForQuestionAnswering"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.25.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 32017
25
+ }
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e7468bccdd3f376b9728d6b2977c0977c40a3c9c2ddc984e1bd3e480537b380
3
+ size 880446181
last-checkpoint/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e709852debbe95a6f40303894a5a28d7f8ecbefa9e7ab388289b5a43a8f9c9c
3
+ size 440235825
last-checkpoint/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f645c23e52a0f83c37f08090ab1bfc030983d8596077e1a318302512fdf0eef6
3
+ size 14503
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:357c660ac16aec412f85ea115a8091d7c93f34ee70a5b0adafa9a5eaa92162c3
3
+ size 623
last-checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<tr>",
4
+ "<td>",
5
+ "<td-r>",
6
+ "<td-l>",
7
+ "<td-d>",
8
+ "<td-u>",
9
+ "<td-rl>",
10
+ "<td-du>",
11
+ "<td-rd>",
12
+ "<td-ld>",
13
+ "<td-ru>",
14
+ "<td-lu>",
15
+ "<td-rld>",
16
+ "<td-rlu>",
17
+ "<td-rdu>",
18
+ "<td-ldu>",
19
+ "<td-rldu>"
20
+ ],
21
+ "cls_token": "[CLS]",
22
+ "mask_token": "[MASK]",
23
+ "pad_token": "[PAD]",
24
+ "sep_token": "[SEP]",
25
+ "unk_token": "[UNK]"
26
+ }
last-checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": false,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "name_or_path": "klue/bert-base",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": "/home/ec2-user/.cache/huggingface/hub/models--klue--bert-base/snapshots/812449f1a6bc736e693db7aa0e513e5e90795a62/special_tokens_map.json",
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
16
+ }
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.05207248489897938,
5
+ "global_step": 500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.05,
12
+ "learning_rate": 2.947927515101021e-05,
13
+ "loss": 2.2374,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.05,
18
+ "eval_exact_match": 49.46965745087811,
19
+ "eval_f1": 54.705922468625346,
20
+ "eval_loss": 1.399248480796814,
21
+ "eval_runtime": 152.8082,
22
+ "eval_samples_per_second": 92.96,
23
+ "eval_steps_per_second": 11.622,
24
+ "step": 500
25
+ }
26
+ ],
27
+ "max_steps": 28806,
28
+ "num_train_epochs": 3,
29
+ "total_flos": 1045187026944000.0,
30
+ "trial_name": null,
31
+ "trial_params": null
32
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:776fa4ef8f6f1f166b75d3366d3aaca8741a284e685b817922477a28bd89701c
3
+ size 3439
last-checkpoint/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
nbest_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e709852debbe95a6f40303894a5a28d7f8ecbefa9e7ab388289b5a43a8f9c9c
3
+ size 440235825
special_tokens_map.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<tr>",
4
+ "<td>",
5
+ "<td-r>",
6
+ "<td-l>",
7
+ "<td-d>",
8
+ "<td-u>",
9
+ "<td-rl>",
10
+ "<td-du>",
11
+ "<td-rd>",
12
+ "<td-ld>",
13
+ "<td-ru>",
14
+ "<td-lu>",
15
+ "<td-rld>",
16
+ "<td-rlu>",
17
+ "<td-rdu>",
18
+ "<td-ldu>",
19
+ "<td-rldu>"
20
+ ],
21
+ "cls_token": "[CLS]",
22
+ "mask_token": "[MASK]",
23
+ "pad_token": "[PAD]",
24
+ "sep_token": "[SEP]",
25
+ "unk_token": "[UNK]"
26
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": false,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "name_or_path": "klue/bert-base",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": "/home/ec2-user/.cache/huggingface/hub/models--klue--bert-base/snapshots/812449f1a6bc736e693db7aa0e513e5e90795a62/special_tokens_map.json",
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
16
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:776fa4ef8f6f1f166b75d3366d3aaca8741a284e685b817922477a28bd89701c
3
+ size 3439
vocab.txt ADDED
The diff for this file is too large to render. See raw diff