Commit
·
577295d
1
Parent(s):
a089475
Training in progress, epoch 1
Browse files- all_results.json +9 -9
- config.json +5 -5
- eval_results.json +5 -5
- pytorch_model.bin +2 -2
- runs/Jan29_15-48-48_node4/events.out.tfevents.1643442552.node4 +2 -2
- runs/Jan29_15-48-48_node4/events.out.tfevents.1643442660.node4 +3 -0
- runs/Jan29_16-31-17_node4/1643445112.230297/events.out.tfevents.1643445112.node4 +3 -0
- runs/Jan29_16-31-17_node4/events.out.tfevents.1643445112.node4 +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +1 -1
- train_results.json +4 -4
- trainer_state.json +59 -59
- training_args.bin +1 -1
all_results.json
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
-
"eval_accuracy": 0.
|
4 |
-
"eval_loss": 0.
|
5 |
-
"eval_runtime": 0.
|
6 |
"eval_samples": 277,
|
7 |
-
"eval_samples_per_second":
|
8 |
-
"eval_steps_per_second":
|
9 |
-
"train_loss": 0.
|
10 |
-
"train_runtime":
|
11 |
"train_samples": 2490,
|
12 |
-
"train_samples_per_second":
|
13 |
-
"train_steps_per_second":
|
14 |
}
|
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
+
"eval_accuracy": 0.5270758122743683,
|
4 |
+
"eval_loss": 0.6930166482925415,
|
5 |
+
"eval_runtime": 0.2259,
|
6 |
"eval_samples": 277,
|
7 |
+
"eval_samples_per_second": 1226.394,
|
8 |
+
"eval_steps_per_second": 13.282,
|
9 |
+
"train_loss": 0.6997218401004106,
|
10 |
+
"train_runtime": 107.0767,
|
11 |
"train_samples": 2490,
|
12 |
+
"train_samples_per_second": 232.544,
|
13 |
+
"train_steps_per_second": 7.284
|
14 |
}
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "../../../../plm_cache/roberta-
|
3 |
"architectures": [
|
4 |
"RobertaForSequenceClassification"
|
5 |
],
|
@@ -10,13 +10,13 @@
|
|
10 |
"finetuning_task": "rte",
|
11 |
"hidden_act": "gelu",
|
12 |
"hidden_dropout_prob": 0.1,
|
13 |
-
"hidden_size":
|
14 |
"id2label": {
|
15 |
"0": "entailment",
|
16 |
"1": "not_entailment"
|
17 |
},
|
18 |
"initializer_range": 0.02,
|
19 |
-
"intermediate_size":
|
20 |
"label2id": {
|
21 |
"entailment": 0,
|
22 |
"not_entailment": 1
|
@@ -24,8 +24,8 @@
|
|
24 |
"layer_norm_eps": 1e-05,
|
25 |
"max_position_embeddings": 514,
|
26 |
"model_type": "roberta",
|
27 |
-
"num_attention_heads":
|
28 |
-
"num_hidden_layers":
|
29 |
"pad_token_id": 1,
|
30 |
"position_embedding_type": "absolute",
|
31 |
"problem_type": "single_label_classification",
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "../../../../plm_cache/roberta-large",
|
3 |
"architectures": [
|
4 |
"RobertaForSequenceClassification"
|
5 |
],
|
|
|
10 |
"finetuning_task": "rte",
|
11 |
"hidden_act": "gelu",
|
12 |
"hidden_dropout_prob": 0.1,
|
13 |
+
"hidden_size": 1024,
|
14 |
"id2label": {
|
15 |
"0": "entailment",
|
16 |
"1": "not_entailment"
|
17 |
},
|
18 |
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4096,
|
20 |
"label2id": {
|
21 |
"entailment": 0,
|
22 |
"not_entailment": 1
|
|
|
24 |
"layer_norm_eps": 1e-05,
|
25 |
"max_position_embeddings": 514,
|
26 |
"model_type": "roberta",
|
27 |
+
"num_attention_heads": 16,
|
28 |
+
"num_hidden_layers": 24,
|
29 |
"pad_token_id": 1,
|
30 |
"position_embedding_type": "absolute",
|
31 |
"problem_type": "single_label_classification",
|
eval_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
-
"eval_accuracy": 0.
|
4 |
-
"eval_loss": 0.
|
5 |
-
"eval_runtime": 0.
|
6 |
"eval_samples": 277,
|
7 |
-
"eval_samples_per_second":
|
8 |
-
"eval_steps_per_second":
|
9 |
}
|
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
+
"eval_accuracy": 0.5270758122743683,
|
4 |
+
"eval_loss": 0.6930166482925415,
|
5 |
+
"eval_runtime": 0.2259,
|
6 |
"eval_samples": 277,
|
7 |
+
"eval_samples_per_second": 1226.394,
|
8 |
+
"eval_steps_per_second": 13.282
|
9 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc154e75055ed4ffe4f6fee8eb3b2c83fd827575bcbab6fd3b65b47693d12588
|
3 |
+
size 17106681
|
runs/Jan29_15-48-48_node4/events.out.tfevents.1643442552.node4
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61d693dfa0eb2f0e4882ba26a780134f70dae796a68b3cdf5bc8eaf2a81b26cc
|
3 |
+
size 7250
|
runs/Jan29_15-48-48_node4/events.out.tfevents.1643442660.node4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a37a54639e8f663b9e48afc468172b55940869c6f34d503fa8f0e51752ff238d
|
3 |
+
size 363
|
runs/Jan29_16-31-17_node4/1643445112.230297/events.out.tfevents.1643445112.node4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82d225b734cb18ede058b02137259711497603f297b73c1984d9fc069f368086
|
3 |
+
size 4585
|
runs/Jan29_16-31-17_node4/events.out.tfevents.1643445112.node4
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:583a5a78f47360afe9e409cd945eeec5395fbe3831f7bd1c7cec649c5bd82552
|
3 |
+
size 3517
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "trim_offsets": true, "special_tokens_map_file": null, "name_or_path": "../../../../plm_cache/roberta-
|
|
|
1 |
+
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "trim_offsets": true, "special_tokens_map_file": null, "name_or_path": "../../../../plm_cache/roberta-large", "tokenizer_class": "RobertaTokenizer"}
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 2490,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second":
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 10.0,
|
3 |
+
"train_loss": 0.6997218401004106,
|
4 |
+
"train_runtime": 107.0767,
|
5 |
"train_samples": 2490,
|
6 |
+
"train_samples_per_second": 232.544,
|
7 |
+
"train_steps_per_second": 7.284
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "outputs/lora/roberta-base/v2/rte/checkpoint-
|
4 |
"epoch": 10.0,
|
5 |
"global_step": 780,
|
6 |
"is_hyper_param_search": false,
|
@@ -10,112 +10,112 @@
|
|
10 |
{
|
11 |
"epoch": 1.0,
|
12 |
"eval_accuracy": 0.4729241877256318,
|
13 |
-
"eval_loss": 0.
|
14 |
-
"eval_runtime": 0.
|
15 |
-
"eval_samples_per_second":
|
16 |
-
"eval_steps_per_second":
|
17 |
"step": 78
|
18 |
},
|
19 |
{
|
20 |
"epoch": 2.0,
|
21 |
-
"eval_accuracy": 0.
|
22 |
-
"eval_loss": 0.
|
23 |
-
"eval_runtime": 0.
|
24 |
-
"eval_samples_per_second":
|
25 |
-
"eval_steps_per_second": 14.
|
26 |
"step": 156
|
27 |
},
|
28 |
{
|
29 |
"epoch": 3.0,
|
30 |
-
"eval_accuracy": 0.
|
31 |
-
"eval_loss": 0.
|
32 |
-
"eval_runtime": 0.
|
33 |
-
"eval_samples_per_second":
|
34 |
-
"eval_steps_per_second":
|
35 |
"step": 234
|
36 |
},
|
37 |
{
|
38 |
"epoch": 4.0,
|
39 |
-
"eval_accuracy": 0.
|
40 |
-
"eval_loss": 0.
|
41 |
-
"eval_runtime": 0.
|
42 |
-
"eval_samples_per_second":
|
43 |
-
"eval_steps_per_second":
|
44 |
"step": 312
|
45 |
},
|
46 |
{
|
47 |
"epoch": 5.0,
|
48 |
-
"eval_accuracy": 0.
|
49 |
-
"eval_loss": 0.
|
50 |
-
"eval_runtime": 0.
|
51 |
-
"eval_samples_per_second":
|
52 |
-
"eval_steps_per_second":
|
53 |
"step": 390
|
54 |
},
|
55 |
{
|
56 |
"epoch": 6.0,
|
57 |
-
"eval_accuracy": 0.
|
58 |
-
"eval_loss": 0.
|
59 |
-
"eval_runtime": 0.
|
60 |
-
"eval_samples_per_second":
|
61 |
-
"eval_steps_per_second":
|
62 |
"step": 468
|
63 |
},
|
64 |
{
|
65 |
"epoch": 6.41,
|
66 |
-
"learning_rate": 0.
|
67 |
-
"loss": 0.
|
68 |
"step": 500
|
69 |
},
|
70 |
{
|
71 |
"epoch": 7.0,
|
72 |
-
"eval_accuracy": 0.
|
73 |
-
"eval_loss": 0.
|
74 |
-
"eval_runtime": 0.
|
75 |
-
"eval_samples_per_second":
|
76 |
-
"eval_steps_per_second":
|
77 |
"step": 546
|
78 |
},
|
79 |
{
|
80 |
"epoch": 8.0,
|
81 |
-
"eval_accuracy": 0.
|
82 |
-
"eval_loss": 0.
|
83 |
-
"eval_runtime": 0.
|
84 |
-
"eval_samples_per_second":
|
85 |
-
"eval_steps_per_second": 14.
|
86 |
"step": 624
|
87 |
},
|
88 |
{
|
89 |
"epoch": 9.0,
|
90 |
-
"eval_accuracy": 0.
|
91 |
-
"eval_loss": 0.
|
92 |
-
"eval_runtime": 0.
|
93 |
-
"eval_samples_per_second":
|
94 |
-
"eval_steps_per_second": 14.
|
95 |
"step": 702
|
96 |
},
|
97 |
{
|
98 |
"epoch": 10.0,
|
99 |
-
"eval_accuracy": 0.
|
100 |
-
"eval_loss": 0.
|
101 |
-
"eval_runtime": 0.
|
102 |
-
"eval_samples_per_second":
|
103 |
-
"eval_steps_per_second":
|
104 |
"step": 780
|
105 |
},
|
106 |
{
|
107 |
"epoch": 10.0,
|
108 |
"step": 780,
|
109 |
-
"total_flos":
|
110 |
-
"train_loss": 0.
|
111 |
-
"train_runtime":
|
112 |
-
"train_samples_per_second":
|
113 |
-
"train_steps_per_second":
|
114 |
}
|
115 |
],
|
116 |
"max_steps": 780,
|
117 |
"num_train_epochs": 10,
|
118 |
-
"total_flos":
|
119 |
"trial_name": null,
|
120 |
"trial_params": null
|
121 |
}
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.5270758122743683,
|
3 |
+
"best_model_checkpoint": "outputs/lora/roberta-base/v2/rte/checkpoint-234",
|
4 |
"epoch": 10.0,
|
5 |
"global_step": 780,
|
6 |
"is_hyper_param_search": false,
|
|
|
10 |
{
|
11 |
"epoch": 1.0,
|
12 |
"eval_accuracy": 0.4729241877256318,
|
13 |
+
"eval_loss": 0.7130928039550781,
|
14 |
+
"eval_runtime": 0.2073,
|
15 |
+
"eval_samples_per_second": 1336.057,
|
16 |
+
"eval_steps_per_second": 14.47,
|
17 |
"step": 78
|
18 |
},
|
19 |
{
|
20 |
"epoch": 2.0,
|
21 |
+
"eval_accuracy": 0.4729241877256318,
|
22 |
+
"eval_loss": 0.6993798613548279,
|
23 |
+
"eval_runtime": 0.2002,
|
24 |
+
"eval_samples_per_second": 1383.767,
|
25 |
+
"eval_steps_per_second": 14.987,
|
26 |
"step": 156
|
27 |
},
|
28 |
{
|
29 |
"epoch": 3.0,
|
30 |
+
"eval_accuracy": 0.5270758122743683,
|
31 |
+
"eval_loss": 0.6930166482925415,
|
32 |
+
"eval_runtime": 0.2199,
|
33 |
+
"eval_samples_per_second": 1259.694,
|
34 |
+
"eval_steps_per_second": 13.643,
|
35 |
"step": 234
|
36 |
},
|
37 |
{
|
38 |
"epoch": 4.0,
|
39 |
+
"eval_accuracy": 0.5270758122743683,
|
40 |
+
"eval_loss": 0.6923182606697083,
|
41 |
+
"eval_runtime": 0.201,
|
42 |
+
"eval_samples_per_second": 1378.119,
|
43 |
+
"eval_steps_per_second": 14.925,
|
44 |
"step": 312
|
45 |
},
|
46 |
{
|
47 |
"epoch": 5.0,
|
48 |
+
"eval_accuracy": 0.4729241877256318,
|
49 |
+
"eval_loss": 0.696506917476654,
|
50 |
+
"eval_runtime": 0.2195,
|
51 |
+
"eval_samples_per_second": 1262.144,
|
52 |
+
"eval_steps_per_second": 13.669,
|
53 |
"step": 390
|
54 |
},
|
55 |
{
|
56 |
"epoch": 6.0,
|
57 |
+
"eval_accuracy": 0.5270758122743683,
|
58 |
+
"eval_loss": 0.6918743252754211,
|
59 |
+
"eval_runtime": 0.2123,
|
60 |
+
"eval_samples_per_second": 1304.889,
|
61 |
+
"eval_steps_per_second": 14.132,
|
62 |
"step": 468
|
63 |
},
|
64 |
{
|
65 |
"epoch": 6.41,
|
66 |
+
"learning_rate": 0.001145975443383356,
|
67 |
+
"loss": 0.7031,
|
68 |
"step": 500
|
69 |
},
|
70 |
{
|
71 |
"epoch": 7.0,
|
72 |
+
"eval_accuracy": 0.4729241877256318,
|
73 |
+
"eval_loss": 0.6935069561004639,
|
74 |
+
"eval_runtime": 0.2336,
|
75 |
+
"eval_samples_per_second": 1185.748,
|
76 |
+
"eval_steps_per_second": 12.842,
|
77 |
"step": 546
|
78 |
},
|
79 |
{
|
80 |
"epoch": 8.0,
|
81 |
+
"eval_accuracy": 0.4729241877256318,
|
82 |
+
"eval_loss": 0.6938801407814026,
|
83 |
+
"eval_runtime": 0.2086,
|
84 |
+
"eval_samples_per_second": 1327.87,
|
85 |
+
"eval_steps_per_second": 14.381,
|
86 |
"step": 624
|
87 |
},
|
88 |
{
|
89 |
"epoch": 9.0,
|
90 |
+
"eval_accuracy": 0.4729241877256318,
|
91 |
+
"eval_loss": 0.6936783194541931,
|
92 |
+
"eval_runtime": 0.2108,
|
93 |
+
"eval_samples_per_second": 1313.785,
|
94 |
+
"eval_steps_per_second": 14.229,
|
95 |
"step": 702
|
96 |
},
|
97 |
{
|
98 |
"epoch": 10.0,
|
99 |
+
"eval_accuracy": 0.4729241877256318,
|
100 |
+
"eval_loss": 0.6938663125038147,
|
101 |
+
"eval_runtime": 0.2046,
|
102 |
+
"eval_samples_per_second": 1353.687,
|
103 |
+
"eval_steps_per_second": 14.661,
|
104 |
"step": 780
|
105 |
},
|
106 |
{
|
107 |
"epoch": 10.0,
|
108 |
"step": 780,
|
109 |
+
"total_flos": 1639048592332800.0,
|
110 |
+
"train_loss": 0.6997218401004106,
|
111 |
+
"train_runtime": 107.0767,
|
112 |
+
"train_samples_per_second": 232.544,
|
113 |
+
"train_steps_per_second": 7.284
|
114 |
}
|
115 |
],
|
116 |
"max_steps": 780,
|
117 |
"num_train_epochs": 10,
|
118 |
+
"total_flos": 1639048592332800.0,
|
119 |
"trial_name": null,
|
120 |
"trial_params": null
|
121 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2991
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d3459cc4e350849119f9397e1e235b5df88dabf546417d4c780d5756a1941640
|
3 |
size 2991
|