consciousAI commited on
Commit
b065f1e
1 Parent(s): 7bd3fcb

Training epoch 10

Browse files
.ipynb_checkpoints/args-checkpoint.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"evaluation_strategy": "epoch", "generation_num_beams": 4, "save_total_limit": 2, "train_file_path": "data/training/oneToManyQsSquad2TrainEncoded.pt", "valid_file_path": "data/training/oneToManyQsSquad2ValidationEncoded.pt", "model_name_or_path": "t5-v1_1-base-squadV2AutoQgen", "max_len": 512, "target_max_len": 512, "output_dir": "t5-v1_1-base-squadV2AutoQgen", "overwrite_output_dir": true, "per_device_train_batch_size": 2, "per_device_eval_batch_size": 2, "learning_rate": 0.0003, "num_train_epochs": 6, "do_train": true, "do_eval": true, "remove_unused_columns": false, "predict_with_generate": true}
.ipynb_checkpoints/eval_results-checkpoint.txt ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ epoch = 4.0
2
+ eval_loss = 1.5862550735473633
3
+ eval_rouge1 = 0.1449
4
+ eval_rouge2 = 0.0701
5
+ eval_rougeL = 0.1268
6
+ eval_rougeLsum = 0.1397
7
+ eval_runtime = 300.07
8
+ eval_samples_per_second = 4.012
9
+ eval_steps_per_second = 1.003
.ipynb_checkpoints/special_tokens_map-checkpoint.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
107
+ }
.ipynb_checkpoints/tokenizer_config-checkpoint.json ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
+ "eos_token": "</s>",
105
+ "extra_ids": 100,
106
+ "model_max_length": 512,
107
+ "name_or_path": "google/t5-v1_1-base",
108
+ "pad_token": "<pad>",
109
+ "sp_model_kwargs": {},
110
+ "special_tokens_map_file": "/root/.cache/huggingface/hub/models--google--t5-v1_1-base/snapshots/650d7745bf1e502d6949b22cc19155cd656d3d4e/special_tokens_map.json",
111
+ "tokenizer_class": "T5Tokenizer",
112
+ "unk_token": "<unk>"
113
+ }
args.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"evaluation_strategy": "epoch", "generation_num_beams": 4, "save_total_limit": 2, "train_file_path": "data/training/oneToManyQsSquad2TrainEncoded.pt", "valid_file_path": "data/training/oneToManyQsSquad2ValidationEncoded.pt", "model_name_or_path": "t5-v1_1-base-squadV2AutoQgen", "max_len": 512, "target_max_len": 512, "output_dir": "t5-v1_1-base-squadV2AutoQgen", "overwrite_output_dir": true, "per_device_train_batch_size": 2, "per_device_eval_batch_size": 2, "learning_rate": 0.0003, "num_train_epochs": 6, "do_train": true, "do_eval": true, "remove_unused_columns": false, "predict_with_generate": true}
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "google/t5-v1_1-base",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "t5-v1_1-base-squadV2AutoQgen",
3
  "architectures": [
4
  "T5ForConditionalGeneration"
5
  ],
emissions.csv CHANGED
@@ -1,2 +1,3 @@
1
  timestamp,experiment_id,project_name,duration,emissions,energy_consumed,country_name,country_iso_code,region,on_cloud,cloud_provider,cloud_region
2
  2022-10-21T07:33:28,211e65ae-9e1b-4708-a048-fd6b1c4649c7,codecarbon,19003.147694826126,0.3301225522754675,0.8975599572470567,USA,USA,South Carolina,Y,gcp,us-east1
 
 
1
  timestamp,experiment_id,project_name,duration,emissions,energy_consumed,country_name,country_iso_code,region,on_cloud,cloud_provider,cloud_region
2
  2022-10-21T07:33:28,211e65ae-9e1b-4708-a048-fd6b1c4649c7,codecarbon,19003.147694826126,0.3301225522754675,0.8975599572470567,USA,USA,South Carolina,Y,gcp,us-east1
3
+ 2022-10-21T21:58:26,cf08465c-2316-40b0-b52f-1102e737170e,codecarbon,28416.326830387115,0.49351839981988527,1.3418118537789159,USA,USA,South Carolina,Y,gcp,us-east1
eval_results.txt CHANGED
@@ -1,9 +1,9 @@
1
- epoch = 4.0
2
- eval_loss = 1.5862550735473633
3
- eval_rouge1 = 0.1449
4
- eval_rouge2 = 0.0701
5
- eval_rougeL = 0.1268
6
- eval_rougeLsum = 0.1397
7
- eval_runtime = 300.07
8
- eval_samples_per_second = 4.012
9
- eval_steps_per_second = 1.003
 
1
+ epoch = 6.0
2
+ eval_loss = 1.6475605964660645
3
+ eval_rouge1 = 0.1407
4
+ eval_rouge2 = 0.0716
5
+ eval_rougeL = 0.1245
6
+ eval_rougeLsum = 0.1356
7
+ eval_runtime = 307.2934
8
+ eval_samples_per_second = 3.918
9
+ eval_steps_per_second = 0.98
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77ccc798116e28e5ab0c4c776fc64ab365bb7c73916bf9ed370c7c4de49567e8
3
  size 990406605
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ba5c4993cad6d9552d2684c595d42d54bf3d2ed20f3828c3893943ff7e921ad
3
  size 990406605
runs/Oct21_14-04-40_6df22a056764/1666361087.782606/events.out.tfevents.1666361087.6df22a056764.14246.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:703fce1d07baee9df20456f51ce0119256178ff96fb9614083a09fdfc626ea8d
3
+ size 5759
runs/Oct21_14-04-40_6df22a056764/events.out.tfevents.1666361087.6df22a056764.14246.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb920631d3f07ab6e039c7ce6860a3ed8c6c40d226b5c976f4cf001c7ae6181d
3
+ size 16272
runs/Oct21_14-04-40_6df22a056764/events.out.tfevents.1666389822.6df22a056764.14246.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13a9fde90bc1eb7e16b781756ecbfd371613130f68f9e43ec722c231b67e5fb9
3
+ size 523
tokenizer_config.json CHANGED
@@ -104,7 +104,7 @@
104
  "eos_token": "</s>",
105
  "extra_ids": 100,
106
  "model_max_length": 512,
107
- "name_or_path": "google/t5-v1_1-base",
108
  "pad_token": "<pad>",
109
  "sp_model_kwargs": {},
110
  "special_tokens_map_file": "/root/.cache/huggingface/hub/models--google--t5-v1_1-base/snapshots/650d7745bf1e502d6949b22cc19155cd656d3d4e/special_tokens_map.json",
 
104
  "eos_token": "</s>",
105
  "extra_ids": 100,
106
  "model_max_length": 512,
107
+ "name_or_path": "t5-v1_1-base-squadV2AutoQgen",
108
  "pad_token": "<pad>",
109
  "sp_model_kwargs": {},
110
  "special_tokens_map_file": "/root/.cache/huggingface/hub/models--google--t5-v1_1-base/snapshots/650d7745bf1e502d6949b22cc19155cd656d3d4e/special_tokens_map.json",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2eeb34c4d3a0a1f58550bfdaafb440b2296df5f916af2281076239194ca6f361
3
  size 3567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c90368d936d779ec3714e42f1d184f328fe63e98185c611e6637f3f3e921b9cb
3
  size 3567