k4black commited on
Commit
3f23f19
1 Parent(s): 30bc812

Training in progress, step 2000

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<NUM>": 32101,
3
+ "ELEM_SEP": 32102,
4
+ "METHOD_NAME": 32100
5
+ }
config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Salesforce/codet5-small",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 512,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "relu",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 2,
14
+ "feed_forward_proj": "relu",
15
+ "gradient_checkpointing": false,
16
+ "id2label": {
17
+ "0": "LABEL_0"
18
+ },
19
+ "initializer_factor": 1.0,
20
+ "is_encoder_decoder": true,
21
+ "is_gated_act": false,
22
+ "label2id": {
23
+ "LABEL_0": 0
24
+ },
25
+ "layer_norm_epsilon": 1e-06,
26
+ "model_type": "t5",
27
+ "n_positions": 512,
28
+ "num_decoder_layers": 6,
29
+ "num_heads": 8,
30
+ "num_layers": 6,
31
+ "output_past": true,
32
+ "pad_token_id": 0,
33
+ "relative_attention_max_distance": 128,
34
+ "relative_attention_num_buckets": 32,
35
+ "task_specific_params": {
36
+ "summarization": {
37
+ "early_stopping": true,
38
+ "length_penalty": 2.0,
39
+ "max_length": 200,
40
+ "min_length": 30,
41
+ "no_repeat_ngram_size": 3,
42
+ "num_beams": 4,
43
+ "prefix": "summarize: "
44
+ },
45
+ "translation_en_to_de": {
46
+ "early_stopping": true,
47
+ "max_length": 300,
48
+ "num_beams": 4,
49
+ "prefix": "translate English to German: "
50
+ },
51
+ "translation_en_to_fr": {
52
+ "early_stopping": true,
53
+ "max_length": 300,
54
+ "num_beams": 4,
55
+ "prefix": "translate English to French: "
56
+ },
57
+ "translation_en_to_ro": {
58
+ "early_stopping": true,
59
+ "max_length": 300,
60
+ "num_beams": 4,
61
+ "prefix": "translate English to Romanian: "
62
+ }
63
+ },
64
+ "torch_dtype": "float32",
65
+ "transformers_version": "4.27.1",
66
+ "use_cache": true,
67
+ "vocab_size": 32103
68
+ }
last-checkpoint/added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<NUM>": 32101,
3
+ "ELEM_SEP": 32102,
4
+ "METHOD_NAME": 32100
5
+ }
last-checkpoint/config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Salesforce/codet5-small",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 512,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "relu",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 2,
14
+ "feed_forward_proj": "relu",
15
+ "gradient_checkpointing": false,
16
+ "id2label": {
17
+ "0": "LABEL_0"
18
+ },
19
+ "initializer_factor": 1.0,
20
+ "is_encoder_decoder": true,
21
+ "is_gated_act": false,
22
+ "label2id": {
23
+ "LABEL_0": 0
24
+ },
25
+ "layer_norm_epsilon": 1e-06,
26
+ "model_type": "t5",
27
+ "n_positions": 512,
28
+ "num_decoder_layers": 6,
29
+ "num_heads": 8,
30
+ "num_layers": 6,
31
+ "output_past": true,
32
+ "pad_token_id": 0,
33
+ "relative_attention_max_distance": 128,
34
+ "relative_attention_num_buckets": 32,
35
+ "task_specific_params": {
36
+ "summarization": {
37
+ "early_stopping": true,
38
+ "length_penalty": 2.0,
39
+ "max_length": 200,
40
+ "min_length": 30,
41
+ "no_repeat_ngram_size": 3,
42
+ "num_beams": 4,
43
+ "prefix": "summarize: "
44
+ },
45
+ "translation_en_to_de": {
46
+ "early_stopping": true,
47
+ "max_length": 300,
48
+ "num_beams": 4,
49
+ "prefix": "translate English to German: "
50
+ },
51
+ "translation_en_to_fr": {
52
+ "early_stopping": true,
53
+ "max_length": 300,
54
+ "num_beams": 4,
55
+ "prefix": "translate English to French: "
56
+ },
57
+ "translation_en_to_ro": {
58
+ "early_stopping": true,
59
+ "max_length": 300,
60
+ "num_beams": 4,
61
+ "prefix": "translate English to Romanian: "
62
+ }
63
+ },
64
+ "torch_dtype": "float32",
65
+ "transformers_version": "4.27.1",
66
+ "use_cache": true,
67
+ "vocab_size": 32103
68
+ }
last-checkpoint/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.27.1"
7
+ }
last-checkpoint/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ac8d53de6ef44d1d3f69b45748893b1dc5bccdd167f72424ab8fe9f00bde48
3
+ size 484059375
last-checkpoint/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de5e3548230ae99ceb68be87aecd9e8e4464ba4a49e0be0a37a68d9a8fed2113
3
+ size 242019067
last-checkpoint/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:415f489b3ca13bf10b426de10c7fd2ffff4eab43ae8dad4b9ea617baa6e5dc10
3
+ size 14503
last-checkpoint/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f5090a9fe1a2692162b6c7a502a38ccad55e430ca4f5f4f67294b8e6a83fa35
3
+ size 559
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dd99f51d00c2e77b7672cb0c0d6b6097887ffbb0427fc5b39de699752efb63f
3
+ size 623
last-checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "ELEM_SEP"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "cls_token": {
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": true,
23
+ "rstrip": false,
24
+ "single_word": false
25
+ },
26
+ "mask_token": {
27
+ "content": "<mask>",
28
+ "lstrip": true,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "pad_token": {
34
+ "content": "<pad>",
35
+ "lstrip": false,
36
+ "normalized": true,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ },
40
+ "sep_token": {
41
+ "content": "</s>",
42
+ "lstrip": false,
43
+ "normalized": true,
44
+ "rstrip": false,
45
+ "single_word": false
46
+ },
47
+ "unk_token": {
48
+ "content": "<unk>",
49
+ "lstrip": false,
50
+ "normalized": true,
51
+ "rstrip": false,
52
+ "single_word": false
53
+ }
54
+ }
last-checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "cls_token": {
12
+ "__type": "AddedToken",
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "__type": "AddedToken",
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "errors": "replace",
28
+ "mask_token": {
29
+ "__type": "AddedToken",
30
+ "content": "<mask>",
31
+ "lstrip": true,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false
35
+ },
36
+ "model_max_length": 512,
37
+ "pad_token": {
38
+ "__type": "AddedToken",
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ "sep_token": {
46
+ "__type": "AddedToken",
47
+ "content": "</s>",
48
+ "lstrip": false,
49
+ "normalized": true,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "special_tokens_map_file": "/home/s5431786/.cache/huggingface/hub/models--Salesforce--codet5-small/snapshots/a642dc934e5475185369d09ac07091dfe72a31fc/special_tokens_map.json",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
+ "unk_token": {
57
+ "__type": "AddedToken",
58
+ "content": "<unk>",
59
+ "lstrip": false,
60
+ "normalized": true,
61
+ "rstrip": false,
62
+ "single_word": false
63
+ }
64
+ }
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.528552770614624,
3
+ "best_model_checkpoint": "/home2/s5431786/jb-internship-repr/results/Salesforce-codet5-small-java-small-selected-wo-tokens/checkpoint-2000",
4
+ "epoch": 0.27743098904147595,
5
+ "global_step": 2000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.28,
12
+ "learning_rate": 9.94363646983237e-05,
13
+ "loss": 2.1009,
14
+ "step": 2000
15
+ },
16
+ {
17
+ "epoch": 0.28,
18
+ "eval_exact_match": 0.31785774198959904,
19
+ "eval_loss": 1.528552770614624,
20
+ "eval_rouge1": 0.5232290073169683,
21
+ "eval_rouge2": 0.24795642157242634,
22
+ "eval_rougeL": 0.5233033679411545,
23
+ "eval_runtime": 163.446,
24
+ "eval_samples_per_second": 145.883,
25
+ "eval_steps_per_second": 1.523,
26
+ "step": 2000
27
+ }
28
+ ],
29
+ "max_steps": 36045,
30
+ "num_train_epochs": 5,
31
+ "total_flos": 2.5985625882624e+16,
32
+ "trial_name": null,
33
+ "trial_params": null
34
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60907d47cf0a1eea61f28a3f32b09b8c8cdfad806bdf2a9d6aa35353d5057709
3
+ size 3887
last-checkpoint/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de5e3548230ae99ceb68be87aecd9e8e4464ba4a49e0be0a37a68d9a8fed2113
3
+ size 242019067
special_tokens_map.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "ELEM_SEP"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "cls_token": {
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": true,
23
+ "rstrip": false,
24
+ "single_word": false
25
+ },
26
+ "mask_token": {
27
+ "content": "<mask>",
28
+ "lstrip": true,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "pad_token": {
34
+ "content": "<pad>",
35
+ "lstrip": false,
36
+ "normalized": true,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ },
40
+ "sep_token": {
41
+ "content": "</s>",
42
+ "lstrip": false,
43
+ "normalized": true,
44
+ "rstrip": false,
45
+ "single_word": false
46
+ },
47
+ "unk_token": {
48
+ "content": "<unk>",
49
+ "lstrip": false,
50
+ "normalized": true,
51
+ "rstrip": false,
52
+ "single_word": false
53
+ }
54
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "cls_token": {
12
+ "__type": "AddedToken",
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "__type": "AddedToken",
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "errors": "replace",
28
+ "mask_token": {
29
+ "__type": "AddedToken",
30
+ "content": "<mask>",
31
+ "lstrip": true,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false
35
+ },
36
+ "model_max_length": 512,
37
+ "pad_token": {
38
+ "__type": "AddedToken",
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ "sep_token": {
46
+ "__type": "AddedToken",
47
+ "content": "</s>",
48
+ "lstrip": false,
49
+ "normalized": true,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "special_tokens_map_file": "/home/s5431786/.cache/huggingface/hub/models--Salesforce--codet5-small/snapshots/a642dc934e5475185369d09ac07091dfe72a31fc/special_tokens_map.json",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
+ "unk_token": {
57
+ "__type": "AddedToken",
58
+ "content": "<unk>",
59
+ "lstrip": false,
60
+ "normalized": true,
61
+ "rstrip": false,
62
+ "single_word": false
63
+ }
64
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60907d47cf0a1eea61f28a3f32b09b8c8cdfad806bdf2a9d6aa35353d5057709
3
+ size 3887
vocab.json ADDED
The diff for this file is too large to render. See raw diff